Adriane Boyd commited on
Commit
d9d78b0
1 Parent(s): 1c6351a

Add xx_udv25_oldfrenchsrcmf_trf-0.0.1

Browse files
.gitattributes CHANGED
@@ -25,3 +25,8 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
28
+ *.whl filter=lfs diff=lfs merge=lfs -text
29
+ *.npz filter=lfs diff=lfs merge=lfs -text
30
+ *strings.json filter=lfs diff=lfs merge=lfs -text
31
+ vectors filter=lfs diff=lfs merge=lfs -text
32
+ model filter=lfs diff=lfs merge=lfs -text
LICENSE.txt ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ The treebank is licensed under the Creative Commons License Attribution-ShareAlike 4.0 International.
2
+
3
+ The complete license text is available at:
4
+ http://creativecommons.org/licenses/by-sa/4.0/legalcode
README.md ADDED
The diff for this file is too large to render. See raw diff
 
config.cfg ADDED
@@ -0,0 +1,254 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [paths]
2
+ train = "corpus/UD_Old_French-SRCMF/train.spacy"
3
+ dev = "corpus/UD_Old_French-SRCMF/dev.spacy"
4
+ vectors = null
5
+ init_tok2vec = null
6
+ tokenizer_source = "training/UD_Old_French-SRCMF/tokenizer/model-best"
7
+ transformer_source = "training/UD_Old_French-SRCMF/transformer/model-best"
8
+
9
+ [system]
10
+ gpu_allocator = "pytorch"
11
+ seed = 0
12
+
13
+ [nlp]
14
+ lang = "xx"
15
+ pipeline = ["experimental_char_ner_tokenizer","transformer","senter","tagger","morphologizer","parser","experimental_edit_tree_lemmatizer"]
16
+ batch_size = 64
17
+ disabled = ["senter"]
18
+ before_creation = null
19
+ after_creation = null
20
+ after_pipeline_creation = null
21
+ tokenizer = {"@tokenizers":"spacy-experimental.char_pretokenizer.v1"}
22
+
23
+ [components]
24
+
25
+ [components.experimental_char_ner_tokenizer]
26
+ factory = "experimental_char_ner_tokenizer"
27
+ scorer = {"@scorers":"spacy-experimental.tokenizer_scorer.v1"}
28
+
29
+ [components.experimental_char_ner_tokenizer.model]
30
+ @architectures = "spacy.TransitionBasedParser.v2"
31
+ state_type = "ner"
32
+ extra_state_tokens = false
33
+ hidden_width = 64
34
+ maxout_pieces = 2
35
+ use_upper = true
36
+ nO = null
37
+
38
+ [components.experimental_char_ner_tokenizer.model.tok2vec]
39
+ @architectures = "spacy.Tok2Vec.v2"
40
+
41
+ [components.experimental_char_ner_tokenizer.model.tok2vec.embed]
42
+ @architectures = "spacy.MultiHashEmbed.v2"
43
+ width = 128
44
+ attrs = ["ORTH","LOWER","IS_DIGIT","IS_ALPHA","IS_SPACE","IS_PUNCT"]
45
+ rows = [1000,500,50,50,50,50]
46
+ include_static_vectors = false
47
+
48
+ [components.experimental_char_ner_tokenizer.model.tok2vec.encode]
49
+ @architectures = "spacy.MaxoutWindowEncoder.v2"
50
+ width = 128
51
+ depth = 4
52
+ window_size = 4
53
+ maxout_pieces = 2
54
+
55
+ [components.experimental_edit_tree_lemmatizer]
56
+ factory = "experimental_edit_tree_lemmatizer"
57
+ backoff = "orth"
58
+ min_tree_freq = 1
59
+ overwrite = false
60
+ scorer = {"@scorers":"spacy.lemmatizer_scorer.v1"}
61
+ top_k = 1
62
+
63
+ [components.experimental_edit_tree_lemmatizer.model]
64
+ @architectures = "spacy.Tagger.v1"
65
+ nO = null
66
+
67
+ [components.experimental_edit_tree_lemmatizer.model.tok2vec]
68
+ @architectures = "spacy-transformers.TransformerListener.v1"
69
+ grad_factor = 1.0
70
+ upstream = "transformer"
71
+ pooling = {"@layers":"reduce_mean.v1"}
72
+
73
+ [components.morphologizer]
74
+ factory = "morphologizer"
75
+ extend = false
76
+ overwrite = false
77
+ scorer = {"@scorers":"spacy.morphologizer_scorer.v1"}
78
+
79
+ [components.morphologizer.model]
80
+ @architectures = "spacy.Tagger.v1"
81
+ nO = null
82
+
83
+ [components.morphologizer.model.tok2vec]
84
+ @architectures = "spacy-transformers.TransformerListener.v1"
85
+ grad_factor = 1.0
86
+ upstream = "transformer"
87
+ pooling = {"@layers":"reduce_mean.v1"}
88
+
89
+ [components.parser]
90
+ factory = "parser"
91
+ learn_tokens = false
92
+ min_action_freq = 5
93
+ moves = null
94
+ scorer = {"@scorers":"spacy.parser_scorer.v1"}
95
+ update_with_oracle_cut_size = 100
96
+
97
+ [components.parser.model]
98
+ @architectures = "spacy.TransitionBasedParser.v2"
99
+ state_type = "parser"
100
+ extra_state_tokens = false
101
+ hidden_width = 64
102
+ maxout_pieces = 3
103
+ use_upper = false
104
+ nO = null
105
+
106
+ [components.parser.model.tok2vec]
107
+ @architectures = "spacy-transformers.TransformerListener.v1"
108
+ grad_factor = 1.0
109
+ upstream = "transformer"
110
+ pooling = {"@layers":"reduce_mean.v1"}
111
+
112
+ [components.senter]
113
+ factory = "senter"
114
+ overwrite = false
115
+ scorer = {"@scorers":"spacy.senter_scorer.v1"}
116
+
117
+ [components.senter.model]
118
+ @architectures = "spacy.Tagger.v1"
119
+ nO = null
120
+
121
+ [components.senter.model.tok2vec]
122
+ @architectures = "spacy-transformers.TransformerListener.v1"
123
+ grad_factor = 1.0
124
+ upstream = "transformer"
125
+ pooling = {"@layers":"reduce_mean.v1"}
126
+
127
+ [components.tagger]
128
+ factory = "tagger"
129
+ neg_prefix = "!"
130
+ overwrite = false
131
+ scorer = {"@scorers":"spacy.tagger_scorer.v1"}
132
+
133
+ [components.tagger.model]
134
+ @architectures = "spacy.Tagger.v1"
135
+ nO = null
136
+
137
+ [components.tagger.model.tok2vec]
138
+ @architectures = "spacy-transformers.TransformerListener.v1"
139
+ grad_factor = 1.0
140
+ upstream = "transformer"
141
+ pooling = {"@layers":"reduce_mean.v1"}
142
+
143
+ [components.transformer]
144
+ factory = "transformer"
145
+ max_batch_items = 4096
146
+ set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
147
+
148
+ [components.transformer.model]
149
+ @architectures = "spacy-transformers.TransformerModel.v3"
150
+ name = "xlm-roberta-base"
151
+ mixed_precision = true
152
+
153
+ [components.transformer.model.get_spans]
154
+ @span_getters = "spacy-transformers.strided_spans.v1"
155
+ window = 128
156
+ stride = 96
157
+
158
+ [components.transformer.model.grad_scaler_config]
159
+
160
+ [components.transformer.model.tokenizer_config]
161
+ use_fast = true
162
+
163
+ [components.transformer.model.transformer_config]
164
+
165
+ [corpora]
166
+
167
+ [corpora.dev]
168
+ @readers = "spacy.Corpus.v1"
169
+ path = ${paths.dev}
170
+ max_length = 0
171
+ gold_preproc = false
172
+ limit = 0
173
+ augmenter = null
174
+
175
+ [corpora.train]
176
+ @readers = "spacy.Corpus.v1"
177
+ path = ${paths.train}
178
+ max_length = 0
179
+ gold_preproc = false
180
+ limit = 0
181
+ augmenter = null
182
+
183
+ [training]
184
+ train_corpus = "corpora.train"
185
+ dev_corpus = "corpora.dev"
186
+ seed = ${system:seed}
187
+ gpu_allocator = ${system:gpu_allocator}
188
+ dropout = 0.1
189
+ accumulate_gradient = 3
190
+ patience = 5000
191
+ max_epochs = 0
192
+ max_steps = 20000
193
+ eval_frequency = 200
194
+ frozen_components = []
195
+ before_to_disk = null
196
+ annotating_components = []
197
+
198
+ [training.batcher]
199
+ @batchers = "spacy.batch_by_padded.v1"
200
+ discard_oversize = true
201
+ get_length = null
202
+ size = 2000
203
+ buffer = 256
204
+
205
+ [training.logger]
206
+ @loggers = "spacy.ConsoleLogger.v1"
207
+ progress_bar = false
208
+
209
+ [training.optimizer]
210
+ @optimizers = "Adam.v1"
211
+ beta1 = 0.9
212
+ beta2 = 0.999
213
+ L2_is_weight_decay = true
214
+ L2 = 0.01
215
+ grad_clip = 1.0
216
+ use_averages = true
217
+ eps = 0.00000001
218
+
219
+ [training.optimizer.learn_rate]
220
+ @schedules = "warmup_linear.v1"
221
+ warmup_steps = 250
222
+ total_steps = 20000
223
+ initial_rate = 0.00005
224
+
225
+ [training.score_weights]
226
+ token_f = 0.0
227
+ token_p = null
228
+ token_r = null
229
+ token_acc = null
230
+ sents_f = 0.05
231
+ sents_p = 0.0
232
+ sents_r = 0.0
233
+ tag_acc = 0.11
234
+ pos_acc = 0.05
235
+ morph_acc = 0.05
236
+ morph_per_feat = null
237
+ dep_uas = 0.11
238
+ dep_las = 0.11
239
+ dep_las_per_type = null
240
+ lemma_acc = 0.52
241
+
242
+ [pretraining]
243
+
244
+ [initialize]
245
+ vectors = ${paths.vectors}
246
+ init_tok2vec = ${paths.init_tok2vec}
247
+ vocab_data = null
248
+ lookups = null
249
+ before_init = null
250
+ after_init = null
251
+
252
+ [initialize.components]
253
+
254
+ [initialize.tokenizer]
experimental_char_ner_tokenizer/cfg ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "moves":null,
3
+ "update_with_oracle_cut_size":100,
4
+ "multitasks":[
5
+
6
+ ],
7
+ "min_action_freq":1,
8
+ "learn_tokens":false,
9
+ "beam_width":1,
10
+ "beam_density":0.0,
11
+ "beam_update_prob":0.0,
12
+ "incorrect_spans_key":null
13
+ }
experimental_char_ner_tokenizer/model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32150f9486c4799acfd9ee89daaa16f056286fb250ddbe286fc92334a91ffe48
3
+ size 6922248
experimental_char_ner_tokenizer/moves ADDED
@@ -0,0 +1 @@
 
 
1
+ ��moves�l{"0":{},"1":{"TOKEN":528469},"2":{"TOKEN":528469},"3":{"TOKEN":528469},"4":{"TOKEN":528469,"":1},"5":{"":1}}�cfg��neg_key�
experimental_edit_tree_lemmatizer/cfg ADDED
The diff for this file is too large to render. See raw diff
 
experimental_edit_tree_lemmatizer/model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53b1f84cedcbd3a14332e24af8a403277555e36ffe17126d09bf90fdba031a4e
3
+ size 49385781
experimental_edit_tree_lemmatizer/trees ADDED
Binary file (932 kB). View file
 
meta.json ADDED
The diff for this file is too large to render. See raw diff
 
morphologizer/cfg ADDED
@@ -0,0 +1,122 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "extend":false,
3
+ "labels_morph":{
4
+ "POS=CCONJ":"",
5
+ "Definite=Def|POS=DET|PronType=Art":"Definite=Def|PronType=Art",
6
+ "POS=NOUN":"",
7
+ "POS=PRON|PronType=Prs":"PronType=Prs",
8
+ "POS=VERB|VerbForm=Fin":"VerbForm=Fin",
9
+ "POS=PROPN":"",
10
+ "POS=PRON|PronType=Prs,Rel":"PronType=Prs,Rel",
11
+ "POS=ADV":"",
12
+ "POS=ADP":"",
13
+ "POS=ADV|PronType=Dem":"PronType=Dem",
14
+ "POS=PRON|PronType=Dem":"PronType=Dem",
15
+ "POS=VERB|Tense=Past|VerbForm=Part":"Tense=Past|VerbForm=Part",
16
+ "POS=AUX|VerbForm=Fin":"VerbForm=Fin",
17
+ "POS=DET|PronType=Int":"PronType=Int",
18
+ "POS=ADJ":"",
19
+ "POS=PRON|PronType=Ind":"PronType=Ind",
20
+ "POS=DET|PronType=Ind":"PronType=Ind",
21
+ "Morph=VPar|POS=ADJ":"Morph=VPar",
22
+ "POS=DET|Poss=Yes":"Poss=Yes",
23
+ "POS=ADV|Polarity=Neg":"Polarity=Neg",
24
+ "Definite=Def|POS=ADP|PronType=Art":"Definite=Def|PronType=Art",
25
+ "POS=PRON|PronType=Int":"PronType=Int",
26
+ "POS=SCONJ":"",
27
+ "POS=VERB|VerbForm=Inf":"VerbForm=Inf",
28
+ "NumType=Card|POS=PRON":"NumType=Card",
29
+ "POS=PRON":"",
30
+ "NumType=Card|POS=DET":"NumType=Card",
31
+ "POS=PRON|Polarity=Neg|PronType=Prs":"Polarity=Neg|PronType=Prs",
32
+ "POS=ADJ|Poss=Yes":"Poss=Yes",
33
+ "POS=PRON|Poss=Yes|PronType=Prs":"Poss=Yes|PronType=Prs",
34
+ "Definite=Ind|POS=DET|PronType=Art":"Definite=Ind|PronType=Art",
35
+ "POS=DET|PronType=Dem":"PronType=Dem",
36
+ "POS=AUX|VerbForm=Inf":"VerbForm=Inf",
37
+ "POS=ADJ|PronType=Ind":"PronType=Ind",
38
+ "Morph=VPar|POS=NOUN":"Morph=VPar",
39
+ "POS=VERB|Tense=Pres|VerbForm=Part":"Tense=Pres|VerbForm=Part",
40
+ "Morph=VPar|POS=PROPN":"Morph=VPar",
41
+ "Morph=VInf|POS=NOUN":"Morph=VInf",
42
+ "NumType=Ord|POS=PRON":"NumType=Ord",
43
+ "POS=INTJ":"",
44
+ "POS=SCONJ|PronType=Prs":"PronType=Prs",
45
+ "Morph=VFin|POS=NOUN":"Morph=VFin",
46
+ "POS=DET|PronType=Rel":"PronType=Rel",
47
+ "NumType=Card|POS=ADJ":"NumType=Card",
48
+ "POS=ADJ|PronType=Ord":"PronType=Ord",
49
+ "Morph=VFin|POS=ADV":"Morph=VFin",
50
+ "Morph=VFin|POS=PROPN":"Morph=VFin",
51
+ "POS=DET":"",
52
+ "Morph=VPar|POS=ADP":"Morph=VPar",
53
+ "Morph=VPar|POS=ADV":"Morph=VPar",
54
+ "NumType=Ord|POS=DET":"NumType=Ord",
55
+ "Morph=VFin|POS=ADP":"Morph=VFin",
56
+ "Morph=VFin|POS=CCONJ":"Morph=VFin",
57
+ "Morph=VInf|POS=ADJ":"Morph=VInf",
58
+ "POS=ADP|PronType=Dem":"PronType=Dem",
59
+ "POS=ADV|Polarity=Int":"Polarity=Int",
60
+ "Morph=VFin|POS=INTJ":"Morph=VFin"
61
+ },
62
+ "labels_pos":{
63
+ "POS=CCONJ":89,
64
+ "Definite=Def|POS=DET|PronType=Art":90,
65
+ "POS=NOUN":92,
66
+ "POS=PRON|PronType=Prs":95,
67
+ "POS=VERB|VerbForm=Fin":100,
68
+ "POS=PROPN":96,
69
+ "POS=PRON|PronType=Prs,Rel":95,
70
+ "POS=ADV":86,
71
+ "POS=ADP":85,
72
+ "POS=ADV|PronType=Dem":86,
73
+ "POS=PRON|PronType=Dem":95,
74
+ "POS=VERB|Tense=Past|VerbForm=Part":100,
75
+ "POS=AUX|VerbForm=Fin":87,
76
+ "POS=DET|PronType=Int":90,
77
+ "POS=ADJ":84,
78
+ "POS=PRON|PronType=Ind":95,
79
+ "POS=DET|PronType=Ind":90,
80
+ "Morph=VPar|POS=ADJ":84,
81
+ "POS=DET|Poss=Yes":90,
82
+ "POS=ADV|Polarity=Neg":86,
83
+ "Definite=Def|POS=ADP|PronType=Art":85,
84
+ "POS=PRON|PronType=Int":95,
85
+ "POS=SCONJ":98,
86
+ "POS=VERB|VerbForm=Inf":100,
87
+ "NumType=Card|POS=PRON":95,
88
+ "POS=PRON":95,
89
+ "NumType=Card|POS=DET":90,
90
+ "POS=PRON|Polarity=Neg|PronType=Prs":95,
91
+ "POS=ADJ|Poss=Yes":84,
92
+ "POS=PRON|Poss=Yes|PronType=Prs":95,
93
+ "Definite=Ind|POS=DET|PronType=Art":90,
94
+ "POS=DET|PronType=Dem":90,
95
+ "POS=AUX|VerbForm=Inf":87,
96
+ "POS=ADJ|PronType=Ind":84,
97
+ "Morph=VPar|POS=NOUN":92,
98
+ "POS=VERB|Tense=Pres|VerbForm=Part":100,
99
+ "Morph=VPar|POS=PROPN":96,
100
+ "Morph=VInf|POS=NOUN":92,
101
+ "NumType=Ord|POS=PRON":95,
102
+ "POS=INTJ":91,
103
+ "POS=SCONJ|PronType=Prs":98,
104
+ "Morph=VFin|POS=NOUN":92,
105
+ "POS=DET|PronType=Rel":90,
106
+ "NumType=Card|POS=ADJ":84,
107
+ "POS=ADJ|PronType=Ord":84,
108
+ "Morph=VFin|POS=ADV":86,
109
+ "Morph=VFin|POS=PROPN":96,
110
+ "POS=DET":90,
111
+ "Morph=VPar|POS=ADP":85,
112
+ "Morph=VPar|POS=ADV":86,
113
+ "NumType=Ord|POS=DET":90,
114
+ "Morph=VFin|POS=ADP":85,
115
+ "Morph=VFin|POS=CCONJ":89,
116
+ "Morph=VInf|POS=ADJ":84,
117
+ "POS=ADP|PronType=Dem":85,
118
+ "POS=ADV|Polarity=Int":86,
119
+ "Morph=VFin|POS=INTJ":91
120
+ },
121
+ "overwrite":false
122
+ }
morphologizer/model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2e73db06b05ece5c866a667df334dffed78ba3b827ce6d56b288a93603402e3
3
+ size 175922
parser/cfg ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "moves":null,
3
+ "update_with_oracle_cut_size":100,
4
+ "multitasks":[
5
+
6
+ ],
7
+ "min_action_freq":5,
8
+ "learn_tokens":false,
9
+ "beam_width":1,
10
+ "beam_density":0.0,
11
+ "beam_update_prob":0.0,
12
+ "incorrect_spans_key":null
13
+ }
parser/model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a12f578d04f09895cbfe7c4df4d516eb685e09d9283a6a2b8b31ae478690c96
3
+ size 1066773
parser/moves ADDED
@@ -0,0 +1 @@
 
 
1
+ ��moves�{"0":{"":87656},"1":{"":34458},"2":{"det":13130,"case":10431,"advmod":10283,"nsubj":10275,"obj":6311,"mark":5679,"obl":5122,"cc:nc":4602,"cc":3674,"aux":3042,"amod":2408,"iobj":2387,"cop":1892,"case:det":1682,"advcl":1182,"expl":1100,"obl:advmod":934,"vocative":683,"nummod":550,"xcomp":282,"parataxis":273,"aux:pass":246,"obj||xcomp":233,"advmod||obj":149,"nmod":115,"discourse":111,"acl":111,"dislocated":92,"obl||xcomp":76,"nsubj:obj":64,"csubj":53,"iobj||xcomp":41,"nmod||obj":40,"advmod||advcl":37,"advmod||obl":26,"advmod||nsubj":25,"ccomp":17,"appos":15,"advmod||advmod":14,"advcl||aux":13,"amod||obj":12,"mark||obj":11,"mark:obl":10,"nmod||nsubj":8,"obj||ccomp":7,"mark:advmod":7,"expl||xcomp":7,"obj:advmod":6,"mark:obj":6,"amod||nsubj":6,"advmod||xcomp":6,"acl:relcl||nsubj":6,"xcomp||xcomp":5,"obl||ccomp":5,"obl:advmod||xcomp":5,"acl||nsubj":5,"dep":0},"3":{"obl":5765,"obj":4256,"conj":3520,"advmod":2800,"nsubj":2482,"nmod":2075,"advcl":1842,"acl:relcl":1837,"xcomp":1701,"ccomp":1063,"amod":972,"acl":880,"flat":822,"appos":466,"cop":460,"aux":392,"iobj":364,"conj||obj":223,"conj||nsubj":175,"acl:relcl||nsubj":167,"acl||obj":142,"csubj":140,"conj||obl":130,"vocative":114,"acl:relcl||obj":110,"cc||conj":107,"parataxis":98,"expl":96,"amod||obj":80,"acl||nsubj":78,"dislocated":72,"nmod||advmod":65,"obl:advmod":59,"acl:relcl||obl":56,"advcl||amod":51,"acl||obl":48,"amod||nsubj":44,"nmod||obj":42,"case":30,"conj||amod":27,"aux:pass":26,"conj||xcomp":21,"advcl||aux":21,"fixed||mark":20,"advmod||obj":20,"obl||xcomp":19,"nummod":19,"conj||advmod":19,"ccomp||aux":19,"appos||nsubj":18,"advcl||advmod":18,"conj||aux":16,"nmod||nsubj":13,"nmod||amod":13,"appos||obj":13,"det":12,"conj||iobj":12,"obj||xcomp":11,"conj||case":11,"acl||amod":10,"acl:relcl||conj":10,"mark":9,"cc":9,"acl||conj":8,"obl||advmod":7,"conj||nummod":7,"conj||conj":7,"acl||expl":7,"acl:relcl||nmod":7,"advmod||xcomp":6,"advmod||advmod":6,"xcomp||xcomp":5,"nmod||obl":5,"ccomp||xcomp":5,"amod||obl":5,"advmod||nsubj":5,"advmod||advcl":5,"advcl||obj":5,"dep":0},"4":{"ROOT":13909}}�cfg��neg_key�
senter/cfg ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "overwrite":false
3
+ }
senter/model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e9c8a175597d5d1bb8e9593f8f10ce42a9c146bc23e7f76bd53970e2cfefa49
3
+ size 6740
tagger/cfg ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "labels":[
3
+ "ADJQUA",
4
+ "ADJcar",
5
+ "ADJind",
6
+ "ADJord",
7
+ "ADJpos",
8
+ "ADJqua",
9
+ "ADVgen",
10
+ "ADVgen.PROadv",
11
+ "ADVgen.PROper",
12
+ "ADVing",
13
+ "ADVint",
14
+ "ADVneg",
15
+ "ADVneg.PROper",
16
+ "ADVsub",
17
+ "CONcoo",
18
+ "CONsub",
19
+ "CONsub.PROper",
20
+ "CONsub_o",
21
+ "CONsub_pre",
22
+ "DETcar",
23
+ "DETdef",
24
+ "DETdem",
25
+ "DETind",
26
+ "DETint",
27
+ "DETndf",
28
+ "DETord",
29
+ "DETpos",
30
+ "DETrel",
31
+ "DETrel_o",
32
+ "ETR",
33
+ "INJ",
34
+ "NOMcom",
35
+ "NOMcom.PROper",
36
+ "NOMpro",
37
+ "PRE",
38
+ "PRE.DETdef",
39
+ "PRE.PROdem",
40
+ "PRE.PROper",
41
+ "PROadv",
42
+ "PROcar",
43
+ "PROdem",
44
+ "PROimp",
45
+ "PROind",
46
+ "PROint",
47
+ "PROint.PROper",
48
+ "PROint_adv",
49
+ "PROord",
50
+ "PROper",
51
+ "PROper.PROper",
52
+ "PROpos",
53
+ "PROrel",
54
+ "PROrel.ADVneg",
55
+ "PROrel.PROadv",
56
+ "PROrel.PROper",
57
+ "PROrel_adv",
58
+ "RED",
59
+ "VERcjg",
60
+ "VERinf",
61
+ "VERppa",
62
+ "VERppe"
63
+ ],
64
+ "neg_prefix":"!",
65
+ "overwrite":false
66
+ }
tagger/model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd48f2365ce10d1738c6d261b6a765853459d37370e7edf17b5112161802b48d
3
+ size 185150
transformer/cfg ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "max_batch_items":4096
3
+ }
transformer/model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:690b268a649be18f30586a1dfac4d0d4fb719731f025d9b6b75a3d47b478a3af
3
+ size 1126406104
vocab/key2row ADDED
@@ -0,0 +1 @@
 
 
1
+
vocab/lookups.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76be8b528d0075f7aae98d6fa57a6d3c83ae480a8469e668d7b0af968995ac71
3
+ size 1
vocab/strings.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:288c67639baa41bedcaf72c2e5fea0ce9468cd59aed948f7222aca3b9aa5b83f
3
+ size 617864
vocab/vectors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14772b683e726436d5948ad3fff2b43d036ef2ebbe3458aafed6004e05a40706
3
+ size 128
vocab/vectors.cfg ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "mode":"default"
3
+ }
xx_udv25_oldfrenchsrcmf_trf-any-py3-none-any.whl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:776a6d5e6a0dbd54b7ee92bb06217b4d8136ff11e11a73ceddcf7edfebfac376
3
+ size 891852582