egumasa commited on
Commit
80bcf22
·
1 Parent(s): 85ac6af

Update spaCy pipeline

Browse files
README.md CHANGED
@@ -5,60 +5,12 @@ language:
5
  - en
6
  model-index:
7
  - name: en_engagement_spl_RoBERTa_acad
8
- results:
9
- - task:
10
- name: NER
11
- type: token-classification
12
- metrics:
13
- - name: NER Precision
14
- type: precision
15
- value: 0.0
16
- - name: NER Recall
17
- type: recall
18
- value: 0.0
19
- - name: NER F Score
20
- type: f_score
21
- value: 0.0
22
- - task:
23
- name: TAG
24
- type: token-classification
25
- metrics:
26
- - name: TAG (XPOS) Accuracy
27
- type: accuracy
28
- value: 0.0
29
- - task:
30
- name: LEMMA
31
- type: token-classification
32
- metrics:
33
- - name: Lemma Accuracy
34
- type: accuracy
35
- value: 0.0
36
- - task:
37
- name: UNLABELED_DEPENDENCIES
38
- type: token-classification
39
- metrics:
40
- - name: Unlabeled Attachment Score (UAS)
41
- type: f_score
42
- value: 0.0
43
- - task:
44
- name: LABELED_DEPENDENCIES
45
- type: token-classification
46
- metrics:
47
- - name: Labeled Attachment Score (LAS)
48
- type: f_score
49
- value: 0.0
50
- - task:
51
- name: SENTS
52
- type: token-classification
53
- metrics:
54
- - name: Sentences F-Score
55
- type: f_score
56
- value: 0.9301042305
57
  ---
58
  | Feature | Description |
59
  | --- | --- |
60
  | **Name** | `en_engagement_spl_RoBERTa_acad` |
61
- | **Version** | `0.3.4.1221` |
62
  | **spaCy** | `>=3.4.4,<3.5.0` |
63
  | **Default Pipeline** | `transformer`, `parser`, `trainable_transformer`, `spancat` |
64
  | **Components** | `transformer`, `parser`, `trainable_transformer`, `spancat` |
@@ -71,12 +23,12 @@ model-index:
71
 
72
  <details>
73
 
74
- <summary>View label scheme (56 labels for 2 components)</summary>
75
 
76
  | Component | Labels |
77
  | --- | --- |
78
  | **`parser`** | `ROOT`, `acl`, `acomp`, `advcl`, `advmod`, `agent`, `amod`, `appos`, `attr`, `aux`, `auxpass`, `case`, `cc`, `ccomp`, `compound`, `conj`, `csubj`, `csubjpass`, `dative`, `dep`, `det`, `dobj`, `expl`, `intj`, `mark`, `meta`, `neg`, `nmod`, `npadvmod`, `nsubj`, `nsubjpass`, `nummod`, `oprd`, `parataxis`, `pcomp`, `pobj`, `poss`, `preconj`, `predet`, `prep`, `prt`, `punct`, `quantmod`, `relcl`, `xcomp` |
79
- | **`spancat`** | `ENTERTAIN`, `DENY`, `ATTRIBUTION`, `COUNTER`, `MONOGLOSS`, `JUSTIFYING`, `SOURCES`, `CITATION`, `ENDOPHORIC`, `PRONOUNCE`, `CONCUR` |
80
 
81
  </details>
82
 
@@ -87,17 +39,17 @@ model-index:
87
  | `DEP_UAS` | 0.00 |
88
  | `DEP_LAS` | 0.00 |
89
  | `DEP_LAS_PER_TYPE` | 0.00 |
90
- | `SENTS_P` | 91.88 |
91
- | `SENTS_R` | 94.17 |
92
- | `SENTS_F` | 93.01 |
93
- | `SPANS_SC_F` | 69.62 |
94
- | `SPANS_SC_P` | 70.85 |
95
- | `SPANS_SC_R` | 68.44 |
96
  | `ENTS_PER_TYPE` | 0.00 |
97
  | `TAG_ACC` | 0.00 |
98
  | `ENTS_F` | 0.00 |
99
  | `ENTS_P` | 0.00 |
100
  | `ENTS_R` | 0.00 |
101
  | `LEMMA_ACC` | 0.00 |
102
- | `TRAINABLE_TRANSFORMER_LOSS` | 578.70 |
103
- | `SPANCAT_LOSS` | 46747.59 |
 
5
  - en
6
  model-index:
7
  - name: en_engagement_spl_RoBERTa_acad
8
+ results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  ---
10
  | Feature | Description |
11
  | --- | --- |
12
  | **Name** | `en_engagement_spl_RoBERTa_acad` |
13
+ | **Version** | `0.2.2.1228` |
14
  | **spaCy** | `>=3.4.4,<3.5.0` |
15
  | **Default Pipeline** | `transformer`, `parser`, `trainable_transformer`, `spancat` |
16
  | **Components** | `transformer`, `parser`, `trainable_transformer`, `spancat` |
 
23
 
24
  <details>
25
 
26
+ <summary>View label scheme (57 labels for 2 components)</summary>
27
 
28
  | Component | Labels |
29
  | --- | --- |
30
  | **`parser`** | `ROOT`, `acl`, `acomp`, `advcl`, `advmod`, `agent`, `amod`, `appos`, `attr`, `aux`, `auxpass`, `case`, `cc`, `ccomp`, `compound`, `conj`, `csubj`, `csubjpass`, `dative`, `dep`, `det`, `dobj`, `expl`, `intj`, `mark`, `meta`, `neg`, `nmod`, `npadvmod`, `nsubj`, `nsubjpass`, `nummod`, `oprd`, `parataxis`, `pcomp`, `pobj`, `poss`, `preconj`, `predet`, `prep`, `prt`, `punct`, `quantmod`, `relcl`, `xcomp` |
31
+ | **`spancat`** | `DENY`, `ENTERTAIN`, `PRONOUNCE`, `JUSTIFYING`, `MONOGLOSS`, `COUNTER`, `ENDORSE`, `ATTRIBUTE`, `ENDOPHORIC`, `SOURCES`, `CONCUR`, `CITATION` |
32
 
33
  </details>
34
 
 
39
  | `DEP_UAS` | 0.00 |
40
  | `DEP_LAS` | 0.00 |
41
  | `DEP_LAS_PER_TYPE` | 0.00 |
42
+ | `SENTS_P` | 90.27 |
43
+ | `SENTS_R` | 93.41 |
44
+ | `SENTS_F` | 91.81 |
45
+ | `SPANS_SC_F` | 69.28 |
46
+ | `SPANS_SC_P` | 72.16 |
47
+ | `SPANS_SC_R` | 66.63 |
48
  | `ENTS_PER_TYPE` | 0.00 |
49
  | `TAG_ACC` | 0.00 |
50
  | `ENTS_F` | 0.00 |
51
  | `ENTS_P` | 0.00 |
52
  | `ENTS_R` | 0.00 |
53
  | `LEMMA_ACC` | 0.00 |
54
+ | `TRAINABLE_TRANSFORMER_LOSS` | 149.48 |
55
+ | `SPANCAT_LOSS` | 106919.98 |
config.cfg CHANGED
@@ -10,8 +10,8 @@ seed = 0
10
 
11
  [nlp]
12
  lang = "en"
13
- pipeline = ["transformer","parser","trainable_transformer","spancat"]
14
- batch_size = 64
15
  disabled = []
16
  before_creation = null
17
  after_creation = null
@@ -20,28 +20,32 @@ tokenizer = {"@tokenizers":"spacy.Tokenizer.v1"}
20
 
21
  [components]
22
 
23
- [components.parser]
24
- factory = "parser"
25
- learn_tokens = false
26
- min_action_freq = 30
27
- moves = null
28
- scorer = {"@scorers":"spacy.parser_scorer.v1"}
29
- update_with_oracle_cut_size = 100
30
-
31
- [components.parser.model]
32
- @architectures = "spacy.TransitionBasedParser.v2"
33
- state_type = "parser"
34
- extra_state_tokens = false
35
- hidden_width = 64
36
- maxout_pieces = 2
37
- use_upper = false
38
- nO = null
39
 
40
- [components.parser.model.tok2vec]
41
  @architectures = "spacy-transformers.TransformerListener.v1"
42
  grad_factor = 1.0
43
- upstream = "transformer"
44
  pooling = {"@layers":"reduce_mean.v1"}
 
 
 
 
 
 
45
 
46
  [components.spancat]
47
  factory = "spancat"
@@ -54,9 +58,8 @@ threshold = 0.4
54
  @architectures = "spacy.SpanCategorizer.v1"
55
 
56
  [components.spancat.model.reducer]
57
- @layers = "mean_max_reducer.v1.5"
58
- hidden_size = 128
59
- dropout = 0.2
60
 
61
  [components.spancat.model.scorer]
62
  @layers = "spacy.LinearLogistic.v1"
@@ -67,28 +70,11 @@ nI = null
67
  @architectures = "spacy-transformers.TransformerListener.v1"
68
  grad_factor = 1.0
69
  pooling = {"@layers":"reduce_mean.v1"}
70
- upstream = "trainable_transformer"
71
 
72
  [components.spancat.suggester]
73
- @misc = "spacy-experimental.ngram_subtree_suggester.v1"
74
- sizes = [1,2,3,4,5,6,7,8,9,10,11,12]
75
-
76
- [components.trainable_transformer]
77
- factory = "transformer"
78
- max_batch_items = 4096
79
- set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
80
-
81
- [components.trainable_transformer.model]
82
- @architectures = "spacy-transformers.TransformerModel.v1"
83
- name = "egumasa/roberta-base-finetuned-academic"
84
-
85
- [components.trainable_transformer.model.get_spans]
86
- @span_getters = "spacy-transformers.strided_spans.v1"
87
- window = 196
88
- stride = 128
89
-
90
- [components.trainable_transformer.model.tokenizer_config]
91
- use_fast = true
92
 
93
  [components.transformer]
94
  factory = "transformer"
@@ -96,22 +82,17 @@ max_batch_items = 4096
96
  set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
97
 
98
  [components.transformer.model]
99
- @architectures = "spacy-transformers.TransformerModel.v3"
100
- name = "roberta-base"
101
- mixed_precision = false
102
 
103
  [components.transformer.model.get_spans]
104
  @span_getters = "spacy-transformers.strided_spans.v1"
105
- window = 128
106
- stride = 96
107
-
108
- [components.transformer.model.grad_scaler_config]
109
 
110
  [components.transformer.model.tokenizer_config]
111
  use_fast = true
112
 
113
- [components.transformer.model.transformer_config]
114
-
115
  [corpora]
116
 
117
  [corpora.dev]
@@ -136,13 +117,13 @@ train_corpus = "corpora.train"
136
  seed = ${system.seed}
137
  gpu_allocator = ${system.gpu_allocator}
138
  dropout = 0.1
139
- accumulate_gradient = 4
140
- patience = 4000
141
  max_epochs = 0
142
  max_steps = 20000
143
  eval_frequency = 200
144
- frozen_components = ["transformer","parser"]
145
- annotating_components = ["parser"]
146
  before_to_disk = null
147
 
148
  [training.batcher]
@@ -153,8 +134,8 @@ get_length = null
153
 
154
  [training.batcher.size]
155
  @schedules = "compounding.v1"
156
- start = 100
157
- stop = 400
158
  compound = 1.0002
159
  t = 0.0
160
 
@@ -175,25 +156,16 @@ eps = 0.00000001
175
  [training.optimizer.learn_rate]
176
  @schedules = "warmup_linear.v1"
177
  warmup_steps = 250
178
- total_steps = 40000
179
  initial_rate = 0.00005
180
 
181
  [training.score_weights]
182
- dep_uas = null
183
- dep_las = null
184
- dep_las_per_type = null
185
- sents_p = null
186
- sents_r = null
187
- sents_f = null
188
  spans_sc_f = 0.6
189
  spans_sc_p = 0.0
190
- spans_sc_r = 0.4
191
- ents_per_type = null
192
- tag_acc = null
193
- ents_f = null
194
- ents_p = null
195
- ents_r = null
196
- lemma_acc = null
197
 
198
  [pretraining]
199
 
 
10
 
11
  [nlp]
12
  lang = "en"
13
+ pipeline = ["transformer","span_finder","spancat"]
14
+ batch_size = 16
15
  disabled = []
16
  before_creation = null
17
  after_creation = null
 
20
 
21
  [components]
22
 
23
+ [components.span_finder]
24
+ factory = "experimental_span_finder"
25
+ max_length = 0
26
+ min_length = 0
27
+ predicted_key = "span_candidates"
28
+ threshold = 0.2
29
+ training_key = ${vars.spans_key}
30
+
31
+ [components.span_finder.model]
32
+ @architectures = "spacy-experimental.SpanFinder.v1"
33
+
34
+ [components.span_finder.model.scorer]
35
+ @layers = "spacy.LinearLogistic.v1"
36
+ nO = 2
37
+ nI = null
 
38
 
39
+ [components.span_finder.model.tok2vec]
40
  @architectures = "spacy-transformers.TransformerListener.v1"
41
  grad_factor = 1.0
 
42
  pooling = {"@layers":"reduce_mean.v1"}
43
+ upstream = "*"
44
+
45
+ [components.span_finder.scorer]
46
+ @scorers = "spacy-experimental.span_finder_scorer.v1"
47
+ predicted_key = ${components.span_finder.predicted_key}
48
+ training_key = ${vars.spans_key}
49
 
50
  [components.spancat]
51
  factory = "spancat"
 
58
  @architectures = "spacy.SpanCategorizer.v1"
59
 
60
  [components.spancat.model.reducer]
61
+ @layers = "spacy.mean_max_reducer.v1"
62
+ hidden_size = 384
 
63
 
64
  [components.spancat.model.scorer]
65
  @layers = "spacy.LinearLogistic.v1"
 
70
  @architectures = "spacy-transformers.TransformerListener.v1"
71
  grad_factor = 1.0
72
  pooling = {"@layers":"reduce_mean.v1"}
73
+ upstream = "*"
74
 
75
  [components.spancat.suggester]
76
+ @misc = "spacy-experimental.span_finder_suggester.v1"
77
+ candidates_key = ${components.span_finder.predicted_key}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
78
 
79
  [components.transformer]
80
  factory = "transformer"
 
82
  set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
83
 
84
  [components.transformer.model]
85
+ @architectures = "spacy-transformers.TransformerModel.v1"
86
+ name = "egumasa/roberta-base-finetuned-academic"
 
87
 
88
  [components.transformer.model.get_spans]
89
  @span_getters = "spacy-transformers.strided_spans.v1"
90
+ window = 169
91
+ stride = 128
 
 
92
 
93
  [components.transformer.model.tokenizer_config]
94
  use_fast = true
95
 
 
 
96
  [corpora]
97
 
98
  [corpora.dev]
 
117
  seed = ${system.seed}
118
  gpu_allocator = ${system.gpu_allocator}
119
  dropout = 0.1
120
+ accumulate_gradient = 1
121
+ patience = 5000
122
  max_epochs = 0
123
  max_steps = 20000
124
  eval_frequency = 200
125
+ frozen_components = []
126
+ annotating_components = ["span_finder"]
127
  before_to_disk = null
128
 
129
  [training.batcher]
 
134
 
135
  [training.batcher.size]
136
  @schedules = "compounding.v1"
137
+ start = 300
138
+ stop = 1000
139
  compound = 1.0002
140
  t = 0.0
141
 
 
156
  [training.optimizer.learn_rate]
157
  @schedules = "warmup_linear.v1"
158
  warmup_steps = 250
159
+ total_steps = 20000
160
  initial_rate = 0.00005
161
 
162
  [training.score_weights]
163
+ span_finder_span_candidates_f = 0.0
164
+ span_finder_span_candidates_p = 0.0
165
+ span_finder_span_candidates_r = 0.2
 
 
 
166
  spans_sc_f = 0.6
167
  spans_sc_p = 0.0
168
+ spans_sc_r = 0.2
 
 
 
 
 
 
169
 
170
  [pretraining]
171
 
en_engagement_spl_RoBERTa_acad-any-py3-none-any.whl CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3c06d780ee3f8a520e30c29297b31d566b4605e1e9070b21ec960b688f2859d9
3
- size 928165493
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9060a8e1c2d993e693cb362144f58ea88767481933403c14a0f6f5ca30cd1af
3
+ size 942481184
meta.json CHANGED
@@ -1,14 +1,14 @@
1
  {
2
  "lang":"en",
3
  "name":"engagement_spl_RoBERTa_acad",
4
- "version":"0.3.4.1221",
5
  "description":"",
6
  "author":"",
7
  "email":"",
8
  "url":"",
9
  "license":"",
10
  "spacy_version":">=3.4.4,<3.5.0",
11
- "spacy_git_version":"Unknown",
12
  "vectors":{
13
  "width":0,
14
  "vectors":0,
@@ -19,106 +19,50 @@
19
  "transformer":[
20
 
21
  ],
22
- "parser":[
23
- "ROOT",
24
- "acl",
25
- "acomp",
26
- "advcl",
27
- "advmod",
28
- "agent",
29
- "amod",
30
- "appos",
31
- "attr",
32
- "aux",
33
- "auxpass",
34
- "case",
35
- "cc",
36
- "ccomp",
37
- "compound",
38
- "conj",
39
- "csubj",
40
- "csubjpass",
41
- "dative",
42
- "dep",
43
- "det",
44
- "dobj",
45
- "expl",
46
- "intj",
47
- "mark",
48
- "meta",
49
- "neg",
50
- "nmod",
51
- "npadvmod",
52
- "nsubj",
53
- "nsubjpass",
54
- "nummod",
55
- "oprd",
56
- "parataxis",
57
- "pcomp",
58
- "pobj",
59
- "poss",
60
- "preconj",
61
- "predet",
62
- "prep",
63
- "prt",
64
- "punct",
65
- "quantmod",
66
- "relcl",
67
- "xcomp"
68
- ],
69
- "trainable_transformer":[
70
 
71
  ],
72
  "spancat":[
73
- "ENTERTAIN",
74
  "DENY",
75
- "ATTRIBUTION",
76
- "COUNTER",
77
- "MONOGLOSS",
78
  "JUSTIFYING",
79
- "SOURCES",
80
- "CITATION",
 
 
81
  "ENDOPHORIC",
82
- "PRONOUNCE",
83
- "CONCUR"
 
84
  ]
85
  },
86
  "pipeline":[
87
  "transformer",
88
- "parser",
89
- "trainable_transformer",
90
  "spancat"
91
  ],
92
  "components":[
93
  "transformer",
94
- "parser",
95
- "trainable_transformer",
96
  "spancat"
97
  ],
98
  "disabled":[
99
 
100
  ],
101
  "performance":{
102
- "dep_uas":0.0,
103
- "dep_las":0.0,
104
- "dep_las_per_type":0.0,
105
- "sents_p":0.9188370684,
106
- "sents_r":0.9416511484,
107
- "sents_f":0.9301042305,
108
- "spans_sc_f":0.696227149,
109
- "spans_sc_p":0.7084900059,
110
- "spans_sc_r":0.6843815714,
111
- "ents_per_type":0.0,
112
- "tag_acc":0.0,
113
- "ents_f":0.0,
114
- "ents_p":0.0,
115
- "ents_r":0.0,
116
- "lemma_acc":0.0,
117
- "trainable_transformer_loss":5.7869890926,
118
- "spancat_loss":467.4759056543
119
  },
120
  "requirements":[
121
  "spacy-experimental>=0.6.1,<0.7.0",
122
- "spacy-transformers>=1.1.7,<1.2.0"
123
  ]
124
  }
 
1
  {
2
  "lang":"en",
3
  "name":"engagement_spl_RoBERTa_acad",
4
+ "version":"0.2.2.1228",
5
  "description":"",
6
  "author":"",
7
  "email":"",
8
  "url":"",
9
  "license":"",
10
  "spacy_version":">=3.4.4,<3.5.0",
11
+ "spacy_git_version":"77833bfef",
12
  "vectors":{
13
  "width":0,
14
  "vectors":0,
 
19
  "transformer":[
20
 
21
  ],
22
+ "span_finder":[
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
 
24
  ],
25
  "spancat":[
 
26
  "DENY",
27
+ "ENTERTAIN",
28
+ "PRONOUNCE",
 
29
  "JUSTIFYING",
30
+ "MONOGLOSS",
31
+ "COUNTER",
32
+ "ENDORSE",
33
+ "ATTRIBUTE",
34
  "ENDOPHORIC",
35
+ "SOURCES",
36
+ "CONCUR",
37
+ "CITATION"
38
  ]
39
  },
40
  "pipeline":[
41
  "transformer",
42
+ "span_finder",
 
43
  "spancat"
44
  ],
45
  "components":[
46
  "transformer",
47
+ "span_finder",
 
48
  "spancat"
49
  ],
50
  "disabled":[
51
 
52
  ],
53
  "performance":{
54
+ "span_finder_span_candidates_f":0.199127869,
55
+ "span_finder_span_candidates_p":0.1139467472,
56
+ "span_finder_span_candidates_r":0.7887877534,
57
+ "spans_sc_f":0.7057724252,
58
+ "spans_sc_p":0.7084201751,
59
+ "spans_sc_r":0.7031443939,
60
+ "transformer_loss":1457529913378.0666503906,
61
+ "span_finder_loss":622.8571459684,
62
+ "spancat_loss":513.4672946446
 
 
 
 
 
 
 
 
63
  },
64
  "requirements":[
65
  "spacy-experimental>=0.6.1,<0.7.0",
66
+ "spacy-transformers>=1.1.8,<1.2.0"
67
  ]
68
  }
span_finder/model ADDED
Binary file (6.9 kB). View file
 
spancat/cfg CHANGED
@@ -1,16 +1,17 @@
1
  {
2
  "labels":[
3
- "ENTERTAIN",
4
  "DENY",
5
- "ATTRIBUTION",
6
- "COUNTER",
7
- "MONOGLOSS",
8
  "JUSTIFYING",
9
- "SOURCES",
10
- "CITATION",
 
 
11
  "ENDOPHORIC",
12
- "PRONOUNCE",
13
- "CONCUR"
 
14
  ],
15
  "spans_key":"sc",
16
  "threshold":0.4,
 
1
  {
2
  "labels":[
 
3
  "DENY",
4
+ "ENTERTAIN",
5
+ "PRONOUNCE",
 
6
  "JUSTIFYING",
7
+ "MONOGLOSS",
8
+ "COUNTER",
9
+ "ENDORSE",
10
+ "ATTRIBUTE",
11
  "ENDOPHORIC",
12
+ "SOURCES",
13
+ "CONCUR",
14
+ "CITATION"
15
  ],
16
  "spans_key":"sc",
17
  "threshold":0.4,
spancat/model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:abc588761e531aca0237766989d1299a6d57be02c168b5a8680c049414414a3e
3
- size 4728651
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64ac1f8c1b52aadba9236caab8a5438c2cb72abf47d95248029664bb15eadd5d
3
+ size 14183772
trainable_transformer/model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:85445ef858c9db826ec3df16d2c2b71d4732468580514c8cd91898ed9c99731a
3
- size 502030706
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d01afabc36b8071df731d74137b8364e09c4fa6c42d0de52b0bcbd84a669581
3
+ size 502027906
transformer/model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2a380b706181d8b3b0a4a7e601a680fdd30a241fcb0a9304bfbdd82816dc44fc
3
- size 502030652
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a76c5f803a0dc9d96af11cce2c5e9d93194c6c2b0df2bb31fc1e9c7401bdef08
3
+ size 502027906
vocab/strings.json CHANGED
The diff for this file is too large to render. See raw diff