egumasa commited on
Commit
77c7dfe
·
1 Parent(s): 80bcf22

Update spaCy pipeline

Browse files
README.md CHANGED
@@ -5,12 +5,60 @@ language:
5
  - en
6
  model-index:
7
  - name: en_engagement_spl_RoBERTa_acad
8
- results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  ---
10
  | Feature | Description |
11
  | --- | --- |
12
  | **Name** | `en_engagement_spl_RoBERTa_acad` |
13
- | **Version** | `0.2.2.1228` |
14
  | **spaCy** | `>=3.4.4,<3.5.0` |
15
  | **Default Pipeline** | `transformer`, `parser`, `trainable_transformer`, `spancat` |
16
  | **Components** | `transformer`, `parser`, `trainable_transformer`, `spancat` |
@@ -42,14 +90,14 @@ model-index:
42
  | `SENTS_P` | 90.27 |
43
  | `SENTS_R` | 93.41 |
44
  | `SENTS_F` | 91.81 |
45
- | `SPANS_SC_F` | 69.28 |
46
- | `SPANS_SC_P` | 72.16 |
47
- | `SPANS_SC_R` | 66.63 |
48
  | `ENTS_PER_TYPE` | 0.00 |
49
  | `TAG_ACC` | 0.00 |
50
  | `ENTS_F` | 0.00 |
51
  | `ENTS_P` | 0.00 |
52
  | `ENTS_R` | 0.00 |
53
  | `LEMMA_ACC` | 0.00 |
54
- | `TRAINABLE_TRANSFORMER_LOSS` | 149.48 |
55
- | `SPANCAT_LOSS` | 106919.98 |
 
5
  - en
6
  model-index:
7
  - name: en_engagement_spl_RoBERTa_acad
8
+ results:
9
+ - task:
10
+ name: NER
11
+ type: token-classification
12
+ metrics:
13
+ - name: NER Precision
14
+ type: precision
15
+ value: 0.0
16
+ - name: NER Recall
17
+ type: recall
18
+ value: 0.0
19
+ - name: NER F Score
20
+ type: f_score
21
+ value: 0.0
22
+ - task:
23
+ name: TAG
24
+ type: token-classification
25
+ metrics:
26
+ - name: TAG (XPOS) Accuracy
27
+ type: accuracy
28
+ value: 0.0
29
+ - task:
30
+ name: LEMMA
31
+ type: token-classification
32
+ metrics:
33
+ - name: Lemma Accuracy
34
+ type: accuracy
35
+ value: 0.0
36
+ - task:
37
+ name: UNLABELED_DEPENDENCIES
38
+ type: token-classification
39
+ metrics:
40
+ - name: Unlabeled Attachment Score (UAS)
41
+ type: f_score
42
+ value: 0.0
43
+ - task:
44
+ name: LABELED_DEPENDENCIES
45
+ type: token-classification
46
+ metrics:
47
+ - name: Labeled Attachment Score (LAS)
48
+ type: f_score
49
+ value: 0.0
50
+ - task:
51
+ name: SENTS
52
+ type: token-classification
53
+ metrics:
54
+ - name: Sentences F-Score
55
+ type: f_score
56
+ value: 0.9181469871
57
  ---
58
  | Feature | Description |
59
  | --- | --- |
60
  | **Name** | `en_engagement_spl_RoBERTa_acad` |
61
+ | **Version** | `0.3.0.1228` |
62
  | **spaCy** | `>=3.4.4,<3.5.0` |
63
  | **Default Pipeline** | `transformer`, `parser`, `trainable_transformer`, `spancat` |
64
  | **Components** | `transformer`, `parser`, `trainable_transformer`, `spancat` |
 
90
  | `SENTS_P` | 90.27 |
91
  | `SENTS_R` | 93.41 |
92
  | `SENTS_F` | 91.81 |
93
+ | `SPANS_SC_F` | 69.58 |
94
+ | `SPANS_SC_P` | 72.23 |
95
+ | `SPANS_SC_R` | 67.11 |
96
  | `ENTS_PER_TYPE` | 0.00 |
97
  | `TAG_ACC` | 0.00 |
98
  | `ENTS_F` | 0.00 |
99
  | `ENTS_P` | 0.00 |
100
  | `ENTS_R` | 0.00 |
101
  | `LEMMA_ACC` | 0.00 |
102
+ | `TRAINABLE_TRANSFORMER_LOSS` | 344.32 |
103
+ | `SPANCAT_LOSS` | 51942.04 |
config.cfg CHANGED
@@ -10,8 +10,8 @@ seed = 0
10
 
11
  [nlp]
12
  lang = "en"
13
- pipeline = ["transformer","span_finder","spancat"]
14
- batch_size = 16
15
  disabled = []
16
  before_creation = null
17
  after_creation = null
@@ -20,32 +20,28 @@ tokenizer = {"@tokenizers":"spacy.Tokenizer.v1"}
20
 
21
  [components]
22
 
23
- [components.span_finder]
24
- factory = "experimental_span_finder"
25
- max_length = 0
26
- min_length = 0
27
- predicted_key = "span_candidates"
28
- threshold = 0.2
29
- training_key = ${vars.spans_key}
30
-
31
- [components.span_finder.model]
32
- @architectures = "spacy-experimental.SpanFinder.v1"
33
-
34
- [components.span_finder.model.scorer]
35
- @layers = "spacy.LinearLogistic.v1"
36
- nO = 2
37
- nI = null
 
38
 
39
- [components.span_finder.model.tok2vec]
40
  @architectures = "spacy-transformers.TransformerListener.v1"
41
  grad_factor = 1.0
 
42
  pooling = {"@layers":"reduce_mean.v1"}
43
- upstream = "*"
44
-
45
- [components.span_finder.scorer]
46
- @scorers = "spacy-experimental.span_finder_scorer.v1"
47
- predicted_key = ${components.span_finder.predicted_key}
48
- training_key = ${vars.spans_key}
49
 
50
  [components.spancat]
51
  factory = "spancat"
@@ -58,8 +54,9 @@ threshold = 0.4
58
  @architectures = "spacy.SpanCategorizer.v1"
59
 
60
  [components.spancat.model.reducer]
61
- @layers = "spacy.mean_max_reducer.v1"
62
- hidden_size = 384
 
63
 
64
  [components.spancat.model.scorer]
65
  @layers = "spacy.LinearLogistic.v1"
@@ -70,29 +67,51 @@ nI = null
70
  @architectures = "spacy-transformers.TransformerListener.v1"
71
  grad_factor = 1.0
72
  pooling = {"@layers":"reduce_mean.v1"}
73
- upstream = "*"
74
 
75
  [components.spancat.suggester]
76
- @misc = "spacy-experimental.span_finder_suggester.v1"
77
- candidates_key = ${components.span_finder.predicted_key}
78
 
79
- [components.transformer]
80
  factory = "transformer"
81
  max_batch_items = 4096
82
  set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
83
 
84
- [components.transformer.model]
85
  @architectures = "spacy-transformers.TransformerModel.v1"
86
  name = "egumasa/roberta-base-finetuned-academic"
87
 
88
- [components.transformer.model.get_spans]
89
  @span_getters = "spacy-transformers.strided_spans.v1"
90
- window = 169
91
  stride = 128
92
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
93
  [components.transformer.model.tokenizer_config]
94
  use_fast = true
95
 
 
 
96
  [corpora]
97
 
98
  [corpora.dev]
@@ -117,13 +136,13 @@ train_corpus = "corpora.train"
117
  seed = ${system.seed}
118
  gpu_allocator = ${system.gpu_allocator}
119
  dropout = 0.1
120
- accumulate_gradient = 1
121
- patience = 5000
122
  max_epochs = 0
123
  max_steps = 20000
124
  eval_frequency = 200
125
- frozen_components = []
126
- annotating_components = ["span_finder"]
127
  before_to_disk = null
128
 
129
  [training.batcher]
@@ -134,8 +153,8 @@ get_length = null
134
 
135
  [training.batcher.size]
136
  @schedules = "compounding.v1"
137
- start = 300
138
- stop = 1000
139
  compound = 1.0002
140
  t = 0.0
141
 
@@ -156,16 +175,25 @@ eps = 0.00000001
156
  [training.optimizer.learn_rate]
157
  @schedules = "warmup_linear.v1"
158
  warmup_steps = 250
159
- total_steps = 20000
160
  initial_rate = 0.00005
161
 
162
  [training.score_weights]
163
- span_finder_span_candidates_f = 0.0
164
- span_finder_span_candidates_p = 0.0
165
- span_finder_span_candidates_r = 0.2
166
- spans_sc_f = 0.6
 
 
 
167
  spans_sc_p = 0.0
168
- spans_sc_r = 0.2
 
 
 
 
 
 
169
 
170
  [pretraining]
171
 
 
10
 
11
  [nlp]
12
  lang = "en"
13
+ pipeline = ["transformer","parser","trainable_transformer","spancat"]
14
+ batch_size = 64
15
  disabled = []
16
  before_creation = null
17
  after_creation = null
 
20
 
21
  [components]
22
 
23
+ [components.parser]
24
+ factory = "parser"
25
+ learn_tokens = false
26
+ min_action_freq = 30
27
+ moves = null
28
+ scorer = {"@scorers":"spacy.parser_scorer.v1"}
29
+ update_with_oracle_cut_size = 100
30
+
31
+ [components.parser.model]
32
+ @architectures = "spacy.TransitionBasedParser.v2"
33
+ state_type = "parser"
34
+ extra_state_tokens = false
35
+ hidden_width = 64
36
+ maxout_pieces = 2
37
+ use_upper = false
38
+ nO = null
39
 
40
+ [components.parser.model.tok2vec]
41
  @architectures = "spacy-transformers.TransformerListener.v1"
42
  grad_factor = 1.0
43
+ upstream = "transformer"
44
  pooling = {"@layers":"reduce_mean.v1"}
 
 
 
 
 
 
45
 
46
  [components.spancat]
47
  factory = "spancat"
 
54
  @architectures = "spacy.SpanCategorizer.v1"
55
 
56
  [components.spancat.model.reducer]
57
+ @layers = "mean_max_reducer.v1.5"
58
+ hidden_size = 256
59
+ dropout = 0.2
60
 
61
  [components.spancat.model.scorer]
62
  @layers = "spacy.LinearLogistic.v1"
 
67
  @architectures = "spacy-transformers.TransformerListener.v1"
68
  grad_factor = 1.0
69
  pooling = {"@layers":"reduce_mean.v1"}
70
+ upstream = "trainable_transformer"
71
 
72
  [components.spancat.suggester]
73
+ @misc = "spacy-experimental.ngram_subtree_suggester.v1"
74
+ sizes = [1,2,3,4,5,6,7,8,9,10,11,12]
75
 
76
+ [components.trainable_transformer]
77
  factory = "transformer"
78
  max_batch_items = 4096
79
  set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
80
 
81
+ [components.trainable_transformer.model]
82
  @architectures = "spacy-transformers.TransformerModel.v1"
83
  name = "egumasa/roberta-base-finetuned-academic"
84
 
85
+ [components.trainable_transformer.model.get_spans]
86
  @span_getters = "spacy-transformers.strided_spans.v1"
87
+ window = 196
88
  stride = 128
89
 
90
+ [components.trainable_transformer.model.tokenizer_config]
91
+ use_fast = true
92
+
93
+ [components.transformer]
94
+ factory = "transformer"
95
+ max_batch_items = 4096
96
+ set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
97
+
98
+ [components.transformer.model]
99
+ @architectures = "spacy-transformers.TransformerModel.v3"
100
+ name = "roberta-base"
101
+ mixed_precision = false
102
+
103
+ [components.transformer.model.get_spans]
104
+ @span_getters = "spacy-transformers.strided_spans.v1"
105
+ window = 128
106
+ stride = 96
107
+
108
+ [components.transformer.model.grad_scaler_config]
109
+
110
  [components.transformer.model.tokenizer_config]
111
  use_fast = true
112
 
113
+ [components.transformer.model.transformer_config]
114
+
115
  [corpora]
116
 
117
  [corpora.dev]
 
136
  seed = ${system.seed}
137
  gpu_allocator = ${system.gpu_allocator}
138
  dropout = 0.1
139
+ accumulate_gradient = 4
140
+ patience = 6000
141
  max_epochs = 0
142
  max_steps = 20000
143
  eval_frequency = 200
144
+ frozen_components = ["transformer","parser"]
145
+ annotating_components = ["parser"]
146
  before_to_disk = null
147
 
148
  [training.batcher]
 
153
 
154
  [training.batcher.size]
155
  @schedules = "compounding.v1"
156
+ start = 100
157
+ stop = 600
158
  compound = 1.0002
159
  t = 0.0
160
 
 
175
  [training.optimizer.learn_rate]
176
  @schedules = "warmup_linear.v1"
177
  warmup_steps = 250
178
+ total_steps = 40000
179
  initial_rate = 0.00005
180
 
181
  [training.score_weights]
182
+ dep_uas = null
183
+ dep_las = null
184
+ dep_las_per_type = null
185
+ sents_p = null
186
+ sents_r = null
187
+ sents_f = null
188
+ spans_sc_f = 0.5
189
  spans_sc_p = 0.0
190
+ spans_sc_r = 0.5
191
+ ents_per_type = null
192
+ tag_acc = null
193
+ ents_f = null
194
+ ents_p = null
195
+ ents_r = null
196
+ lemma_acc = null
197
 
198
  [pretraining]
199
 
custom_functions.py CHANGED
@@ -59,7 +59,8 @@ def build_mean_max_reducer2(hidden_size: int,
59
  cast(Model[Ragged, Floats2d], reduce_first()),
60
  reduce_mean(),
61
  reduce_max(),
62
- ), Maxout(nO=hidden_size, normalize=True, dropout=dropout),
 
63
  Maxout(nO=hidden_size, normalize=True, dropout=dropout))
64
 
65
 
 
59
  cast(Model[Ragged, Floats2d], reduce_first()),
60
  reduce_mean(),
61
  reduce_max(),
62
+ ),
63
+ Maxout(nO=hidden_size, normalize=True, dropout=dropout),
64
  Maxout(nO=hidden_size, normalize=True, dropout=dropout))
65
 
66
 
en_engagement_spl_RoBERTa_acad-any-py3-none-any.whl CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f9060a8e1c2d993e693cb362144f58ea88767481933403c14a0f6f5ca30cd1af
3
- size 942481184
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2772317b3e8261f1ca495961af24180ca0910be23cea34ca00020666836a3fee
3
+ size 932543299
meta.json CHANGED
@@ -1,14 +1,14 @@
1
  {
2
  "lang":"en",
3
  "name":"engagement_spl_RoBERTa_acad",
4
- "version":"0.2.2.1228",
5
  "description":"",
6
  "author":"",
7
  "email":"",
8
  "url":"",
9
  "license":"",
10
  "spacy_version":">=3.4.4,<3.5.0",
11
- "spacy_git_version":"77833bfef",
12
  "vectors":{
13
  "width":0,
14
  "vectors":0,
@@ -19,7 +19,54 @@
19
  "transformer":[
20
 
21
  ],
22
- "span_finder":[
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
 
24
  ],
25
  "spancat":[
@@ -39,30 +86,40 @@
39
  },
40
  "pipeline":[
41
  "transformer",
42
- "span_finder",
 
43
  "spancat"
44
  ],
45
  "components":[
46
  "transformer",
47
- "span_finder",
 
48
  "spancat"
49
  ],
50
  "disabled":[
51
 
52
  ],
53
  "performance":{
54
- "span_finder_span_candidates_f":0.199127869,
55
- "span_finder_span_candidates_p":0.1139467472,
56
- "span_finder_span_candidates_r":0.7887877534,
57
- "spans_sc_f":0.7057724252,
58
- "spans_sc_p":0.7084201751,
59
- "spans_sc_r":0.7031443939,
60
- "transformer_loss":1457529913378.0666503906,
61
- "span_finder_loss":622.8571459684,
62
- "spancat_loss":513.4672946446
 
 
 
 
 
 
 
 
63
  },
64
  "requirements":[
65
- "spacy-experimental>=0.6.1,<0.7.0",
66
- "spacy-transformers>=1.1.8,<1.2.0"
67
  ]
68
  }
 
1
  {
2
  "lang":"en",
3
  "name":"engagement_spl_RoBERTa_acad",
4
+ "version":"0.3.0.1228",
5
  "description":"",
6
  "author":"",
7
  "email":"",
8
  "url":"",
9
  "license":"",
10
  "spacy_version":">=3.4.4,<3.5.0",
11
+ "spacy_git_version":"Unknown",
12
  "vectors":{
13
  "width":0,
14
  "vectors":0,
 
19
  "transformer":[
20
 
21
  ],
22
+ "parser":[
23
+ "ROOT",
24
+ "acl",
25
+ "acomp",
26
+ "advcl",
27
+ "advmod",
28
+ "agent",
29
+ "amod",
30
+ "appos",
31
+ "attr",
32
+ "aux",
33
+ "auxpass",
34
+ "case",
35
+ "cc",
36
+ "ccomp",
37
+ "compound",
38
+ "conj",
39
+ "csubj",
40
+ "csubjpass",
41
+ "dative",
42
+ "dep",
43
+ "det",
44
+ "dobj",
45
+ "expl",
46
+ "intj",
47
+ "mark",
48
+ "meta",
49
+ "neg",
50
+ "nmod",
51
+ "npadvmod",
52
+ "nsubj",
53
+ "nsubjpass",
54
+ "nummod",
55
+ "oprd",
56
+ "parataxis",
57
+ "pcomp",
58
+ "pobj",
59
+ "poss",
60
+ "preconj",
61
+ "predet",
62
+ "prep",
63
+ "prt",
64
+ "punct",
65
+ "quantmod",
66
+ "relcl",
67
+ "xcomp"
68
+ ],
69
+ "trainable_transformer":[
70
 
71
  ],
72
  "spancat":[
 
86
  },
87
  "pipeline":[
88
  "transformer",
89
+ "parser",
90
+ "trainable_transformer",
91
  "spancat"
92
  ],
93
  "components":[
94
  "transformer",
95
+ "parser",
96
+ "trainable_transformer",
97
  "spancat"
98
  ],
99
  "disabled":[
100
 
101
  ],
102
  "performance":{
103
+ "dep_uas":0.0,
104
+ "dep_las":0.0,
105
+ "dep_las_per_type":0.0,
106
+ "sents_p":0.902739726,
107
+ "sents_r":0.9340892984,
108
+ "sents_f":0.9181469871,
109
+ "spans_sc_f":0.6957640751,
110
+ "spans_sc_p":0.722333556,
111
+ "spans_sc_r":0.6710798511,
112
+ "ents_per_type":0.0,
113
+ "tag_acc":0.0,
114
+ "ents_f":0.0,
115
+ "ents_p":0.0,
116
+ "ents_r":0.0,
117
+ "lemma_acc":0.0,
118
+ "trainable_transformer_loss":3.4431899822,
119
+ "spancat_loss":519.4204027318
120
  },
121
  "requirements":[
122
+ "spacy-transformers>=1.1.8,<1.2.0",
123
+ "spacy-experimental>=0.6.1,<0.7.0"
124
  ]
125
  }
spancat/model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:64ac1f8c1b52aadba9236caab8a5438c2cb72abf47d95248029664bb15eadd5d
3
- size 14183772
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6760360b1c0cbf336fb1d473bd0db2ea6562f91f9bd42b0a41140ba1ad0c129
3
+ size 9456476
trainable_transformer/model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d01afabc36b8071df731d74137b8364e09c4fa6c42d0de52b0bcbd84a669581
3
- size 502027906
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa52feb953fba36c73a8277246dd4235d80239fefa4fb4c04b6891db4cd77db0
3
+ size 502030722
transformer/model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a76c5f803a0dc9d96af11cce2c5e9d93194c6c2b0df2bb31fc1e9c7401bdef08
3
- size 502027906
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f3bd9c07fd7e10e50c3eea3a9f90a08f4f1d12f0ad4839b6949b3e9d3d91662
3
+ size 502030652
vocab/strings.json CHANGED
The diff for this file is too large to render. See raw diff