SupstarZh commited on
Commit
00fc045
·
1 Parent(s): 89f344d

Upload 13 files

Browse files
unsup-whitenedcse-bert-base-uncased/config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-uncased",
3
+ "architectures": [
4
+ "BertForCL"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "attention_probs_dropout_prob_noise": 0.1,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_dropout_prob_noise": 0.1,
12
+ "hidden_dropout_prob_noise2": 0.2,
13
+ "hidden_size": 768,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 3072,
16
+ "layer_norm_eps": 1e-12,
17
+ "max_position_embeddings": 512,
18
+ "model_type": "bert",
19
+ "multi_dropout": false,
20
+ "num_attention_heads": 12,
21
+ "num_hidden_layers": 12,
22
+ "pad_token_id": 0,
23
+ "position_embedding_type": "absolute",
24
+ "transformers_version": "4.2.1",
25
+ "type_vocab_size": 2,
26
+ "use_cache": true,
27
+ "vocab_size": 30522
28
+ }
unsup-whitenedcse-bert-base-uncased/eval_results.txt ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch = 1.0
2
+ eval_CR = 88.47
3
+ eval_MPQA = 88.56
4
+ eval_MR = 83.73
5
+ eval_MRPC = 74.34
6
+ eval_SST2 = 88.42
7
+ eval_SUBJ = 94.47
8
+ eval_TREC = 80.36
9
+ eval_avg_sts = 0.7561232564078078
10
+ eval_avg_transfer = 85.47857142857143
11
+ eval_sickr_spearman = 0.7206682316875482
12
+ eval_stsb_spearman = 0.7915782811280672
unsup-whitenedcse-bert-base-uncased/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
unsup-whitenedcse-bert-base-uncased/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b4fd94a826bbad7e019e3973538b5725a58ae316429e5d76e56385ef5cb9b4b
3
+ size 875973285
unsup-whitenedcse-bert-base-uncased/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0185e4dbb506a8307a9da59d0bef90f31e3e929afc2d9ebd3daf1281fa14ac46
3
+ size 438012465
unsup-whitenedcse-bert-base-uncased/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e44519aa377dfc697372b150b09cfa3048148d6bf5a0ea461a9012cf304d40e4
3
+ size 623
unsup-whitenedcse-bert-base-uncased/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
unsup-whitenedcse-bert-base-uncased/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "name_or_path": "bert-base-uncased"}
unsup-whitenedcse-bert-base-uncased/train_results.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ epoch = 1.0
2
+ train_runtime = 7194.6299
3
+ train_samples_per_second = 1.086
unsup-whitenedcse-bert-base-uncased/trainer_state.json ADDED
@@ -0,0 +1,241 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.8064888341848981,
3
+ "best_model_checkpoint": "result/my-unsup-simcse-bert-base-uncased",
4
+ "epoch": 0.8638341438443818,
5
+ "global_step": 3375,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.03,
12
+ "eval_avg_sts": 0.7035617984039024,
13
+ "eval_sickr_spearman": 0.6848745407030954,
14
+ "eval_stsb_spearman": 0.7222490561047095,
15
+ "step": 125
16
+ },
17
+ {
18
+ "epoch": 0.06,
19
+ "eval_avg_sts": 0.7266642057578656,
20
+ "eval_sickr_spearman": 0.70274752572599,
21
+ "eval_stsb_spearman": 0.7505808857897414,
22
+ "step": 250
23
+ },
24
+ {
25
+ "epoch": 0.1,
26
+ "eval_avg_sts": 0.7461885709126848,
27
+ "eval_sickr_spearman": 0.7217250002652101,
28
+ "eval_stsb_spearman": 0.7706521415601595,
29
+ "step": 375
30
+ },
31
+ {
32
+ "epoch": 0.13,
33
+ "learning_rate": 2.6160737138469415e-05,
34
+ "loss": 0.0013,
35
+ "step": 500
36
+ },
37
+ {
38
+ "epoch": 0.13,
39
+ "eval_avg_sts": 0.7463389976451559,
40
+ "eval_sickr_spearman": 0.7199349731494692,
41
+ "eval_stsb_spearman": 0.7727430221408426,
42
+ "step": 500
43
+ },
44
+ {
45
+ "epoch": 0.16,
46
+ "eval_avg_sts": 0.743938555014451,
47
+ "eval_sickr_spearman": 0.7180073918502566,
48
+ "eval_stsb_spearman": 0.7698697181786454,
49
+ "step": 625
50
+ },
51
+ {
52
+ "epoch": 0.19,
53
+ "eval_avg_sts": 0.7550432481097468,
54
+ "eval_sickr_spearman": 0.7234769647664954,
55
+ "eval_stsb_spearman": 0.7866095314529982,
56
+ "step": 750
57
+ },
58
+ {
59
+ "epoch": 0.22,
60
+ "eval_avg_sts": 0.7565707728294095,
61
+ "eval_sickr_spearman": 0.7258336460737318,
62
+ "eval_stsb_spearman": 0.7873078995850874,
63
+ "step": 875
64
+ },
65
+ {
66
+ "epoch": 0.26,
67
+ "learning_rate": 2.232147427693883e-05,
68
+ "loss": 0.0004,
69
+ "step": 1000
70
+ },
71
+ {
72
+ "epoch": 0.26,
73
+ "eval_avg_sts": 0.7492950228664624,
74
+ "eval_sickr_spearman": 0.7233381976577096,
75
+ "eval_stsb_spearman": 0.7752518480752153,
76
+ "step": 1000
77
+ },
78
+ {
79
+ "epoch": 0.29,
80
+ "eval_avg_sts": 0.7522516406505854,
81
+ "eval_sickr_spearman": 0.7287197031901267,
82
+ "eval_stsb_spearman": 0.775783578111044,
83
+ "step": 1125
84
+ },
85
+ {
86
+ "epoch": 0.32,
87
+ "eval_avg_sts": 0.7552964382615647,
88
+ "eval_sickr_spearman": 0.729468085275092,
89
+ "eval_stsb_spearman": 0.7811247912480374,
90
+ "step": 1250
91
+ },
92
+ {
93
+ "epoch": 0.35,
94
+ "eval_avg_sts": 0.7568556891093262,
95
+ "eval_sickr_spearman": 0.7272718476322126,
96
+ "eval_stsb_spearman": 0.7864395305864398,
97
+ "step": 1375
98
+ },
99
+ {
100
+ "epoch": 0.38,
101
+ "learning_rate": 1.8482211415408245e-05,
102
+ "loss": 0.0004,
103
+ "step": 1500
104
+ },
105
+ {
106
+ "epoch": 0.38,
107
+ "eval_avg_sts": 0.7657081366190621,
108
+ "eval_sickr_spearman": 0.7298737076524129,
109
+ "eval_stsb_spearman": 0.8015425655857115,
110
+ "step": 1500
111
+ },
112
+ {
113
+ "epoch": 0.42,
114
+ "eval_avg_sts": 0.7645138872097019,
115
+ "eval_sickr_spearman": 0.7293550763436135,
116
+ "eval_stsb_spearman": 0.7996726980757902,
117
+ "step": 1625
118
+ },
119
+ {
120
+ "epoch": 0.45,
121
+ "eval_avg_sts": 0.7604696897809164,
122
+ "eval_sickr_spearman": 0.7309107216292885,
123
+ "eval_stsb_spearman": 0.7900286579325445,
124
+ "step": 1750
125
+ },
126
+ {
127
+ "epoch": 0.48,
128
+ "eval_avg_sts": 0.7665958188961237,
129
+ "eval_sickr_spearman": 0.7308310370977922,
130
+ "eval_stsb_spearman": 0.8023606006944553,
131
+ "step": 1875
132
+ },
133
+ {
134
+ "epoch": 0.51,
135
+ "learning_rate": 1.4642948553877656e-05,
136
+ "loss": 0.0004,
137
+ "step": 2000
138
+ },
139
+ {
140
+ "epoch": 0.51,
141
+ "eval_avg_sts": 0.7650091226879439,
142
+ "eval_sickr_spearman": 0.729455872126414,
143
+ "eval_stsb_spearman": 0.8005623732494735,
144
+ "step": 2000
145
+ },
146
+ {
147
+ "epoch": 0.54,
148
+ "eval_avg_sts": 0.7652971356117957,
149
+ "eval_sickr_spearman": 0.7297830140299209,
150
+ "eval_stsb_spearman": 0.8008112571936706,
151
+ "step": 2125
152
+ },
153
+ {
154
+ "epoch": 0.58,
155
+ "eval_avg_sts": 0.7629965694314595,
156
+ "eval_sickr_spearman": 0.7272991579004704,
157
+ "eval_stsb_spearman": 0.7986939809624487,
158
+ "step": 2250
159
+ },
160
+ {
161
+ "epoch": 0.61,
162
+ "eval_avg_sts": 0.7639744709473972,
163
+ "eval_sickr_spearman": 0.7219387730993263,
164
+ "eval_stsb_spearman": 0.8060101687954682,
165
+ "step": 2375
166
+ },
167
+ {
168
+ "epoch": 0.64,
169
+ "learning_rate": 1.080368569234707e-05,
170
+ "loss": 0.0003,
171
+ "step": 2500
172
+ },
173
+ {
174
+ "epoch": 0.64,
175
+ "eval_avg_sts": 0.7649654793413658,
176
+ "eval_sickr_spearman": 0.7252475539569496,
177
+ "eval_stsb_spearman": 0.8046834047257821,
178
+ "step": 2500
179
+ },
180
+ {
181
+ "epoch": 0.67,
182
+ "eval_avg_sts": 0.7645559824476773,
183
+ "eval_sickr_spearman": 0.7272004062812213,
184
+ "eval_stsb_spearman": 0.8019115586141332,
185
+ "step": 2625
186
+ },
187
+ {
188
+ "epoch": 0.7,
189
+ "eval_avg_sts": 0.76482808857313,
190
+ "eval_sickr_spearman": 0.7270758876759156,
191
+ "eval_stsb_spearman": 0.8025802894703444,
192
+ "step": 2750
193
+ },
194
+ {
195
+ "epoch": 0.74,
196
+ "eval_avg_sts": 0.7659614109106493,
197
+ "eval_sickr_spearman": 0.7289854911304676,
198
+ "eval_stsb_spearman": 0.802937330690831,
199
+ "step": 2875
200
+ },
201
+ {
202
+ "epoch": 0.77,
203
+ "learning_rate": 6.964422830816484e-06,
204
+ "loss": 0.0005,
205
+ "step": 3000
206
+ },
207
+ {
208
+ "epoch": 0.77,
209
+ "eval_avg_sts": 0.7669678121477455,
210
+ "eval_sickr_spearman": 0.7313520084790959,
211
+ "eval_stsb_spearman": 0.802583615816395,
212
+ "step": 3000
213
+ },
214
+ {
215
+ "epoch": 0.8,
216
+ "eval_avg_sts": 0.767953736021239,
217
+ "eval_sickr_spearman": 0.7305751992550179,
218
+ "eval_stsb_spearman": 0.8053322727874599,
219
+ "step": 3125
220
+ },
221
+ {
222
+ "epoch": 0.83,
223
+ "eval_avg_sts": 0.7669771718476196,
224
+ "eval_sickr_spearman": 0.7286785670473275,
225
+ "eval_stsb_spearman": 0.8052757766479117,
226
+ "step": 3250
227
+ },
228
+ {
229
+ "epoch": 0.86,
230
+ "eval_avg_sts": 0.7669738896262592,
231
+ "eval_sickr_spearman": 0.7274589450676203,
232
+ "eval_stsb_spearman": 0.8064888341848981,
233
+ "step": 3375
234
+ }
235
+ ],
236
+ "max_steps": 3907,
237
+ "num_train_epochs": 1,
238
+ "total_flos": 0,
239
+ "trial_name": null,
240
+ "trial_params": null
241
+ }
unsup-whitenedcse-bert-base-uncased/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc017acc014637a03799adfe67fafbf28a757d517eb124cfdc7199e10ef49e09
3
+ size 2095
unsup-whitenedcse-bert-base-uncased/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
unsup-whitenedcse-bert-base-uncased/vocab.txt ADDED
The diff for this file is too large to render. See raw diff