Upload 13 files
Browse files- unsup-whitenedcse-bert-base-uncased/config.json +28 -0
- unsup-whitenedcse-bert-base-uncased/eval_results.txt +12 -0
- unsup-whitenedcse-bert-base-uncased/merges.txt +0 -0
- unsup-whitenedcse-bert-base-uncased/optimizer.pt +3 -0
- unsup-whitenedcse-bert-base-uncased/pytorch_model.bin +3 -0
- unsup-whitenedcse-bert-base-uncased/scheduler.pt +3 -0
- unsup-whitenedcse-bert-base-uncased/special_tokens_map.json +1 -0
- unsup-whitenedcse-bert-base-uncased/tokenizer_config.json +1 -0
- unsup-whitenedcse-bert-base-uncased/train_results.txt +3 -0
- unsup-whitenedcse-bert-base-uncased/trainer_state.json +241 -0
- unsup-whitenedcse-bert-base-uncased/training_args.bin +3 -0
- unsup-whitenedcse-bert-base-uncased/vocab.json +0 -0
- unsup-whitenedcse-bert-base-uncased/vocab.txt +0 -0
unsup-whitenedcse-bert-base-uncased/config.json
ADDED
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "bert-base-uncased",
|
3 |
+
"architectures": [
|
4 |
+
"BertForCL"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"attention_probs_dropout_prob_noise": 0.1,
|
8 |
+
"gradient_checkpointing": false,
|
9 |
+
"hidden_act": "gelu",
|
10 |
+
"hidden_dropout_prob": 0.1,
|
11 |
+
"hidden_dropout_prob_noise": 0.1,
|
12 |
+
"hidden_dropout_prob_noise2": 0.2,
|
13 |
+
"hidden_size": 768,
|
14 |
+
"initializer_range": 0.02,
|
15 |
+
"intermediate_size": 3072,
|
16 |
+
"layer_norm_eps": 1e-12,
|
17 |
+
"max_position_embeddings": 512,
|
18 |
+
"model_type": "bert",
|
19 |
+
"multi_dropout": false,
|
20 |
+
"num_attention_heads": 12,
|
21 |
+
"num_hidden_layers": 12,
|
22 |
+
"pad_token_id": 0,
|
23 |
+
"position_embedding_type": "absolute",
|
24 |
+
"transformers_version": "4.2.1",
|
25 |
+
"type_vocab_size": 2,
|
26 |
+
"use_cache": true,
|
27 |
+
"vocab_size": 30522
|
28 |
+
}
|
unsup-whitenedcse-bert-base-uncased/eval_results.txt
ADDED
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
epoch = 1.0
|
2 |
+
eval_CR = 88.47
|
3 |
+
eval_MPQA = 88.56
|
4 |
+
eval_MR = 83.73
|
5 |
+
eval_MRPC = 74.34
|
6 |
+
eval_SST2 = 88.42
|
7 |
+
eval_SUBJ = 94.47
|
8 |
+
eval_TREC = 80.36
|
9 |
+
eval_avg_sts = 0.7561232564078078
|
10 |
+
eval_avg_transfer = 85.47857142857143
|
11 |
+
eval_sickr_spearman = 0.7206682316875482
|
12 |
+
eval_stsb_spearman = 0.7915782811280672
|
unsup-whitenedcse-bert-base-uncased/merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
unsup-whitenedcse-bert-base-uncased/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7b4fd94a826bbad7e019e3973538b5725a58ae316429e5d76e56385ef5cb9b4b
|
3 |
+
size 875973285
|
unsup-whitenedcse-bert-base-uncased/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0185e4dbb506a8307a9da59d0bef90f31e3e929afc2d9ebd3daf1281fa14ac46
|
3 |
+
size 438012465
|
unsup-whitenedcse-bert-base-uncased/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e44519aa377dfc697372b150b09cfa3048148d6bf5a0ea461a9012cf304d40e4
|
3 |
+
size 623
|
unsup-whitenedcse-bert-base-uncased/special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
|
unsup-whitenedcse-bert-base-uncased/tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "name_or_path": "bert-base-uncased"}
|
unsup-whitenedcse-bert-base-uncased/train_results.txt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
epoch = 1.0
|
2 |
+
train_runtime = 7194.6299
|
3 |
+
train_samples_per_second = 1.086
|
unsup-whitenedcse-bert-base-uncased/trainer_state.json
ADDED
@@ -0,0 +1,241 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"best_metric": 0.8064888341848981,
|
3 |
+
"best_model_checkpoint": "result/my-unsup-simcse-bert-base-uncased",
|
4 |
+
"epoch": 0.8638341438443818,
|
5 |
+
"global_step": 3375,
|
6 |
+
"is_hyper_param_search": false,
|
7 |
+
"is_local_process_zero": true,
|
8 |
+
"is_world_process_zero": true,
|
9 |
+
"log_history": [
|
10 |
+
{
|
11 |
+
"epoch": 0.03,
|
12 |
+
"eval_avg_sts": 0.7035617984039024,
|
13 |
+
"eval_sickr_spearman": 0.6848745407030954,
|
14 |
+
"eval_stsb_spearman": 0.7222490561047095,
|
15 |
+
"step": 125
|
16 |
+
},
|
17 |
+
{
|
18 |
+
"epoch": 0.06,
|
19 |
+
"eval_avg_sts": 0.7266642057578656,
|
20 |
+
"eval_sickr_spearman": 0.70274752572599,
|
21 |
+
"eval_stsb_spearman": 0.7505808857897414,
|
22 |
+
"step": 250
|
23 |
+
},
|
24 |
+
{
|
25 |
+
"epoch": 0.1,
|
26 |
+
"eval_avg_sts": 0.7461885709126848,
|
27 |
+
"eval_sickr_spearman": 0.7217250002652101,
|
28 |
+
"eval_stsb_spearman": 0.7706521415601595,
|
29 |
+
"step": 375
|
30 |
+
},
|
31 |
+
{
|
32 |
+
"epoch": 0.13,
|
33 |
+
"learning_rate": 2.6160737138469415e-05,
|
34 |
+
"loss": 0.0013,
|
35 |
+
"step": 500
|
36 |
+
},
|
37 |
+
{
|
38 |
+
"epoch": 0.13,
|
39 |
+
"eval_avg_sts": 0.7463389976451559,
|
40 |
+
"eval_sickr_spearman": 0.7199349731494692,
|
41 |
+
"eval_stsb_spearman": 0.7727430221408426,
|
42 |
+
"step": 500
|
43 |
+
},
|
44 |
+
{
|
45 |
+
"epoch": 0.16,
|
46 |
+
"eval_avg_sts": 0.743938555014451,
|
47 |
+
"eval_sickr_spearman": 0.7180073918502566,
|
48 |
+
"eval_stsb_spearman": 0.7698697181786454,
|
49 |
+
"step": 625
|
50 |
+
},
|
51 |
+
{
|
52 |
+
"epoch": 0.19,
|
53 |
+
"eval_avg_sts": 0.7550432481097468,
|
54 |
+
"eval_sickr_spearman": 0.7234769647664954,
|
55 |
+
"eval_stsb_spearman": 0.7866095314529982,
|
56 |
+
"step": 750
|
57 |
+
},
|
58 |
+
{
|
59 |
+
"epoch": 0.22,
|
60 |
+
"eval_avg_sts": 0.7565707728294095,
|
61 |
+
"eval_sickr_spearman": 0.7258336460737318,
|
62 |
+
"eval_stsb_spearman": 0.7873078995850874,
|
63 |
+
"step": 875
|
64 |
+
},
|
65 |
+
{
|
66 |
+
"epoch": 0.26,
|
67 |
+
"learning_rate": 2.232147427693883e-05,
|
68 |
+
"loss": 0.0004,
|
69 |
+
"step": 1000
|
70 |
+
},
|
71 |
+
{
|
72 |
+
"epoch": 0.26,
|
73 |
+
"eval_avg_sts": 0.7492950228664624,
|
74 |
+
"eval_sickr_spearman": 0.7233381976577096,
|
75 |
+
"eval_stsb_spearman": 0.7752518480752153,
|
76 |
+
"step": 1000
|
77 |
+
},
|
78 |
+
{
|
79 |
+
"epoch": 0.29,
|
80 |
+
"eval_avg_sts": 0.7522516406505854,
|
81 |
+
"eval_sickr_spearman": 0.7287197031901267,
|
82 |
+
"eval_stsb_spearman": 0.775783578111044,
|
83 |
+
"step": 1125
|
84 |
+
},
|
85 |
+
{
|
86 |
+
"epoch": 0.32,
|
87 |
+
"eval_avg_sts": 0.7552964382615647,
|
88 |
+
"eval_sickr_spearman": 0.729468085275092,
|
89 |
+
"eval_stsb_spearman": 0.7811247912480374,
|
90 |
+
"step": 1250
|
91 |
+
},
|
92 |
+
{
|
93 |
+
"epoch": 0.35,
|
94 |
+
"eval_avg_sts": 0.7568556891093262,
|
95 |
+
"eval_sickr_spearman": 0.7272718476322126,
|
96 |
+
"eval_stsb_spearman": 0.7864395305864398,
|
97 |
+
"step": 1375
|
98 |
+
},
|
99 |
+
{
|
100 |
+
"epoch": 0.38,
|
101 |
+
"learning_rate": 1.8482211415408245e-05,
|
102 |
+
"loss": 0.0004,
|
103 |
+
"step": 1500
|
104 |
+
},
|
105 |
+
{
|
106 |
+
"epoch": 0.38,
|
107 |
+
"eval_avg_sts": 0.7657081366190621,
|
108 |
+
"eval_sickr_spearman": 0.7298737076524129,
|
109 |
+
"eval_stsb_spearman": 0.8015425655857115,
|
110 |
+
"step": 1500
|
111 |
+
},
|
112 |
+
{
|
113 |
+
"epoch": 0.42,
|
114 |
+
"eval_avg_sts": 0.7645138872097019,
|
115 |
+
"eval_sickr_spearman": 0.7293550763436135,
|
116 |
+
"eval_stsb_spearman": 0.7996726980757902,
|
117 |
+
"step": 1625
|
118 |
+
},
|
119 |
+
{
|
120 |
+
"epoch": 0.45,
|
121 |
+
"eval_avg_sts": 0.7604696897809164,
|
122 |
+
"eval_sickr_spearman": 0.7309107216292885,
|
123 |
+
"eval_stsb_spearman": 0.7900286579325445,
|
124 |
+
"step": 1750
|
125 |
+
},
|
126 |
+
{
|
127 |
+
"epoch": 0.48,
|
128 |
+
"eval_avg_sts": 0.7665958188961237,
|
129 |
+
"eval_sickr_spearman": 0.7308310370977922,
|
130 |
+
"eval_stsb_spearman": 0.8023606006944553,
|
131 |
+
"step": 1875
|
132 |
+
},
|
133 |
+
{
|
134 |
+
"epoch": 0.51,
|
135 |
+
"learning_rate": 1.4642948553877656e-05,
|
136 |
+
"loss": 0.0004,
|
137 |
+
"step": 2000
|
138 |
+
},
|
139 |
+
{
|
140 |
+
"epoch": 0.51,
|
141 |
+
"eval_avg_sts": 0.7650091226879439,
|
142 |
+
"eval_sickr_spearman": 0.729455872126414,
|
143 |
+
"eval_stsb_spearman": 0.8005623732494735,
|
144 |
+
"step": 2000
|
145 |
+
},
|
146 |
+
{
|
147 |
+
"epoch": 0.54,
|
148 |
+
"eval_avg_sts": 0.7652971356117957,
|
149 |
+
"eval_sickr_spearman": 0.7297830140299209,
|
150 |
+
"eval_stsb_spearman": 0.8008112571936706,
|
151 |
+
"step": 2125
|
152 |
+
},
|
153 |
+
{
|
154 |
+
"epoch": 0.58,
|
155 |
+
"eval_avg_sts": 0.7629965694314595,
|
156 |
+
"eval_sickr_spearman": 0.7272991579004704,
|
157 |
+
"eval_stsb_spearman": 0.7986939809624487,
|
158 |
+
"step": 2250
|
159 |
+
},
|
160 |
+
{
|
161 |
+
"epoch": 0.61,
|
162 |
+
"eval_avg_sts": 0.7639744709473972,
|
163 |
+
"eval_sickr_spearman": 0.7219387730993263,
|
164 |
+
"eval_stsb_spearman": 0.8060101687954682,
|
165 |
+
"step": 2375
|
166 |
+
},
|
167 |
+
{
|
168 |
+
"epoch": 0.64,
|
169 |
+
"learning_rate": 1.080368569234707e-05,
|
170 |
+
"loss": 0.0003,
|
171 |
+
"step": 2500
|
172 |
+
},
|
173 |
+
{
|
174 |
+
"epoch": 0.64,
|
175 |
+
"eval_avg_sts": 0.7649654793413658,
|
176 |
+
"eval_sickr_spearman": 0.7252475539569496,
|
177 |
+
"eval_stsb_spearman": 0.8046834047257821,
|
178 |
+
"step": 2500
|
179 |
+
},
|
180 |
+
{
|
181 |
+
"epoch": 0.67,
|
182 |
+
"eval_avg_sts": 0.7645559824476773,
|
183 |
+
"eval_sickr_spearman": 0.7272004062812213,
|
184 |
+
"eval_stsb_spearman": 0.8019115586141332,
|
185 |
+
"step": 2625
|
186 |
+
},
|
187 |
+
{
|
188 |
+
"epoch": 0.7,
|
189 |
+
"eval_avg_sts": 0.76482808857313,
|
190 |
+
"eval_sickr_spearman": 0.7270758876759156,
|
191 |
+
"eval_stsb_spearman": 0.8025802894703444,
|
192 |
+
"step": 2750
|
193 |
+
},
|
194 |
+
{
|
195 |
+
"epoch": 0.74,
|
196 |
+
"eval_avg_sts": 0.7659614109106493,
|
197 |
+
"eval_sickr_spearman": 0.7289854911304676,
|
198 |
+
"eval_stsb_spearman": 0.802937330690831,
|
199 |
+
"step": 2875
|
200 |
+
},
|
201 |
+
{
|
202 |
+
"epoch": 0.77,
|
203 |
+
"learning_rate": 6.964422830816484e-06,
|
204 |
+
"loss": 0.0005,
|
205 |
+
"step": 3000
|
206 |
+
},
|
207 |
+
{
|
208 |
+
"epoch": 0.77,
|
209 |
+
"eval_avg_sts": 0.7669678121477455,
|
210 |
+
"eval_sickr_spearman": 0.7313520084790959,
|
211 |
+
"eval_stsb_spearman": 0.802583615816395,
|
212 |
+
"step": 3000
|
213 |
+
},
|
214 |
+
{
|
215 |
+
"epoch": 0.8,
|
216 |
+
"eval_avg_sts": 0.767953736021239,
|
217 |
+
"eval_sickr_spearman": 0.7305751992550179,
|
218 |
+
"eval_stsb_spearman": 0.8053322727874599,
|
219 |
+
"step": 3125
|
220 |
+
},
|
221 |
+
{
|
222 |
+
"epoch": 0.83,
|
223 |
+
"eval_avg_sts": 0.7669771718476196,
|
224 |
+
"eval_sickr_spearman": 0.7286785670473275,
|
225 |
+
"eval_stsb_spearman": 0.8052757766479117,
|
226 |
+
"step": 3250
|
227 |
+
},
|
228 |
+
{
|
229 |
+
"epoch": 0.86,
|
230 |
+
"eval_avg_sts": 0.7669738896262592,
|
231 |
+
"eval_sickr_spearman": 0.7274589450676203,
|
232 |
+
"eval_stsb_spearman": 0.8064888341848981,
|
233 |
+
"step": 3375
|
234 |
+
}
|
235 |
+
],
|
236 |
+
"max_steps": 3907,
|
237 |
+
"num_train_epochs": 1,
|
238 |
+
"total_flos": 0,
|
239 |
+
"trial_name": null,
|
240 |
+
"trial_params": null
|
241 |
+
}
|
unsup-whitenedcse-bert-base-uncased/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bc017acc014637a03799adfe67fafbf28a757d517eb124cfdc7199e10ef49e09
|
3 |
+
size 2095
|
unsup-whitenedcse-bert-base-uncased/vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
unsup-whitenedcse-bert-base-uncased/vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|