bobox commited on
Commit
116c83d
·
verified ·
1 Parent(s): 3ea06b4

Training in progress, step 657, checkpoint

Browse files
checkpoint-657/1_AdvancedWeightedPooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "embed_dim": 1024,
3
+ "num_heads": 8,
4
+ "dropout": 0.05,
5
+ "bias": true,
6
+ "use_layernorm": false,
7
+ "use_MLP": false,
8
+ "MLP_h_size": 1024,
9
+ "use_residual": false
10
+ }
checkpoint-657/1_AdvancedWeightedPooling/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:192031eaa51e6c7e16d624ce1ecf51c07d10e2497f42917ba584b010b704897a
3
+ size 16795776
checkpoint-657/README.md ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-657/config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "BAAI/bge-m3",
3
+ "architectures": [
4
+ "XLMRobertaModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 1024,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 4096,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 8194,
17
+ "model_type": "xlm-roberta",
18
+ "num_attention_heads": 16,
19
+ "num_hidden_layers": 24,
20
+ "output_past": true,
21
+ "pad_token_id": 1,
22
+ "position_embedding_type": "absolute",
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.47.0",
25
+ "type_vocab_size": 1,
26
+ "use_cache": true,
27
+ "vocab_size": 250002
28
+ }
checkpoint-657/config_sentence_transformers.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "3.3.1",
4
+ "transformers": "4.47.0",
5
+ "pytorch": "2.5.1+cu121"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null,
9
+ "similarity_fn_name": "cosine"
10
+ }
checkpoint-657/modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_AdvancedWeightedPooling",
12
+ "type": "__main__.AdvancedWeightedPooling"
13
+ }
14
+ ]
checkpoint-657/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea45d56f6bf0c5f88c1c88e3d7878fa8e4e4aabc99f6e28772d6284c38aa8208
3
+ size 33591506
checkpoint-657/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e969b03be79a3f792216f1febef058e86c2b7f27374fb1e044d87938ceea1ba
3
+ size 2271151270
checkpoint-657/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53528fda7876f4fd66afde130a1e9bfa327dc9073de838efad40ae811ca44f16
3
+ size 14244
checkpoint-657/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:662519926e1ae5ab445c3bb73f7486e2a4033e221530866a06c7c0cbbcf6a969
3
+ size 1000
checkpoint-657/sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 8192,
3
+ "do_lower_case": false
4
+ }
checkpoint-657/sentencepiece.bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
3
+ size 5069051
checkpoint-657/special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "<unk>",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
checkpoint-657/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4f7e21bec3fb0044ca0bb2d50eb5d4d8c596273c422baef84466d2c73748b9c
3
+ size 17083053
checkpoint-657/tokenizer_config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "250001": {
36
+ "content": "<mask>",
37
+ "lstrip": true,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "<s>",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "<s>",
47
+ "eos_token": "</s>",
48
+ "extra_special_tokens": {},
49
+ "mask_token": "<mask>",
50
+ "model_max_length": 8192,
51
+ "pad_token": "<pad>",
52
+ "sep_token": "</s>",
53
+ "sp_model_kwargs": {},
54
+ "tokenizer_class": "XLMRobertaTokenizer",
55
+ "unk_token": "<unk>"
56
+ }
checkpoint-657/trainer_state.json ADDED
@@ -0,0 +1,1051 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.9012345679012346,
5
+ "eval_steps": 110,
6
+ "global_step": 657,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.015089163237311385,
13
+ "grad_norm": 3.8793580532073975,
14
+ "learning_rate": 1.8281535648994517e-05,
15
+ "loss": 0.7558,
16
+ "step": 11
17
+ },
18
+ {
19
+ "epoch": 0.03017832647462277,
20
+ "grad_norm": 3.666482448577881,
21
+ "learning_rate": 3.839122486288849e-05,
22
+ "loss": 0.7776,
23
+ "step": 22
24
+ },
25
+ {
26
+ "epoch": 0.04526748971193416,
27
+ "grad_norm": 3.493319511413574,
28
+ "learning_rate": 5.850091407678245e-05,
29
+ "loss": 0.379,
30
+ "step": 33
31
+ },
32
+ {
33
+ "epoch": 0.06035665294924554,
34
+ "grad_norm": 2.435302972793579,
35
+ "learning_rate": 7.861060329067642e-05,
36
+ "loss": 0.4947,
37
+ "step": 44
38
+ },
39
+ {
40
+ "epoch": 0.07544581618655692,
41
+ "grad_norm": 2.2073612213134766,
42
+ "learning_rate": 9.872029250457039e-05,
43
+ "loss": 0.3406,
44
+ "step": 55
45
+ },
46
+ {
47
+ "epoch": 0.09053497942386832,
48
+ "grad_norm": 0.5107505917549133,
49
+ "learning_rate": 0.00011882998171846434,
50
+ "loss": 0.2391,
51
+ "step": 66
52
+ },
53
+ {
54
+ "epoch": 0.1056241426611797,
55
+ "grad_norm": 2.4110350608825684,
56
+ "learning_rate": 0.00013893967093235832,
57
+ "loss": 0.2154,
58
+ "step": 77
59
+ },
60
+ {
61
+ "epoch": 0.12071330589849108,
62
+ "grad_norm": 3.4400408267974854,
63
+ "learning_rate": 0.00015904936014625229,
64
+ "loss": 0.2662,
65
+ "step": 88
66
+ },
67
+ {
68
+ "epoch": 0.13580246913580246,
69
+ "grad_norm": 3.31318998336792,
70
+ "learning_rate": 0.00017915904936014626,
71
+ "loss": 0.283,
72
+ "step": 99
73
+ },
74
+ {
75
+ "epoch": 0.15089163237311384,
76
+ "grad_norm": 2.624321699142456,
77
+ "learning_rate": 0.00019926873857404023,
78
+ "loss": 0.2391,
79
+ "step": 110
80
+ },
81
+ {
82
+ "epoch": 0.15089163237311384,
83
+ "eval_Qnli-dev_cosine_accuracy": 0.7353515625,
84
+ "eval_Qnli-dev_cosine_accuracy_threshold": 0.641769528388977,
85
+ "eval_Qnli-dev_cosine_ap": 0.7934694922676566,
86
+ "eval_Qnli-dev_cosine_f1": 0.7255734919286321,
87
+ "eval_Qnli-dev_cosine_f1_threshold": 0.533623456954956,
88
+ "eval_Qnli-dev_cosine_precision": 0.6170520231213873,
89
+ "eval_Qnli-dev_cosine_recall": 0.8804123711340206,
90
+ "eval_allNLI-dev_cosine_accuracy": 0.7451171875,
91
+ "eval_allNLI-dev_cosine_accuracy_threshold": 0.7058684229850769,
92
+ "eval_allNLI-dev_cosine_ap": 0.6358738534384165,
93
+ "eval_allNLI-dev_cosine_f1": 0.6633039092055485,
94
+ "eval_allNLI-dev_cosine_f1_threshold": 0.6644865274429321,
95
+ "eval_allNLI-dev_cosine_precision": 0.579295154185022,
96
+ "eval_allNLI-dev_cosine_recall": 0.775811209439528,
97
+ "eval_sequential_score": 0.7934694922676566,
98
+ "eval_sts-test_pearson_cosine": 0.8508165029528609,
99
+ "eval_sts-test_spearman_cosine": 0.8665992028008191,
100
+ "eval_vitaminc-pairs_loss": 1.550615668296814,
101
+ "eval_vitaminc-pairs_runtime": 24.2459,
102
+ "eval_vitaminc-pairs_samples_per_second": 10.558,
103
+ "eval_vitaminc-pairs_steps_per_second": 0.041,
104
+ "step": 110
105
+ },
106
+ {
107
+ "epoch": 0.15089163237311384,
108
+ "eval_negation-triplets_loss": 0.8121126294136047,
109
+ "eval_negation-triplets_runtime": 4.2821,
110
+ "eval_negation-triplets_samples_per_second": 59.784,
111
+ "eval_negation-triplets_steps_per_second": 0.234,
112
+ "step": 110
113
+ },
114
+ {
115
+ "epoch": 0.15089163237311384,
116
+ "eval_scitail-pairs-pos_loss": 0.05080736428499222,
117
+ "eval_scitail-pairs-pos_runtime": 3.0909,
118
+ "eval_scitail-pairs-pos_samples_per_second": 82.824,
119
+ "eval_scitail-pairs-pos_steps_per_second": 0.324,
120
+ "step": 110
121
+ },
122
+ {
123
+ "epoch": 0.15089163237311384,
124
+ "eval_scitail-pairs-qa_loss": 0.011191274970769882,
125
+ "eval_scitail-pairs-qa_runtime": 2.3574,
126
+ "eval_scitail-pairs-qa_samples_per_second": 108.596,
127
+ "eval_scitail-pairs-qa_steps_per_second": 0.424,
128
+ "step": 110
129
+ },
130
+ {
131
+ "epoch": 0.15089163237311384,
132
+ "eval_xsum-pairs_loss": 0.2039160132408142,
133
+ "eval_xsum-pairs_runtime": 12.8416,
134
+ "eval_xsum-pairs_samples_per_second": 19.935,
135
+ "eval_xsum-pairs_steps_per_second": 0.078,
136
+ "step": 110
137
+ },
138
+ {
139
+ "epoch": 0.15089163237311384,
140
+ "eval_sciq_pairs_loss": 0.023365184664726257,
141
+ "eval_sciq_pairs_runtime": 20.6659,
142
+ "eval_sciq_pairs_samples_per_second": 12.388,
143
+ "eval_sciq_pairs_steps_per_second": 0.048,
144
+ "step": 110
145
+ },
146
+ {
147
+ "epoch": 0.15089163237311384,
148
+ "eval_qasc_pairs_loss": 0.558290421962738,
149
+ "eval_qasc_pairs_runtime": 3.009,
150
+ "eval_qasc_pairs_samples_per_second": 85.077,
151
+ "eval_qasc_pairs_steps_per_second": 0.332,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.15089163237311384,
156
+ "eval_openbookqa_pairs_loss": 1.253723382949829,
157
+ "eval_openbookqa_pairs_runtime": 2.2457,
158
+ "eval_openbookqa_pairs_samples_per_second": 113.993,
159
+ "eval_openbookqa_pairs_steps_per_second": 0.445,
160
+ "step": 110
161
+ },
162
+ {
163
+ "epoch": 0.15089163237311384,
164
+ "eval_nq_pairs_loss": 0.10213108360767365,
165
+ "eval_nq_pairs_runtime": 18.094,
166
+ "eval_nq_pairs_samples_per_second": 14.148,
167
+ "eval_nq_pairs_steps_per_second": 0.055,
168
+ "step": 110
169
+ },
170
+ {
171
+ "epoch": 0.15089163237311384,
172
+ "eval_trivia_pairs_loss": 0.4372706711292267,
173
+ "eval_trivia_pairs_runtime": 16.9326,
174
+ "eval_trivia_pairs_samples_per_second": 15.119,
175
+ "eval_trivia_pairs_steps_per_second": 0.059,
176
+ "step": 110
177
+ },
178
+ {
179
+ "epoch": 0.15089163237311384,
180
+ "eval_gooaq_pairs_loss": 0.2727060914039612,
181
+ "eval_gooaq_pairs_runtime": 3.6277,
182
+ "eval_gooaq_pairs_samples_per_second": 70.568,
183
+ "eval_gooaq_pairs_steps_per_second": 0.276,
184
+ "step": 110
185
+ },
186
+ {
187
+ "epoch": 0.15089163237311384,
188
+ "eval_paws-pos_loss": 0.040396444499492645,
189
+ "eval_paws-pos_runtime": 2.9381,
190
+ "eval_paws-pos_samples_per_second": 87.132,
191
+ "eval_paws-pos_steps_per_second": 0.34,
192
+ "step": 110
193
+ },
194
+ {
195
+ "epoch": 0.15089163237311384,
196
+ "eval_global_dataset_loss": 0.1820984184741974,
197
+ "eval_global_dataset_runtime": 125.361,
198
+ "eval_global_dataset_samples_per_second": 9.764,
199
+ "eval_global_dataset_steps_per_second": 0.04,
200
+ "step": 110
201
+ },
202
+ {
203
+ "epoch": 0.16598079561042525,
204
+ "grad_norm": 3.345057725906372,
205
+ "learning_rate": 0.00021937842778793417,
206
+ "loss": 0.2456,
207
+ "step": 121
208
+ },
209
+ {
210
+ "epoch": 0.18106995884773663,
211
+ "grad_norm": 2.2555935382843018,
212
+ "learning_rate": 0.00023948811700182814,
213
+ "loss": 0.4199,
214
+ "step": 132
215
+ },
216
+ {
217
+ "epoch": 0.19615912208504802,
218
+ "grad_norm": 2.3139288425445557,
219
+ "learning_rate": 0.0002595978062157221,
220
+ "loss": 0.2809,
221
+ "step": 143
222
+ },
223
+ {
224
+ "epoch": 0.2112482853223594,
225
+ "grad_norm": 3.585463047027588,
226
+ "learning_rate": 0.0002797074954296161,
227
+ "loss": 0.5773,
228
+ "step": 154
229
+ },
230
+ {
231
+ "epoch": 0.22633744855967078,
232
+ "grad_norm": 2.4281251430511475,
233
+ "learning_rate": 0.00029981718464351003,
234
+ "loss": 0.3734,
235
+ "step": 165
236
+ },
237
+ {
238
+ "epoch": 0.24142661179698216,
239
+ "grad_norm": 0.2383209466934204,
240
+ "learning_rate": 0.000319926873857404,
241
+ "loss": 0.2348,
242
+ "step": 176
243
+ },
244
+ {
245
+ "epoch": 0.25651577503429357,
246
+ "grad_norm": 2.4634456634521484,
247
+ "learning_rate": 0.00034003656307129797,
248
+ "loss": 0.4421,
249
+ "step": 187
250
+ },
251
+ {
252
+ "epoch": 0.2716049382716049,
253
+ "grad_norm": 3.1270384788513184,
254
+ "learning_rate": 0.00036014625228519197,
255
+ "loss": 0.5076,
256
+ "step": 198
257
+ },
258
+ {
259
+ "epoch": 0.28669410150891633,
260
+ "grad_norm": 0.7871516346931458,
261
+ "learning_rate": 0.0003802559414990859,
262
+ "loss": 0.211,
263
+ "step": 209
264
+ },
265
+ {
266
+ "epoch": 0.3017832647462277,
267
+ "grad_norm": 2.159247636795044,
268
+ "learning_rate": 0.00040036563071297986,
269
+ "loss": 0.3514,
270
+ "step": 220
271
+ },
272
+ {
273
+ "epoch": 0.3017832647462277,
274
+ "eval_Qnli-dev_cosine_accuracy": 0.736328125,
275
+ "eval_Qnli-dev_cosine_accuracy_threshold": 0.6637322902679443,
276
+ "eval_Qnli-dev_cosine_ap": 0.7960534826633536,
277
+ "eval_Qnli-dev_cosine_f1": 0.7298050139275766,
278
+ "eval_Qnli-dev_cosine_f1_threshold": 0.5937396287918091,
279
+ "eval_Qnli-dev_cosine_precision": 0.6638513513513513,
280
+ "eval_Qnli-dev_cosine_recall": 0.8103092783505155,
281
+ "eval_allNLI-dev_cosine_accuracy": 0.755859375,
282
+ "eval_allNLI-dev_cosine_accuracy_threshold": 0.728554904460907,
283
+ "eval_allNLI-dev_cosine_ap": 0.6433680273177467,
284
+ "eval_allNLI-dev_cosine_f1": 0.661818181818182,
285
+ "eval_allNLI-dev_cosine_f1_threshold": 0.6872978210449219,
286
+ "eval_allNLI-dev_cosine_precision": 0.5617283950617284,
287
+ "eval_allNLI-dev_cosine_recall": 0.8053097345132744,
288
+ "eval_sequential_score": 0.7960534826633536,
289
+ "eval_sts-test_pearson_cosine": 0.8381104761555598,
290
+ "eval_sts-test_spearman_cosine": 0.8624525294470655,
291
+ "eval_vitaminc-pairs_loss": 1.5675894021987915,
292
+ "eval_vitaminc-pairs_runtime": 24.2649,
293
+ "eval_vitaminc-pairs_samples_per_second": 10.55,
294
+ "eval_vitaminc-pairs_steps_per_second": 0.041,
295
+ "step": 220
296
+ },
297
+ {
298
+ "epoch": 0.3017832647462277,
299
+ "eval_negation-triplets_loss": 0.923573911190033,
300
+ "eval_negation-triplets_runtime": 4.311,
301
+ "eval_negation-triplets_samples_per_second": 59.384,
302
+ "eval_negation-triplets_steps_per_second": 0.232,
303
+ "step": 220
304
+ },
305
+ {
306
+ "epoch": 0.3017832647462277,
307
+ "eval_scitail-pairs-pos_loss": 0.0530293844640255,
308
+ "eval_scitail-pairs-pos_runtime": 3.1182,
309
+ "eval_scitail-pairs-pos_samples_per_second": 82.098,
310
+ "eval_scitail-pairs-pos_steps_per_second": 0.321,
311
+ "step": 220
312
+ },
313
+ {
314
+ "epoch": 0.3017832647462277,
315
+ "eval_scitail-pairs-qa_loss": 0.008582310751080513,
316
+ "eval_scitail-pairs-qa_runtime": 2.402,
317
+ "eval_scitail-pairs-qa_samples_per_second": 106.58,
318
+ "eval_scitail-pairs-qa_steps_per_second": 0.416,
319
+ "step": 220
320
+ },
321
+ {
322
+ "epoch": 0.3017832647462277,
323
+ "eval_xsum-pairs_loss": 0.19049452245235443,
324
+ "eval_xsum-pairs_runtime": 12.8763,
325
+ "eval_xsum-pairs_samples_per_second": 19.882,
326
+ "eval_xsum-pairs_steps_per_second": 0.078,
327
+ "step": 220
328
+ },
329
+ {
330
+ "epoch": 0.3017832647462277,
331
+ "eval_sciq_pairs_loss": 0.02437273971736431,
332
+ "eval_sciq_pairs_runtime": 20.6537,
333
+ "eval_sciq_pairs_samples_per_second": 12.395,
334
+ "eval_sciq_pairs_steps_per_second": 0.048,
335
+ "step": 220
336
+ },
337
+ {
338
+ "epoch": 0.3017832647462277,
339
+ "eval_qasc_pairs_loss": 0.6957117915153503,
340
+ "eval_qasc_pairs_runtime": 3.0168,
341
+ "eval_qasc_pairs_samples_per_second": 84.858,
342
+ "eval_qasc_pairs_steps_per_second": 0.331,
343
+ "step": 220
344
+ },
345
+ {
346
+ "epoch": 0.3017832647462277,
347
+ "eval_openbookqa_pairs_loss": 1.2585959434509277,
348
+ "eval_openbookqa_pairs_runtime": 2.2495,
349
+ "eval_openbookqa_pairs_samples_per_second": 113.801,
350
+ "eval_openbookqa_pairs_steps_per_second": 0.445,
351
+ "step": 220
352
+ },
353
+ {
354
+ "epoch": 0.3017832647462277,
355
+ "eval_nq_pairs_loss": 0.12953564524650574,
356
+ "eval_nq_pairs_runtime": 18.127,
357
+ "eval_nq_pairs_samples_per_second": 14.123,
358
+ "eval_nq_pairs_steps_per_second": 0.055,
359
+ "step": 220
360
+ },
361
+ {
362
+ "epoch": 0.3017832647462277,
363
+ "eval_trivia_pairs_loss": 0.46085307002067566,
364
+ "eval_trivia_pairs_runtime": 16.9635,
365
+ "eval_trivia_pairs_samples_per_second": 15.091,
366
+ "eval_trivia_pairs_steps_per_second": 0.059,
367
+ "step": 220
368
+ },
369
+ {
370
+ "epoch": 0.3017832647462277,
371
+ "eval_gooaq_pairs_loss": 0.2918424606323242,
372
+ "eval_gooaq_pairs_runtime": 3.6275,
373
+ "eval_gooaq_pairs_samples_per_second": 70.573,
374
+ "eval_gooaq_pairs_steps_per_second": 0.276,
375
+ "step": 220
376
+ },
377
+ {
378
+ "epoch": 0.3017832647462277,
379
+ "eval_paws-pos_loss": 0.041661862283945084,
380
+ "eval_paws-pos_runtime": 2.9518,
381
+ "eval_paws-pos_samples_per_second": 86.727,
382
+ "eval_paws-pos_steps_per_second": 0.339,
383
+ "step": 220
384
+ },
385
+ {
386
+ "epoch": 0.3017832647462277,
387
+ "eval_global_dataset_loss": 0.3610426187515259,
388
+ "eval_global_dataset_runtime": 125.4174,
389
+ "eval_global_dataset_samples_per_second": 9.759,
390
+ "eval_global_dataset_steps_per_second": 0.04,
391
+ "step": 220
392
+ },
393
+ {
394
+ "epoch": 0.3168724279835391,
395
+ "grad_norm": 2.7020387649536133,
396
+ "learning_rate": 0.00042047531992687385,
397
+ "loss": 0.2981,
398
+ "step": 231
399
+ },
400
+ {
401
+ "epoch": 0.3319615912208505,
402
+ "grad_norm": 2.0000767707824707,
403
+ "learning_rate": 0.0004405850091407678,
404
+ "loss": 0.3045,
405
+ "step": 242
406
+ },
407
+ {
408
+ "epoch": 0.34705075445816186,
409
+ "grad_norm": 2.539660692214966,
410
+ "learning_rate": 0.00046069469835466185,
411
+ "loss": 0.3126,
412
+ "step": 253
413
+ },
414
+ {
415
+ "epoch": 0.36213991769547327,
416
+ "grad_norm": 2.418445348739624,
417
+ "learning_rate": 0.0004808043875685558,
418
+ "loss": 0.7813,
419
+ "step": 264
420
+ },
421
+ {
422
+ "epoch": 0.3772290809327846,
423
+ "grad_norm": 0.13016735017299652,
424
+ "learning_rate": 0.0005009140767824497,
425
+ "loss": 0.547,
426
+ "step": 275
427
+ },
428
+ {
429
+ "epoch": 0.39231824417009603,
430
+ "grad_norm": 2.2323102951049805,
431
+ "learning_rate": 0.0005210237659963437,
432
+ "loss": 0.4698,
433
+ "step": 286
434
+ },
435
+ {
436
+ "epoch": 0.4074074074074074,
437
+ "grad_norm": 3.097975492477417,
438
+ "learning_rate": 0.0005411334552102377,
439
+ "loss": 0.5427,
440
+ "step": 297
441
+ },
442
+ {
443
+ "epoch": 0.4224965706447188,
444
+ "grad_norm": 0.7084994316101074,
445
+ "learning_rate": 0.0005612431444241316,
446
+ "loss": 0.3151,
447
+ "step": 308
448
+ },
449
+ {
450
+ "epoch": 0.4375857338820302,
451
+ "grad_norm": 1.7643369436264038,
452
+ "learning_rate": 0.0005813528336380256,
453
+ "loss": 0.4687,
454
+ "step": 319
455
+ },
456
+ {
457
+ "epoch": 0.45267489711934156,
458
+ "grad_norm": 1.7608978748321533,
459
+ "learning_rate": 0.0006014625228519196,
460
+ "loss": 0.3769,
461
+ "step": 330
462
+ },
463
+ {
464
+ "epoch": 0.45267489711934156,
465
+ "eval_Qnli-dev_cosine_accuracy": 0.7255859375,
466
+ "eval_Qnli-dev_cosine_accuracy_threshold": 0.6892818212509155,
467
+ "eval_Qnli-dev_cosine_ap": 0.7884120709809157,
468
+ "eval_Qnli-dev_cosine_f1": 0.7228464419475655,
469
+ "eval_Qnli-dev_cosine_f1_threshold": 0.6395477056503296,
470
+ "eval_Qnli-dev_cosine_precision": 0.6620926243567753,
471
+ "eval_Qnli-dev_cosine_recall": 0.7958762886597938,
472
+ "eval_allNLI-dev_cosine_accuracy": 0.7421875,
473
+ "eval_allNLI-dev_cosine_accuracy_threshold": 0.7702663540840149,
474
+ "eval_allNLI-dev_cosine_ap": 0.6347046378974335,
475
+ "eval_allNLI-dev_cosine_f1": 0.6455542021924483,
476
+ "eval_allNLI-dev_cosine_f1_threshold": 0.7014378309249878,
477
+ "eval_allNLI-dev_cosine_precision": 0.549792531120332,
478
+ "eval_allNLI-dev_cosine_recall": 0.7817109144542773,
479
+ "eval_sequential_score": 0.7884120709809157,
480
+ "eval_sts-test_pearson_cosine": 0.8389808770066287,
481
+ "eval_sts-test_spearman_cosine": 0.861226668384837,
482
+ "eval_vitaminc-pairs_loss": 1.5675371885299683,
483
+ "eval_vitaminc-pairs_runtime": 24.2072,
484
+ "eval_vitaminc-pairs_samples_per_second": 10.575,
485
+ "eval_vitaminc-pairs_steps_per_second": 0.041,
486
+ "step": 330
487
+ },
488
+ {
489
+ "epoch": 0.45267489711934156,
490
+ "eval_negation-triplets_loss": 0.7654371857643127,
491
+ "eval_negation-triplets_runtime": 4.2588,
492
+ "eval_negation-triplets_samples_per_second": 60.111,
493
+ "eval_negation-triplets_steps_per_second": 0.235,
494
+ "step": 330
495
+ },
496
+ {
497
+ "epoch": 0.45267489711934156,
498
+ "eval_scitail-pairs-pos_loss": 0.04239173233509064,
499
+ "eval_scitail-pairs-pos_runtime": 3.0875,
500
+ "eval_scitail-pairs-pos_samples_per_second": 82.915,
501
+ "eval_scitail-pairs-pos_steps_per_second": 0.324,
502
+ "step": 330
503
+ },
504
+ {
505
+ "epoch": 0.45267489711934156,
506
+ "eval_scitail-pairs-qa_loss": 0.010665436275303364,
507
+ "eval_scitail-pairs-qa_runtime": 2.3433,
508
+ "eval_scitail-pairs-qa_samples_per_second": 109.246,
509
+ "eval_scitail-pairs-qa_steps_per_second": 0.427,
510
+ "step": 330
511
+ },
512
+ {
513
+ "epoch": 0.45267489711934156,
514
+ "eval_xsum-pairs_loss": 0.20898626744747162,
515
+ "eval_xsum-pairs_runtime": 12.8471,
516
+ "eval_xsum-pairs_samples_per_second": 19.927,
517
+ "eval_xsum-pairs_steps_per_second": 0.078,
518
+ "step": 330
519
+ },
520
+ {
521
+ "epoch": 0.45267489711934156,
522
+ "eval_sciq_pairs_loss": 0.03412973880767822,
523
+ "eval_sciq_pairs_runtime": 20.7279,
524
+ "eval_sciq_pairs_samples_per_second": 12.351,
525
+ "eval_sciq_pairs_steps_per_second": 0.048,
526
+ "step": 330
527
+ },
528
+ {
529
+ "epoch": 0.45267489711934156,
530
+ "eval_qasc_pairs_loss": 0.7678776383399963,
531
+ "eval_qasc_pairs_runtime": 3.0154,
532
+ "eval_qasc_pairs_samples_per_second": 84.896,
533
+ "eval_qasc_pairs_steps_per_second": 0.332,
534
+ "step": 330
535
+ },
536
+ {
537
+ "epoch": 0.45267489711934156,
538
+ "eval_openbookqa_pairs_loss": 1.3723315000534058,
539
+ "eval_openbookqa_pairs_runtime": 2.2474,
540
+ "eval_openbookqa_pairs_samples_per_second": 113.91,
541
+ "eval_openbookqa_pairs_steps_per_second": 0.445,
542
+ "step": 330
543
+ },
544
+ {
545
+ "epoch": 0.45267489711934156,
546
+ "eval_nq_pairs_loss": 0.15752817690372467,
547
+ "eval_nq_pairs_runtime": 18.1484,
548
+ "eval_nq_pairs_samples_per_second": 14.106,
549
+ "eval_nq_pairs_steps_per_second": 0.055,
550
+ "step": 330
551
+ },
552
+ {
553
+ "epoch": 0.45267489711934156,
554
+ "eval_trivia_pairs_loss": 0.6312745213508606,
555
+ "eval_trivia_pairs_runtime": 16.9738,
556
+ "eval_trivia_pairs_samples_per_second": 15.082,
557
+ "eval_trivia_pairs_steps_per_second": 0.059,
558
+ "step": 330
559
+ },
560
+ {
561
+ "epoch": 0.45267489711934156,
562
+ "eval_gooaq_pairs_loss": 0.47655048966407776,
563
+ "eval_gooaq_pairs_runtime": 3.6272,
564
+ "eval_gooaq_pairs_samples_per_second": 70.579,
565
+ "eval_gooaq_pairs_steps_per_second": 0.276,
566
+ "step": 330
567
+ },
568
+ {
569
+ "epoch": 0.45267489711934156,
570
+ "eval_paws-pos_loss": 0.04226630926132202,
571
+ "eval_paws-pos_runtime": 2.9441,
572
+ "eval_paws-pos_samples_per_second": 86.954,
573
+ "eval_paws-pos_steps_per_second": 0.34,
574
+ "step": 330
575
+ },
576
+ {
577
+ "epoch": 0.45267489711934156,
578
+ "eval_global_dataset_loss": 0.23270446062088013,
579
+ "eval_global_dataset_runtime": 125.3499,
580
+ "eval_global_dataset_samples_per_second": 9.765,
581
+ "eval_global_dataset_steps_per_second": 0.04,
582
+ "step": 330
583
+ },
584
+ {
585
+ "epoch": 0.46776406035665297,
586
+ "grad_norm": 0.37502893805503845,
587
+ "learning_rate": 0.0006215722120658135,
588
+ "loss": 0.3832,
589
+ "step": 341
590
+ },
591
+ {
592
+ "epoch": 0.4828532235939643,
593
+ "grad_norm": 2.488353967666626,
594
+ "learning_rate": 0.0006416819012797075,
595
+ "loss": 0.4392,
596
+ "step": 352
597
+ },
598
+ {
599
+ "epoch": 0.49794238683127573,
600
+ "grad_norm": 0.0037633629981428385,
601
+ "learning_rate": 0.0006617915904936015,
602
+ "loss": 0.3929,
603
+ "step": 363
604
+ },
605
+ {
606
+ "epoch": 0.5130315500685871,
607
+ "grad_norm": 0.467970073223114,
608
+ "learning_rate": 0.0006819012797074955,
609
+ "loss": 0.2611,
610
+ "step": 374
611
+ },
612
+ {
613
+ "epoch": 0.5281207133058985,
614
+ "grad_norm": 2.020796537399292,
615
+ "learning_rate": 0.0007020109689213894,
616
+ "loss": 0.3528,
617
+ "step": 385
618
+ },
619
+ {
620
+ "epoch": 0.5432098765432098,
621
+ "grad_norm": 2.0242817401885986,
622
+ "learning_rate": 0.0007221206581352834,
623
+ "loss": 0.738,
624
+ "step": 396
625
+ },
626
+ {
627
+ "epoch": 0.5582990397805213,
628
+ "grad_norm": 2.9996001720428467,
629
+ "learning_rate": 0.0007422303473491774,
630
+ "loss": 0.4016,
631
+ "step": 407
632
+ },
633
+ {
634
+ "epoch": 0.5733882030178327,
635
+ "grad_norm": 1.8861972093582153,
636
+ "learning_rate": 0.0007623400365630713,
637
+ "loss": 0.3589,
638
+ "step": 418
639
+ },
640
+ {
641
+ "epoch": 0.588477366255144,
642
+ "grad_norm": 0.24432632327079773,
643
+ "learning_rate": 0.0007824497257769653,
644
+ "loss": 0.3057,
645
+ "step": 429
646
+ },
647
+ {
648
+ "epoch": 0.6035665294924554,
649
+ "grad_norm": 1.1939449310302734,
650
+ "learning_rate": 0.0008025594149908593,
651
+ "loss": 0.6095,
652
+ "step": 440
653
+ },
654
+ {
655
+ "epoch": 0.6035665294924554,
656
+ "eval_Qnli-dev_cosine_accuracy": 0.7294921875,
657
+ "eval_Qnli-dev_cosine_accuracy_threshold": 0.7398079633712769,
658
+ "eval_Qnli-dev_cosine_ap": 0.7745603475556573,
659
+ "eval_Qnli-dev_cosine_f1": 0.7162790697674419,
660
+ "eval_Qnli-dev_cosine_f1_threshold": 0.6965705752372742,
661
+ "eval_Qnli-dev_cosine_precision": 0.652542372881356,
662
+ "eval_Qnli-dev_cosine_recall": 0.7938144329896907,
663
+ "eval_allNLI-dev_cosine_accuracy": 0.7431640625,
664
+ "eval_allNLI-dev_cosine_accuracy_threshold": 0.7988871335983276,
665
+ "eval_allNLI-dev_cosine_ap": 0.6431097859832409,
666
+ "eval_allNLI-dev_cosine_f1": 0.6502793296089385,
667
+ "eval_allNLI-dev_cosine_f1_threshold": 0.7334020733833313,
668
+ "eval_allNLI-dev_cosine_precision": 0.5233812949640287,
669
+ "eval_allNLI-dev_cosine_recall": 0.8584070796460177,
670
+ "eval_sequential_score": 0.7745603475556573,
671
+ "eval_sts-test_pearson_cosine": 0.8337056827457451,
672
+ "eval_sts-test_spearman_cosine": 0.8627129106149749,
673
+ "eval_vitaminc-pairs_loss": 1.9911772012710571,
674
+ "eval_vitaminc-pairs_runtime": 24.6526,
675
+ "eval_vitaminc-pairs_samples_per_second": 10.384,
676
+ "eval_vitaminc-pairs_steps_per_second": 0.041,
677
+ "step": 440
678
+ },
679
+ {
680
+ "epoch": 0.6035665294924554,
681
+ "eval_negation-triplets_loss": 0.7211357951164246,
682
+ "eval_negation-triplets_runtime": 4.2956,
683
+ "eval_negation-triplets_samples_per_second": 59.595,
684
+ "eval_negation-triplets_steps_per_second": 0.233,
685
+ "step": 440
686
+ },
687
+ {
688
+ "epoch": 0.6035665294924554,
689
+ "eval_scitail-pairs-pos_loss": 0.08597714453935623,
690
+ "eval_scitail-pairs-pos_runtime": 3.1006,
691
+ "eval_scitail-pairs-pos_samples_per_second": 82.565,
692
+ "eval_scitail-pairs-pos_steps_per_second": 0.323,
693
+ "step": 440
694
+ },
695
+ {
696
+ "epoch": 0.6035665294924554,
697
+ "eval_scitail-pairs-qa_loss": 0.01645529642701149,
698
+ "eval_scitail-pairs-qa_runtime": 2.3457,
699
+ "eval_scitail-pairs-qa_samples_per_second": 109.136,
700
+ "eval_scitail-pairs-qa_steps_per_second": 0.426,
701
+ "step": 440
702
+ },
703
+ {
704
+ "epoch": 0.6035665294924554,
705
+ "eval_xsum-pairs_loss": 0.41817042231559753,
706
+ "eval_xsum-pairs_runtime": 12.8711,
707
+ "eval_xsum-pairs_samples_per_second": 19.89,
708
+ "eval_xsum-pairs_steps_per_second": 0.078,
709
+ "step": 440
710
+ },
711
+ {
712
+ "epoch": 0.6035665294924554,
713
+ "eval_sciq_pairs_loss": 0.03533514216542244,
714
+ "eval_sciq_pairs_runtime": 20.7828,
715
+ "eval_sciq_pairs_samples_per_second": 12.318,
716
+ "eval_sciq_pairs_steps_per_second": 0.048,
717
+ "step": 440
718
+ },
719
+ {
720
+ "epoch": 0.6035665294924554,
721
+ "eval_qasc_pairs_loss": 0.6740420460700989,
722
+ "eval_qasc_pairs_runtime": 3.0138,
723
+ "eval_qasc_pairs_samples_per_second": 84.942,
724
+ "eval_qasc_pairs_steps_per_second": 0.332,
725
+ "step": 440
726
+ },
727
+ {
728
+ "epoch": 0.6035665294924554,
729
+ "eval_openbookqa_pairs_loss": 1.6906702518463135,
730
+ "eval_openbookqa_pairs_runtime": 2.257,
731
+ "eval_openbookqa_pairs_samples_per_second": 113.425,
732
+ "eval_openbookqa_pairs_steps_per_second": 0.443,
733
+ "step": 440
734
+ },
735
+ {
736
+ "epoch": 0.6035665294924554,
737
+ "eval_nq_pairs_loss": 0.1867213398218155,
738
+ "eval_nq_pairs_runtime": 18.1321,
739
+ "eval_nq_pairs_samples_per_second": 14.119,
740
+ "eval_nq_pairs_steps_per_second": 0.055,
741
+ "step": 440
742
+ },
743
+ {
744
+ "epoch": 0.6035665294924554,
745
+ "eval_trivia_pairs_loss": 0.6214608550071716,
746
+ "eval_trivia_pairs_runtime": 17.0099,
747
+ "eval_trivia_pairs_samples_per_second": 15.05,
748
+ "eval_trivia_pairs_steps_per_second": 0.059,
749
+ "step": 440
750
+ },
751
+ {
752
+ "epoch": 0.6035665294924554,
753
+ "eval_gooaq_pairs_loss": 0.4756861925125122,
754
+ "eval_gooaq_pairs_runtime": 3.6268,
755
+ "eval_gooaq_pairs_samples_per_second": 70.587,
756
+ "eval_gooaq_pairs_steps_per_second": 0.276,
757
+ "step": 440
758
+ },
759
+ {
760
+ "epoch": 0.6035665294924554,
761
+ "eval_paws-pos_loss": 0.04365835338830948,
762
+ "eval_paws-pos_runtime": 2.9421,
763
+ "eval_paws-pos_samples_per_second": 87.012,
764
+ "eval_paws-pos_steps_per_second": 0.34,
765
+ "step": 440
766
+ },
767
+ {
768
+ "epoch": 0.6035665294924554,
769
+ "eval_global_dataset_loss": 0.29543137550354004,
770
+ "eval_global_dataset_runtime": 125.3874,
771
+ "eval_global_dataset_samples_per_second": 9.762,
772
+ "eval_global_dataset_steps_per_second": 0.04,
773
+ "step": 440
774
+ },
775
+ {
776
+ "epoch": 0.6186556927297668,
777
+ "grad_norm": 1.2445456981658936,
778
+ "learning_rate": 0.0008226691042047533,
779
+ "loss": 0.3255,
780
+ "step": 451
781
+ },
782
+ {
783
+ "epoch": 0.6337448559670782,
784
+ "grad_norm": 1.707679033279419,
785
+ "learning_rate": 0.0008427787934186472,
786
+ "loss": 0.4211,
787
+ "step": 462
788
+ },
789
+ {
790
+ "epoch": 0.6488340192043895,
791
+ "grad_norm": 1.942112922668457,
792
+ "learning_rate": 0.0008628884826325412,
793
+ "loss": 0.5372,
794
+ "step": 473
795
+ },
796
+ {
797
+ "epoch": 0.663923182441701,
798
+ "grad_norm": 1.4559303522109985,
799
+ "learning_rate": 0.0008829981718464352,
800
+ "loss": 0.5747,
801
+ "step": 484
802
+ },
803
+ {
804
+ "epoch": 0.6790123456790124,
805
+ "grad_norm": 1.95571768283844,
806
+ "learning_rate": 0.000903107861060329,
807
+ "loss": 0.4851,
808
+ "step": 495
809
+ },
810
+ {
811
+ "epoch": 0.6941015089163237,
812
+ "grad_norm": 2.090074300765991,
813
+ "learning_rate": 0.000923217550274223,
814
+ "loss": 0.243,
815
+ "step": 506
816
+ },
817
+ {
818
+ "epoch": 0.7091906721536351,
819
+ "grad_norm": 2.7206690311431885,
820
+ "learning_rate": 0.000943327239488117,
821
+ "loss": 0.4899,
822
+ "step": 517
823
+ },
824
+ {
825
+ "epoch": 0.7242798353909465,
826
+ "grad_norm": 0.44097644090652466,
827
+ "learning_rate": 0.0009634369287020109,
828
+ "loss": 0.2475,
829
+ "step": 528
830
+ },
831
+ {
832
+ "epoch": 0.7393689986282579,
833
+ "grad_norm": 2.089887857437134,
834
+ "learning_rate": 0.000983546617915905,
835
+ "loss": 0.5144,
836
+ "step": 539
837
+ },
838
+ {
839
+ "epoch": 0.7544581618655692,
840
+ "grad_norm": 1.4973657131195068,
841
+ "learning_rate": 0.0009999966974148216,
842
+ "loss": 0.7029,
843
+ "step": 550
844
+ },
845
+ {
846
+ "epoch": 0.7544581618655692,
847
+ "eval_Qnli-dev_cosine_accuracy": 0.7177734375,
848
+ "eval_Qnli-dev_cosine_accuracy_threshold": 0.7129597663879395,
849
+ "eval_Qnli-dev_cosine_ap": 0.7715845808061284,
850
+ "eval_Qnli-dev_cosine_f1": 0.7221719457013575,
851
+ "eval_Qnli-dev_cosine_f1_threshold": 0.6851584911346436,
852
+ "eval_Qnli-dev_cosine_precision": 0.6435483870967742,
853
+ "eval_Qnli-dev_cosine_recall": 0.822680412371134,
854
+ "eval_allNLI-dev_cosine_accuracy": 0.73828125,
855
+ "eval_allNLI-dev_cosine_accuracy_threshold": 0.8462234139442444,
856
+ "eval_allNLI-dev_cosine_ap": 0.6244503911184303,
857
+ "eval_allNLI-dev_cosine_f1": 0.6362545018007203,
858
+ "eval_allNLI-dev_cosine_f1_threshold": 0.7372293472290039,
859
+ "eval_allNLI-dev_cosine_precision": 0.5364372469635628,
860
+ "eval_allNLI-dev_cosine_recall": 0.7817109144542773,
861
+ "eval_sequential_score": 0.7715845808061284,
862
+ "eval_sts-test_pearson_cosine": 0.8245725507043073,
863
+ "eval_sts-test_spearman_cosine": 0.8556805260032072,
864
+ "eval_vitaminc-pairs_loss": 2.2667033672332764,
865
+ "eval_vitaminc-pairs_runtime": 24.1991,
866
+ "eval_vitaminc-pairs_samples_per_second": 10.579,
867
+ "eval_vitaminc-pairs_steps_per_second": 0.041,
868
+ "step": 550
869
+ },
870
+ {
871
+ "epoch": 0.7544581618655692,
872
+ "eval_negation-triplets_loss": 0.8184330463409424,
873
+ "eval_negation-triplets_runtime": 4.2711,
874
+ "eval_negation-triplets_samples_per_second": 59.938,
875
+ "eval_negation-triplets_steps_per_second": 0.234,
876
+ "step": 550
877
+ },
878
+ {
879
+ "epoch": 0.7544581618655692,
880
+ "eval_scitail-pairs-pos_loss": 0.11127087473869324,
881
+ "eval_scitail-pairs-pos_runtime": 3.0897,
882
+ "eval_scitail-pairs-pos_samples_per_second": 82.855,
883
+ "eval_scitail-pairs-pos_steps_per_second": 0.324,
884
+ "step": 550
885
+ },
886
+ {
887
+ "epoch": 0.7544581618655692,
888
+ "eval_scitail-pairs-qa_loss": 0.018279315903782845,
889
+ "eval_scitail-pairs-qa_runtime": 2.3553,
890
+ "eval_scitail-pairs-qa_samples_per_second": 108.691,
891
+ "eval_scitail-pairs-qa_steps_per_second": 0.425,
892
+ "step": 550
893
+ },
894
+ {
895
+ "epoch": 0.7544581618655692,
896
+ "eval_xsum-pairs_loss": 0.42622849345207214,
897
+ "eval_xsum-pairs_runtime": 12.8556,
898
+ "eval_xsum-pairs_samples_per_second": 19.914,
899
+ "eval_xsum-pairs_steps_per_second": 0.078,
900
+ "step": 550
901
+ },
902
+ {
903
+ "epoch": 0.7544581618655692,
904
+ "eval_sciq_pairs_loss": 0.039594996720552444,
905
+ "eval_sciq_pairs_runtime": 20.7135,
906
+ "eval_sciq_pairs_samples_per_second": 12.359,
907
+ "eval_sciq_pairs_steps_per_second": 0.048,
908
+ "step": 550
909
+ },
910
+ {
911
+ "epoch": 0.7544581618655692,
912
+ "eval_qasc_pairs_loss": 0.88427734375,
913
+ "eval_qasc_pairs_runtime": 3.014,
914
+ "eval_qasc_pairs_samples_per_second": 84.937,
915
+ "eval_qasc_pairs_steps_per_second": 0.332,
916
+ "step": 550
917
+ },
918
+ {
919
+ "epoch": 0.7544581618655692,
920
+ "eval_openbookqa_pairs_loss": 1.3467748165130615,
921
+ "eval_openbookqa_pairs_runtime": 2.245,
922
+ "eval_openbookqa_pairs_samples_per_second": 114.032,
923
+ "eval_openbookqa_pairs_steps_per_second": 0.445,
924
+ "step": 550
925
+ },
926
+ {
927
+ "epoch": 0.7544581618655692,
928
+ "eval_nq_pairs_loss": 0.1950540691614151,
929
+ "eval_nq_pairs_runtime": 18.1193,
930
+ "eval_nq_pairs_samples_per_second": 14.129,
931
+ "eval_nq_pairs_steps_per_second": 0.055,
932
+ "step": 550
933
+ },
934
+ {
935
+ "epoch": 0.7544581618655692,
936
+ "eval_trivia_pairs_loss": 0.8971078395843506,
937
+ "eval_trivia_pairs_runtime": 16.9594,
938
+ "eval_trivia_pairs_samples_per_second": 15.095,
939
+ "eval_trivia_pairs_steps_per_second": 0.059,
940
+ "step": 550
941
+ },
942
+ {
943
+ "epoch": 0.7544581618655692,
944
+ "eval_gooaq_pairs_loss": 0.5568686127662659,
945
+ "eval_gooaq_pairs_runtime": 3.6329,
946
+ "eval_gooaq_pairs_samples_per_second": 70.468,
947
+ "eval_gooaq_pairs_steps_per_second": 0.275,
948
+ "step": 550
949
+ },
950
+ {
951
+ "epoch": 0.7544581618655692,
952
+ "eval_paws-pos_loss": 0.04390861093997955,
953
+ "eval_paws-pos_runtime": 2.9399,
954
+ "eval_paws-pos_samples_per_second": 87.078,
955
+ "eval_paws-pos_steps_per_second": 0.34,
956
+ "step": 550
957
+ },
958
+ {
959
+ "epoch": 0.7544581618655692,
960
+ "eval_global_dataset_loss": 0.2895234227180481,
961
+ "eval_global_dataset_runtime": 125.42,
962
+ "eval_global_dataset_samples_per_second": 9.759,
963
+ "eval_global_dataset_steps_per_second": 0.04,
964
+ "step": 550
965
+ },
966
+ {
967
+ "epoch": 0.7695473251028807,
968
+ "grad_norm": 1.4781888723373413,
969
+ "learning_rate": 0.0009998604728164411,
970
+ "loss": 0.4442,
971
+ "step": 561
972
+ },
973
+ {
974
+ "epoch": 0.7846364883401921,
975
+ "grad_norm": 1.0631585121154785,
976
+ "learning_rate": 0.0009995245109131869,
977
+ "loss": 0.3803,
978
+ "step": 572
979
+ },
980
+ {
981
+ "epoch": 0.7997256515775034,
982
+ "grad_norm": 1.074621319770813,
983
+ "learning_rate": 0.000998988960871588,
984
+ "loss": 0.5295,
985
+ "step": 583
986
+ },
987
+ {
988
+ "epoch": 0.8148148148148148,
989
+ "grad_norm": 1.6665034294128418,
990
+ "learning_rate": 0.0009982540604749751,
991
+ "loss": 0.3499,
992
+ "step": 594
993
+ },
994
+ {
995
+ "epoch": 0.8299039780521262,
996
+ "grad_norm": 1.859250545501709,
997
+ "learning_rate": 0.000997320136017908,
998
+ "loss": 0.4366,
999
+ "step": 605
1000
+ },
1001
+ {
1002
+ "epoch": 0.8449931412894376,
1003
+ "grad_norm": 1.294621467590332,
1004
+ "learning_rate": 0.0009961876021612984,
1005
+ "loss": 0.941,
1006
+ "step": 616
1007
+ },
1008
+ {
1009
+ "epoch": 0.8600823045267489,
1010
+ "grad_norm": 0.0,
1011
+ "learning_rate": 0.000994856961748303,
1012
+ "loss": 0.3918,
1013
+ "step": 627
1014
+ },
1015
+ {
1016
+ "epoch": 0.8751714677640604,
1017
+ "grad_norm": 0.0,
1018
+ "learning_rate": 0.0009933288055810616,
1019
+ "loss": 0.5476,
1020
+ "step": 638
1021
+ },
1022
+ {
1023
+ "epoch": 0.8902606310013718,
1024
+ "grad_norm": 0.9047508835792542,
1025
+ "learning_rate": 0.000991603812158381,
1026
+ "loss": 0.3118,
1027
+ "step": 649
1028
+ }
1029
+ ],
1030
+ "logging_steps": 11,
1031
+ "max_steps": 2187,
1032
+ "num_input_tokens_seen": 0,
1033
+ "num_train_epochs": 3,
1034
+ "save_steps": 219,
1035
+ "stateful_callbacks": {
1036
+ "TrainerControl": {
1037
+ "args": {
1038
+ "should_epoch_stop": false,
1039
+ "should_evaluate": false,
1040
+ "should_log": false,
1041
+ "should_save": true,
1042
+ "should_training_stop": false
1043
+ },
1044
+ "attributes": {}
1045
+ }
1046
+ },
1047
+ "total_flos": 0.0,
1048
+ "train_batch_size": 128,
1049
+ "trial_name": null,
1050
+ "trial_params": null
1051
+ }
checkpoint-657/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc4e2f1b0b6ec0b954c31a7a2d7392e2f2f817d60f3db0f6f0a80bdf431c25bc
3
+ size 5880