AmberYifan commited on
Commit
934f873
·
verified ·
1 Parent(s): 057d64e

Model save

Browse files
README.md ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: alignment-handbook/zephyr-7b-sft-full
4
+ tags:
5
+ - trl
6
+ - dpo
7
+ - generated_from_trainer
8
+ model-index:
9
+ - name: dpo-v-trans
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # dpo-v-trans
17
+
18
+ This model is a fine-tuned version of [alignment-handbook/zephyr-7b-sft-full](https://huggingface.co/alignment-handbook/zephyr-7b-sft-full) on the None dataset.
19
+
20
+ ## Model description
21
+
22
+ More information needed
23
+
24
+ ## Intended uses & limitations
25
+
26
+ More information needed
27
+
28
+ ## Training and evaluation data
29
+
30
+ More information needed
31
+
32
+ ## Training procedure
33
+
34
+ ### Training hyperparameters
35
+
36
+ The following hyperparameters were used during training:
37
+ - learning_rate: 5e-07
38
+ - train_batch_size: 8
39
+ - eval_batch_size: 4
40
+ - seed: 42
41
+ - distributed_type: multi-GPU
42
+ - num_devices: 4
43
+ - total_train_batch_size: 32
44
+ - total_eval_batch_size: 16
45
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
+ - lr_scheduler_type: linear
47
+ - lr_scheduler_warmup_ratio: 0.1
48
+ - num_epochs: 1
49
+
50
+ ### Training results
51
+
52
+
53
+
54
+ ### Framework versions
55
+
56
+ - Transformers 4.39.0.dev0
57
+ - Pytorch 2.3.0+cu121
58
+ - Datasets 2.14.6
59
+ - Tokenizers 0.15.2
all_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "train_loss": 0.013828172848911992,
4
+ "train_runtime": 8967.2368,
5
+ "train_samples": 50000,
6
+ "train_samples_per_second": 5.576,
7
+ "train_steps_per_second": 0.174
8
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.39.0.dev0"
6
+ }
model-00001-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11d72a2e7956c0ae2f1a10e8a0551d299c117056f009b5c273081de1c6ae926d
3
+ size 4943162336
model-00002-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29ffdce2fede116916d894a3fa535c93abbed5b73fd8d5e621b47e87925a15e3
3
+ size 4999819336
model-00003-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9025628c250717b9ca3d8bacea11fdd0c8aa848b72a99927d3e3ffe5d2c8641
3
+ size 4540516344
model.safetensors.index.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 14483464192
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00003-of-00003.safetensors",
7
+ "model.embed_tokens.weight": "model-00001-of-00003.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00003.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
10
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
11
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
12
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
13
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
14
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
15
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
16
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
17
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00003.safetensors",
18
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
19
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
20
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
21
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
22
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
23
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
24
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
25
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
26
+ "model.layers.10.input_layernorm.weight": "model-00002-of-00003.safetensors",
27
+ "model.layers.10.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
28
+ "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
29
+ "model.layers.10.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
30
+ "model.layers.10.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
31
+ "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
32
+ "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
33
+ "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
34
+ "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
35
+ "model.layers.11.input_layernorm.weight": "model-00002-of-00003.safetensors",
36
+ "model.layers.11.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
37
+ "model.layers.11.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
38
+ "model.layers.11.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
39
+ "model.layers.11.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
40
+ "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
41
+ "model.layers.11.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
42
+ "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
43
+ "model.layers.11.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
44
+ "model.layers.12.input_layernorm.weight": "model-00002-of-00003.safetensors",
45
+ "model.layers.12.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
46
+ "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
47
+ "model.layers.12.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
48
+ "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
49
+ "model.layers.12.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
50
+ "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
51
+ "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
52
+ "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
53
+ "model.layers.13.input_layernorm.weight": "model-00002-of-00003.safetensors",
54
+ "model.layers.13.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
55
+ "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
56
+ "model.layers.13.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
57
+ "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
58
+ "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
59
+ "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
60
+ "model.layers.13.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
61
+ "model.layers.13.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
62
+ "model.layers.14.input_layernorm.weight": "model-00002-of-00003.safetensors",
63
+ "model.layers.14.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
64
+ "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
65
+ "model.layers.14.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
66
+ "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
67
+ "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
68
+ "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
69
+ "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
70
+ "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
71
+ "model.layers.15.input_layernorm.weight": "model-00002-of-00003.safetensors",
72
+ "model.layers.15.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
73
+ "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
74
+ "model.layers.15.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
75
+ "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
76
+ "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
77
+ "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
78
+ "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
79
+ "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
80
+ "model.layers.16.input_layernorm.weight": "model-00002-of-00003.safetensors",
81
+ "model.layers.16.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
82
+ "model.layers.16.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
83
+ "model.layers.16.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
84
+ "model.layers.16.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
85
+ "model.layers.16.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
86
+ "model.layers.16.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
87
+ "model.layers.16.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
88
+ "model.layers.16.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
89
+ "model.layers.17.input_layernorm.weight": "model-00002-of-00003.safetensors",
90
+ "model.layers.17.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
91
+ "model.layers.17.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
92
+ "model.layers.17.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
93
+ "model.layers.17.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
94
+ "model.layers.17.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
95
+ "model.layers.17.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
96
+ "model.layers.17.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
97
+ "model.layers.17.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
98
+ "model.layers.18.input_layernorm.weight": "model-00002-of-00003.safetensors",
99
+ "model.layers.18.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
100
+ "model.layers.18.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
101
+ "model.layers.18.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
102
+ "model.layers.18.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
103
+ "model.layers.18.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
104
+ "model.layers.18.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
105
+ "model.layers.18.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
106
+ "model.layers.18.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
107
+ "model.layers.19.input_layernorm.weight": "model-00002-of-00003.safetensors",
108
+ "model.layers.19.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
109
+ "model.layers.19.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
110
+ "model.layers.19.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
111
+ "model.layers.19.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
112
+ "model.layers.19.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
113
+ "model.layers.19.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
114
+ "model.layers.19.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
115
+ "model.layers.19.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
116
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00003.safetensors",
117
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
118
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
119
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
120
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
121
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
122
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
123
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
124
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
125
+ "model.layers.20.input_layernorm.weight": "model-00002-of-00003.safetensors",
126
+ "model.layers.20.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
127
+ "model.layers.20.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
128
+ "model.layers.20.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
129
+ "model.layers.20.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
130
+ "model.layers.20.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
131
+ "model.layers.20.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
132
+ "model.layers.20.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
133
+ "model.layers.20.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
134
+ "model.layers.21.input_layernorm.weight": "model-00002-of-00003.safetensors",
135
+ "model.layers.21.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
136
+ "model.layers.21.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
137
+ "model.layers.21.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
138
+ "model.layers.21.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
139
+ "model.layers.21.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
140
+ "model.layers.21.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
141
+ "model.layers.21.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
142
+ "model.layers.21.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
143
+ "model.layers.22.input_layernorm.weight": "model-00003-of-00003.safetensors",
144
+ "model.layers.22.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
145
+ "model.layers.22.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
146
+ "model.layers.22.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
147
+ "model.layers.22.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
148
+ "model.layers.22.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
149
+ "model.layers.22.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
150
+ "model.layers.22.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
151
+ "model.layers.22.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
152
+ "model.layers.23.input_layernorm.weight": "model-00003-of-00003.safetensors",
153
+ "model.layers.23.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
154
+ "model.layers.23.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
155
+ "model.layers.23.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
156
+ "model.layers.23.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
157
+ "model.layers.23.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
158
+ "model.layers.23.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
159
+ "model.layers.23.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
160
+ "model.layers.23.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
161
+ "model.layers.24.input_layernorm.weight": "model-00003-of-00003.safetensors",
162
+ "model.layers.24.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
163
+ "model.layers.24.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
164
+ "model.layers.24.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
165
+ "model.layers.24.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
166
+ "model.layers.24.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
167
+ "model.layers.24.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
168
+ "model.layers.24.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
169
+ "model.layers.24.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
170
+ "model.layers.25.input_layernorm.weight": "model-00003-of-00003.safetensors",
171
+ "model.layers.25.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
172
+ "model.layers.25.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
173
+ "model.layers.25.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
174
+ "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
175
+ "model.layers.25.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
176
+ "model.layers.25.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
177
+ "model.layers.25.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
178
+ "model.layers.25.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
179
+ "model.layers.26.input_layernorm.weight": "model-00003-of-00003.safetensors",
180
+ "model.layers.26.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
181
+ "model.layers.26.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
182
+ "model.layers.26.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
183
+ "model.layers.26.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
184
+ "model.layers.26.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
185
+ "model.layers.26.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
186
+ "model.layers.26.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
187
+ "model.layers.26.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
188
+ "model.layers.27.input_layernorm.weight": "model-00003-of-00003.safetensors",
189
+ "model.layers.27.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
190
+ "model.layers.27.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
191
+ "model.layers.27.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
192
+ "model.layers.27.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
193
+ "model.layers.27.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
194
+ "model.layers.27.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
195
+ "model.layers.27.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
196
+ "model.layers.27.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
197
+ "model.layers.28.input_layernorm.weight": "model-00003-of-00003.safetensors",
198
+ "model.layers.28.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
199
+ "model.layers.28.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
200
+ "model.layers.28.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
201
+ "model.layers.28.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
202
+ "model.layers.28.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
203
+ "model.layers.28.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
204
+ "model.layers.28.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
205
+ "model.layers.28.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
206
+ "model.layers.29.input_layernorm.weight": "model-00003-of-00003.safetensors",
207
+ "model.layers.29.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
208
+ "model.layers.29.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
209
+ "model.layers.29.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
210
+ "model.layers.29.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
211
+ "model.layers.29.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
212
+ "model.layers.29.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
213
+ "model.layers.29.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
214
+ "model.layers.29.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
215
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00003.safetensors",
216
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
217
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
218
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
219
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
220
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
221
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
222
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
223
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
224
+ "model.layers.30.input_layernorm.weight": "model-00003-of-00003.safetensors",
225
+ "model.layers.30.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
226
+ "model.layers.30.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
227
+ "model.layers.30.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
228
+ "model.layers.30.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
229
+ "model.layers.30.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
230
+ "model.layers.30.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
231
+ "model.layers.30.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
232
+ "model.layers.30.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
233
+ "model.layers.31.input_layernorm.weight": "model-00003-of-00003.safetensors",
234
+ "model.layers.31.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
235
+ "model.layers.31.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
236
+ "model.layers.31.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
237
+ "model.layers.31.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
238
+ "model.layers.31.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
239
+ "model.layers.31.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
240
+ "model.layers.31.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
241
+ "model.layers.31.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
242
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00003.safetensors",
243
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
244
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
245
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
246
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
247
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
248
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
249
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
250
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
251
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00003.safetensors",
252
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
253
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
254
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
255
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
256
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
257
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
258
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
259
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
260
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00003.safetensors",
261
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
262
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
263
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
264
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
265
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
266
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
267
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
268
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
269
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00003.safetensors",
270
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
271
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
272
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
273
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
274
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
275
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
276
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
277
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
278
+ "model.layers.8.input_layernorm.weight": "model-00001-of-00003.safetensors",
279
+ "model.layers.8.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
280
+ "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
281
+ "model.layers.8.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
282
+ "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
283
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
284
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
285
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
286
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
287
+ "model.layers.9.input_layernorm.weight": "model-00001-of-00003.safetensors",
288
+ "model.layers.9.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
289
+ "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
290
+ "model.layers.9.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
291
+ "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
292
+ "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
293
+ "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
294
+ "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
295
+ "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
296
+ "model.norm.weight": "model-00003-of-00003.safetensors"
297
+ }
298
+ }
runs/Jun12_23-04-31_gilbreth-j001.rcac.purdue.edu/events.out.tfevents.1718248519.gilbreth-j001.rcac.purdue.edu.252532.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3fd51a6b3a26741d3d5f088393e3143dafbb57f47d7491521f3dea0060bd1d60
3
- size 108452
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e5f664313e6d9ac928a8018aac2c018e4ee76ba1cb19fca51f1649f40edd1e0
3
+ size 112934
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "train_loss": 0.013828172848911992,
4
+ "train_runtime": 8967.2368,
5
+ "train_samples": 50000,
6
+ "train_samples_per_second": 5.576,
7
+ "train_steps_per_second": 0.174
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,2385 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
+ "eval_steps": 100,
6
+ "global_step": 1563,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.0,
13
+ "grad_norm": 269.7842833735451,
14
+ "learning_rate": 3.1847133757961784e-09,
15
+ "logits/chosen": -2.8781068325042725,
16
+ "logits/rejected": -3.18619966506958,
17
+ "logps/chosen": -132.49632263183594,
18
+ "logps/rejected": -97.49405670166016,
19
+ "loss": 0.6931,
20
+ "rewards/accuracies": 0.0,
21
+ "rewards/chosen": 0.0,
22
+ "rewards/margins": 0.0,
23
+ "rewards/rejected": 0.0,
24
+ "step": 1
25
+ },
26
+ {
27
+ "epoch": 0.01,
28
+ "grad_norm": 225.840088756475,
29
+ "learning_rate": 3.184713375796178e-08,
30
+ "logits/chosen": -2.7182540893554688,
31
+ "logits/rejected": -3.014042854309082,
32
+ "logps/chosen": -114.23727416992188,
33
+ "logps/rejected": -89.29263305664062,
34
+ "loss": 0.6761,
35
+ "rewards/accuracies": 0.7361111044883728,
36
+ "rewards/chosen": 0.005112478509545326,
37
+ "rewards/margins": 0.04120528697967529,
38
+ "rewards/rejected": -0.03609280660748482,
39
+ "step": 10
40
+ },
41
+ {
42
+ "epoch": 0.01,
43
+ "grad_norm": 139.5446217816026,
44
+ "learning_rate": 6.369426751592356e-08,
45
+ "logits/chosen": -2.7113776206970215,
46
+ "logits/rejected": -2.9822986125946045,
47
+ "logps/chosen": -120.40122985839844,
48
+ "logps/rejected": -87.35284423828125,
49
+ "loss": 0.4887,
50
+ "rewards/accuracies": 1.0,
51
+ "rewards/chosen": 0.11869080364704132,
52
+ "rewards/margins": 0.47487640380859375,
53
+ "rewards/rejected": -0.3561856150627136,
54
+ "step": 20
55
+ },
56
+ {
57
+ "epoch": 0.02,
58
+ "grad_norm": 52.502479723007994,
59
+ "learning_rate": 9.554140127388536e-08,
60
+ "logits/chosen": -2.7034497261047363,
61
+ "logits/rejected": -2.912091016769409,
62
+ "logps/chosen": -112.85099792480469,
63
+ "logps/rejected": -92.99781036376953,
64
+ "loss": 0.239,
65
+ "rewards/accuracies": 1.0,
66
+ "rewards/chosen": 0.27735695242881775,
67
+ "rewards/margins": 1.6260454654693604,
68
+ "rewards/rejected": -1.3486883640289307,
69
+ "step": 30
70
+ },
71
+ {
72
+ "epoch": 0.03,
73
+ "grad_norm": 24.342254507057344,
74
+ "learning_rate": 1.2738853503184713e-07,
75
+ "logits/chosen": -2.6779820919036865,
76
+ "logits/rejected": -2.9538276195526123,
77
+ "logps/chosen": -114.9598617553711,
78
+ "logps/rejected": -109.4762191772461,
79
+ "loss": 0.1007,
80
+ "rewards/accuracies": 1.0,
81
+ "rewards/chosen": 0.3945399522781372,
82
+ "rewards/margins": 2.8680927753448486,
83
+ "rewards/rejected": -2.47355318069458,
84
+ "step": 40
85
+ },
86
+ {
87
+ "epoch": 0.03,
88
+ "grad_norm": 16.84971880044338,
89
+ "learning_rate": 1.592356687898089e-07,
90
+ "logits/chosen": -2.63224720954895,
91
+ "logits/rejected": -2.869575023651123,
92
+ "logps/chosen": -120.35627746582031,
93
+ "logps/rejected": -114.70686340332031,
94
+ "loss": 0.0746,
95
+ "rewards/accuracies": 0.987500011920929,
96
+ "rewards/chosen": 0.559775173664093,
97
+ "rewards/margins": 3.929781436920166,
98
+ "rewards/rejected": -3.3700060844421387,
99
+ "step": 50
100
+ },
101
+ {
102
+ "epoch": 0.04,
103
+ "grad_norm": 22.345296735538415,
104
+ "learning_rate": 1.9108280254777072e-07,
105
+ "logits/chosen": -2.649770975112915,
106
+ "logits/rejected": -2.9390110969543457,
107
+ "logps/chosen": -118.53907775878906,
108
+ "logps/rejected": -128.2223358154297,
109
+ "loss": 0.049,
110
+ "rewards/accuracies": 0.987500011920929,
111
+ "rewards/chosen": 0.25389423966407776,
112
+ "rewards/margins": 4.323164939880371,
113
+ "rewards/rejected": -4.069270610809326,
114
+ "step": 60
115
+ },
116
+ {
117
+ "epoch": 0.04,
118
+ "grad_norm": 20.651969023179603,
119
+ "learning_rate": 2.2292993630573247e-07,
120
+ "logits/chosen": -2.620734453201294,
121
+ "logits/rejected": -2.904691696166992,
122
+ "logps/chosen": -118.86378479003906,
123
+ "logps/rejected": -142.30442810058594,
124
+ "loss": 0.0448,
125
+ "rewards/accuracies": 1.0,
126
+ "rewards/chosen": 0.2997106611728668,
127
+ "rewards/margins": 5.346270561218262,
128
+ "rewards/rejected": -5.046560287475586,
129
+ "step": 70
130
+ },
131
+ {
132
+ "epoch": 0.05,
133
+ "grad_norm": 18.756699820611576,
134
+ "learning_rate": 2.5477707006369425e-07,
135
+ "logits/chosen": -2.644515037536621,
136
+ "logits/rejected": -2.8679375648498535,
137
+ "logps/chosen": -141.45867919921875,
138
+ "logps/rejected": -139.5041961669922,
139
+ "loss": 0.0338,
140
+ "rewards/accuracies": 1.0,
141
+ "rewards/chosen": 0.016187960281968117,
142
+ "rewards/margins": 5.637503147125244,
143
+ "rewards/rejected": -5.6213154792785645,
144
+ "step": 80
145
+ },
146
+ {
147
+ "epoch": 0.06,
148
+ "grad_norm": 17.304042934264913,
149
+ "learning_rate": 2.86624203821656e-07,
150
+ "logits/chosen": -2.6533827781677246,
151
+ "logits/rejected": -2.8769655227661133,
152
+ "logps/chosen": -131.0616455078125,
153
+ "logps/rejected": -137.1123046875,
154
+ "loss": 0.0279,
155
+ "rewards/accuracies": 1.0,
156
+ "rewards/chosen": 0.2292163372039795,
157
+ "rewards/margins": 5.93363094329834,
158
+ "rewards/rejected": -5.704414367675781,
159
+ "step": 90
160
+ },
161
+ {
162
+ "epoch": 0.06,
163
+ "grad_norm": 17.418215277278303,
164
+ "learning_rate": 3.184713375796178e-07,
165
+ "logits/chosen": -2.5998005867004395,
166
+ "logits/rejected": -2.8436856269836426,
167
+ "logps/chosen": -118.36405181884766,
168
+ "logps/rejected": -150.16636657714844,
169
+ "loss": 0.0283,
170
+ "rewards/accuracies": 0.987500011920929,
171
+ "rewards/chosen": -0.1456707864999771,
172
+ "rewards/margins": 6.479343414306641,
173
+ "rewards/rejected": -6.625013828277588,
174
+ "step": 100
175
+ },
176
+ {
177
+ "epoch": 0.07,
178
+ "grad_norm": 12.48057297022973,
179
+ "learning_rate": 3.5031847133757957e-07,
180
+ "logits/chosen": -2.5239391326904297,
181
+ "logits/rejected": -2.7482004165649414,
182
+ "logps/chosen": -124.32521057128906,
183
+ "logps/rejected": -143.4300537109375,
184
+ "loss": 0.0177,
185
+ "rewards/accuracies": 1.0,
186
+ "rewards/chosen": 0.06046473979949951,
187
+ "rewards/margins": 6.917716026306152,
188
+ "rewards/rejected": -6.8572516441345215,
189
+ "step": 110
190
+ },
191
+ {
192
+ "epoch": 0.08,
193
+ "grad_norm": 11.25165220079385,
194
+ "learning_rate": 3.8216560509554143e-07,
195
+ "logits/chosen": -2.6064770221710205,
196
+ "logits/rejected": -2.8154265880584717,
197
+ "logps/chosen": -127.79212951660156,
198
+ "logps/rejected": -166.4564971923828,
199
+ "loss": 0.0185,
200
+ "rewards/accuracies": 1.0,
201
+ "rewards/chosen": 0.18089526891708374,
202
+ "rewards/margins": 8.502016067504883,
203
+ "rewards/rejected": -8.321121215820312,
204
+ "step": 120
205
+ },
206
+ {
207
+ "epoch": 0.08,
208
+ "grad_norm": 9.982779910930713,
209
+ "learning_rate": 4.140127388535032e-07,
210
+ "logits/chosen": -2.629976749420166,
211
+ "logits/rejected": -2.829049587249756,
212
+ "logps/chosen": -119.15159606933594,
213
+ "logps/rejected": -171.82168579101562,
214
+ "loss": 0.0242,
215
+ "rewards/accuracies": 0.987500011920929,
216
+ "rewards/chosen": 0.10970219224691391,
217
+ "rewards/margins": 8.503213882446289,
218
+ "rewards/rejected": -8.393510818481445,
219
+ "step": 130
220
+ },
221
+ {
222
+ "epoch": 0.09,
223
+ "grad_norm": 1.068441186553867,
224
+ "learning_rate": 4.4585987261146494e-07,
225
+ "logits/chosen": -2.4788918495178223,
226
+ "logits/rejected": -2.7960944175720215,
227
+ "logps/chosen": -122.7088623046875,
228
+ "logps/rejected": -180.638427734375,
229
+ "loss": 0.0047,
230
+ "rewards/accuracies": 1.0,
231
+ "rewards/chosen": -0.3950616419315338,
232
+ "rewards/margins": 8.885313987731934,
233
+ "rewards/rejected": -9.280375480651855,
234
+ "step": 140
235
+ },
236
+ {
237
+ "epoch": 0.1,
238
+ "grad_norm": 1.6980178715972738,
239
+ "learning_rate": 4.777070063694267e-07,
240
+ "logits/chosen": -2.533334732055664,
241
+ "logits/rejected": -2.7179126739501953,
242
+ "logps/chosen": -114.43672943115234,
243
+ "logps/rejected": -182.99908447265625,
244
+ "loss": 0.0115,
245
+ "rewards/accuracies": 0.987500011920929,
246
+ "rewards/chosen": 0.3883129954338074,
247
+ "rewards/margins": 10.101364135742188,
248
+ "rewards/rejected": -9.713052749633789,
249
+ "step": 150
250
+ },
251
+ {
252
+ "epoch": 0.1,
253
+ "grad_norm": 5.589702998473681,
254
+ "learning_rate": 4.989331436699858e-07,
255
+ "logits/chosen": -2.5194196701049805,
256
+ "logits/rejected": -2.7274131774902344,
257
+ "logps/chosen": -146.7877655029297,
258
+ "logps/rejected": -185.3484649658203,
259
+ "loss": 0.0171,
260
+ "rewards/accuracies": 0.987500011920929,
261
+ "rewards/chosen": -1.5904136896133423,
262
+ "rewards/margins": 8.3977632522583,
263
+ "rewards/rejected": -9.988175392150879,
264
+ "step": 160
265
+ },
266
+ {
267
+ "epoch": 0.11,
268
+ "grad_norm": 1.0748044802435313,
269
+ "learning_rate": 4.953769559032717e-07,
270
+ "logits/chosen": -2.478966474533081,
271
+ "logits/rejected": -2.706465005874634,
272
+ "logps/chosen": -123.1487045288086,
273
+ "logps/rejected": -187.28634643554688,
274
+ "loss": 0.0074,
275
+ "rewards/accuracies": 1.0,
276
+ "rewards/chosen": -0.411639928817749,
277
+ "rewards/margins": 9.865466117858887,
278
+ "rewards/rejected": -10.277105331420898,
279
+ "step": 170
280
+ },
281
+ {
282
+ "epoch": 0.12,
283
+ "grad_norm": 1.303203442204381,
284
+ "learning_rate": 4.918207681365576e-07,
285
+ "logits/chosen": -2.5271456241607666,
286
+ "logits/rejected": -2.7161953449249268,
287
+ "logps/chosen": -139.56097412109375,
288
+ "logps/rejected": -192.1598358154297,
289
+ "loss": 0.011,
290
+ "rewards/accuracies": 1.0,
291
+ "rewards/chosen": -1.4142029285430908,
292
+ "rewards/margins": 9.24979019165039,
293
+ "rewards/rejected": -10.663992881774902,
294
+ "step": 180
295
+ },
296
+ {
297
+ "epoch": 0.12,
298
+ "grad_norm": 0.8568137300390842,
299
+ "learning_rate": 4.882645803698435e-07,
300
+ "logits/chosen": -2.4333901405334473,
301
+ "logits/rejected": -2.620060682296753,
302
+ "logps/chosen": -140.07345581054688,
303
+ "logps/rejected": -180.193603515625,
304
+ "loss": 0.0123,
305
+ "rewards/accuracies": 1.0,
306
+ "rewards/chosen": -1.630637526512146,
307
+ "rewards/margins": 8.6264066696167,
308
+ "rewards/rejected": -10.257043838500977,
309
+ "step": 190
310
+ },
311
+ {
312
+ "epoch": 0.13,
313
+ "grad_norm": 13.412950576813955,
314
+ "learning_rate": 4.847083926031294e-07,
315
+ "logits/chosen": -2.461732864379883,
316
+ "logits/rejected": -2.675943374633789,
317
+ "logps/chosen": -137.39089965820312,
318
+ "logps/rejected": -195.60218811035156,
319
+ "loss": 0.0112,
320
+ "rewards/accuracies": 1.0,
321
+ "rewards/chosen": -0.9913924336433411,
322
+ "rewards/margins": 10.286381721496582,
323
+ "rewards/rejected": -11.277772903442383,
324
+ "step": 200
325
+ },
326
+ {
327
+ "epoch": 0.13,
328
+ "grad_norm": 21.7222990738424,
329
+ "learning_rate": 4.811522048364154e-07,
330
+ "logits/chosen": -2.5159268379211426,
331
+ "logits/rejected": -2.7606940269470215,
332
+ "logps/chosen": -162.57406616210938,
333
+ "logps/rejected": -213.39517211914062,
334
+ "loss": 0.0029,
335
+ "rewards/accuracies": 1.0,
336
+ "rewards/chosen": -0.11065955460071564,
337
+ "rewards/margins": 12.143023490905762,
338
+ "rewards/rejected": -12.253683090209961,
339
+ "step": 210
340
+ },
341
+ {
342
+ "epoch": 0.14,
343
+ "grad_norm": 0.046180032897506273,
344
+ "learning_rate": 4.775960170697012e-07,
345
+ "logits/chosen": -2.4487884044647217,
346
+ "logits/rejected": -2.664316415786743,
347
+ "logps/chosen": -129.77688598632812,
348
+ "logps/rejected": -209.32772827148438,
349
+ "loss": 0.0021,
350
+ "rewards/accuracies": 1.0,
351
+ "rewards/chosen": -0.11752267926931381,
352
+ "rewards/margins": 12.149419784545898,
353
+ "rewards/rejected": -12.266942977905273,
354
+ "step": 220
355
+ },
356
+ {
357
+ "epoch": 0.15,
358
+ "grad_norm": 1.4824700240745254,
359
+ "learning_rate": 4.7403982930298717e-07,
360
+ "logits/chosen": -2.4436874389648438,
361
+ "logits/rejected": -2.694324016571045,
362
+ "logps/chosen": -135.8723602294922,
363
+ "logps/rejected": -216.75387573242188,
364
+ "loss": 0.0057,
365
+ "rewards/accuracies": 1.0,
366
+ "rewards/chosen": -1.2215015888214111,
367
+ "rewards/margins": 11.901634216308594,
368
+ "rewards/rejected": -13.123135566711426,
369
+ "step": 230
370
+ },
371
+ {
372
+ "epoch": 0.15,
373
+ "grad_norm": 17.463403344491365,
374
+ "learning_rate": 4.7048364153627306e-07,
375
+ "logits/chosen": -2.4271080493927,
376
+ "logits/rejected": -2.691582441329956,
377
+ "logps/chosen": -156.45681762695312,
378
+ "logps/rejected": -219.70529174804688,
379
+ "loss": 0.0074,
380
+ "rewards/accuracies": 1.0,
381
+ "rewards/chosen": -2.2694718837738037,
382
+ "rewards/margins": 10.988931655883789,
383
+ "rewards/rejected": -13.258402824401855,
384
+ "step": 240
385
+ },
386
+ {
387
+ "epoch": 0.16,
388
+ "grad_norm": 2.8943147376404292,
389
+ "learning_rate": 4.66927453769559e-07,
390
+ "logits/chosen": -2.4071226119995117,
391
+ "logits/rejected": -2.6235671043395996,
392
+ "logps/chosen": -136.54635620117188,
393
+ "logps/rejected": -195.32444763183594,
394
+ "loss": 0.0033,
395
+ "rewards/accuracies": 1.0,
396
+ "rewards/chosen": -0.8205227851867676,
397
+ "rewards/margins": 10.77901840209961,
398
+ "rewards/rejected": -11.599540710449219,
399
+ "step": 250
400
+ },
401
+ {
402
+ "epoch": 0.17,
403
+ "grad_norm": 1.3617461125493031,
404
+ "learning_rate": 4.633712660028449e-07,
405
+ "logits/chosen": -2.439023971557617,
406
+ "logits/rejected": -2.676661729812622,
407
+ "logps/chosen": -132.08621215820312,
408
+ "logps/rejected": -209.95077514648438,
409
+ "loss": 0.0032,
410
+ "rewards/accuracies": 1.0,
411
+ "rewards/chosen": -0.34371912479400635,
412
+ "rewards/margins": 12.549942970275879,
413
+ "rewards/rejected": -12.893661499023438,
414
+ "step": 260
415
+ },
416
+ {
417
+ "epoch": 0.17,
418
+ "grad_norm": 1.0505376091417922,
419
+ "learning_rate": 4.5981507823613085e-07,
420
+ "logits/chosen": -2.4203948974609375,
421
+ "logits/rejected": -2.6439831256866455,
422
+ "logps/chosen": -143.63368225097656,
423
+ "logps/rejected": -215.7870330810547,
424
+ "loss": 0.0043,
425
+ "rewards/accuracies": 1.0,
426
+ "rewards/chosen": -1.3733088970184326,
427
+ "rewards/margins": 12.261556625366211,
428
+ "rewards/rejected": -13.634866714477539,
429
+ "step": 270
430
+ },
431
+ {
432
+ "epoch": 0.18,
433
+ "grad_norm": 0.30085985079958266,
434
+ "learning_rate": 4.562588904694168e-07,
435
+ "logits/chosen": -2.4349520206451416,
436
+ "logits/rejected": -2.671058177947998,
437
+ "logps/chosen": -138.3717498779297,
438
+ "logps/rejected": -223.06948852539062,
439
+ "loss": 0.0046,
440
+ "rewards/accuracies": 1.0,
441
+ "rewards/chosen": -0.9617056846618652,
442
+ "rewards/margins": 13.077425003051758,
443
+ "rewards/rejected": -14.039129257202148,
444
+ "step": 280
445
+ },
446
+ {
447
+ "epoch": 0.19,
448
+ "grad_norm": 1.459881483419176,
449
+ "learning_rate": 4.5270270270270264e-07,
450
+ "logits/chosen": -2.5198299884796143,
451
+ "logits/rejected": -2.6836562156677246,
452
+ "logps/chosen": -146.54750061035156,
453
+ "logps/rejected": -236.26171875,
454
+ "loss": 0.0045,
455
+ "rewards/accuracies": 1.0,
456
+ "rewards/chosen": -1.627685308456421,
457
+ "rewards/margins": 13.237665176391602,
458
+ "rewards/rejected": -14.865351676940918,
459
+ "step": 290
460
+ },
461
+ {
462
+ "epoch": 0.19,
463
+ "grad_norm": 0.10587938886796702,
464
+ "learning_rate": 4.491465149359886e-07,
465
+ "logits/chosen": -2.531820774078369,
466
+ "logits/rejected": -2.704437017440796,
467
+ "logps/chosen": -152.57876586914062,
468
+ "logps/rejected": -237.67868041992188,
469
+ "loss": 0.0017,
470
+ "rewards/accuracies": 1.0,
471
+ "rewards/chosen": -2.0694496631622314,
472
+ "rewards/margins": 12.964448928833008,
473
+ "rewards/rejected": -15.033899307250977,
474
+ "step": 300
475
+ },
476
+ {
477
+ "epoch": 0.2,
478
+ "grad_norm": 3.1066767259351065,
479
+ "learning_rate": 4.4559032716927454e-07,
480
+ "logits/chosen": -2.4913744926452637,
481
+ "logits/rejected": -2.7113704681396484,
482
+ "logps/chosen": -150.282958984375,
483
+ "logps/rejected": -228.7188262939453,
484
+ "loss": 0.009,
485
+ "rewards/accuracies": 1.0,
486
+ "rewards/chosen": -2.2832634449005127,
487
+ "rewards/margins": 12.166834831237793,
488
+ "rewards/rejected": -14.450098991394043,
489
+ "step": 310
490
+ },
491
+ {
492
+ "epoch": 0.2,
493
+ "grad_norm": 0.28682507599353846,
494
+ "learning_rate": 4.420341394025605e-07,
495
+ "logits/chosen": -2.3985180854797363,
496
+ "logits/rejected": -2.6870627403259277,
497
+ "logps/chosen": -160.65838623046875,
498
+ "logps/rejected": -225.05581665039062,
499
+ "loss": 0.0048,
500
+ "rewards/accuracies": 1.0,
501
+ "rewards/chosen": -3.4125454425811768,
502
+ "rewards/margins": 10.527777671813965,
503
+ "rewards/rejected": -13.940322875976562,
504
+ "step": 320
505
+ },
506
+ {
507
+ "epoch": 0.21,
508
+ "grad_norm": 5.225802985518444,
509
+ "learning_rate": 4.384779516358463e-07,
510
+ "logits/chosen": -2.39935040473938,
511
+ "logits/rejected": -2.6781036853790283,
512
+ "logps/chosen": -170.07534790039062,
513
+ "logps/rejected": -235.141357421875,
514
+ "loss": 0.0075,
515
+ "rewards/accuracies": 1.0,
516
+ "rewards/chosen": -3.992198944091797,
517
+ "rewards/margins": 11.101171493530273,
518
+ "rewards/rejected": -15.09337043762207,
519
+ "step": 330
520
+ },
521
+ {
522
+ "epoch": 0.22,
523
+ "grad_norm": 3.405010658068852,
524
+ "learning_rate": 4.3492176386913227e-07,
525
+ "logits/chosen": -2.3850109577178955,
526
+ "logits/rejected": -2.697495937347412,
527
+ "logps/chosen": -162.45657348632812,
528
+ "logps/rejected": -248.86428833007812,
529
+ "loss": 0.0024,
530
+ "rewards/accuracies": 1.0,
531
+ "rewards/chosen": -3.1875181198120117,
532
+ "rewards/margins": 13.499272346496582,
533
+ "rewards/rejected": -16.686786651611328,
534
+ "step": 340
535
+ },
536
+ {
537
+ "epoch": 0.22,
538
+ "grad_norm": 1.533659428300562,
539
+ "learning_rate": 4.313655761024182e-07,
540
+ "logits/chosen": -2.2804832458496094,
541
+ "logits/rejected": -2.6101315021514893,
542
+ "logps/chosen": -167.80911254882812,
543
+ "logps/rejected": -265.25482177734375,
544
+ "loss": 0.0057,
545
+ "rewards/accuracies": 1.0,
546
+ "rewards/chosen": -5.248920440673828,
547
+ "rewards/margins": 12.490852355957031,
548
+ "rewards/rejected": -17.73977279663086,
549
+ "step": 350
550
+ },
551
+ {
552
+ "epoch": 0.23,
553
+ "grad_norm": 1.8069977925253333,
554
+ "learning_rate": 4.278093883357041e-07,
555
+ "logits/chosen": -2.348785400390625,
556
+ "logits/rejected": -2.618760824203491,
557
+ "logps/chosen": -191.90760803222656,
558
+ "logps/rejected": -262.60369873046875,
559
+ "loss": 0.0023,
560
+ "rewards/accuracies": 1.0,
561
+ "rewards/chosen": -4.235640048980713,
562
+ "rewards/margins": 13.658535957336426,
563
+ "rewards/rejected": -17.894176483154297,
564
+ "step": 360
565
+ },
566
+ {
567
+ "epoch": 0.24,
568
+ "grad_norm": 1.1519087510764392,
569
+ "learning_rate": 4.2425320056899e-07,
570
+ "logits/chosen": -2.3367717266082764,
571
+ "logits/rejected": -2.6459882259368896,
572
+ "logps/chosen": -157.1217041015625,
573
+ "logps/rejected": -270.79559326171875,
574
+ "loss": 0.0015,
575
+ "rewards/accuracies": 1.0,
576
+ "rewards/chosen": -3.419424057006836,
577
+ "rewards/margins": 14.605476379394531,
578
+ "rewards/rejected": -18.024898529052734,
579
+ "step": 370
580
+ },
581
+ {
582
+ "epoch": 0.24,
583
+ "grad_norm": 0.12503486691431573,
584
+ "learning_rate": 4.2069701280227595e-07,
585
+ "logits/chosen": -2.2905983924865723,
586
+ "logits/rejected": -2.6188652515411377,
587
+ "logps/chosen": -161.6908416748047,
588
+ "logps/rejected": -277.34710693359375,
589
+ "loss": 0.0037,
590
+ "rewards/accuracies": 1.0,
591
+ "rewards/chosen": -4.108786582946777,
592
+ "rewards/margins": 14.594747543334961,
593
+ "rewards/rejected": -18.703533172607422,
594
+ "step": 380
595
+ },
596
+ {
597
+ "epoch": 0.25,
598
+ "grad_norm": 1.5481281251493877,
599
+ "learning_rate": 4.1714082503556185e-07,
600
+ "logits/chosen": -2.266719341278076,
601
+ "logits/rejected": -2.5488734245300293,
602
+ "logps/chosen": -176.82908630371094,
603
+ "logps/rejected": -274.1158447265625,
604
+ "loss": 0.0042,
605
+ "rewards/accuracies": 1.0,
606
+ "rewards/chosen": -6.371943950653076,
607
+ "rewards/margins": 12.869401931762695,
608
+ "rewards/rejected": -19.24134635925293,
609
+ "step": 390
610
+ },
611
+ {
612
+ "epoch": 0.26,
613
+ "grad_norm": 0.5072108881730257,
614
+ "learning_rate": 4.135846372688478e-07,
615
+ "logits/chosen": -2.29506254196167,
616
+ "logits/rejected": -2.585667133331299,
617
+ "logps/chosen": -166.04281616210938,
618
+ "logps/rejected": -287.45941162109375,
619
+ "loss": 0.001,
620
+ "rewards/accuracies": 1.0,
621
+ "rewards/chosen": -3.006166934967041,
622
+ "rewards/margins": 16.814212799072266,
623
+ "rewards/rejected": -19.820377349853516,
624
+ "step": 400
625
+ },
626
+ {
627
+ "epoch": 0.26,
628
+ "grad_norm": 14.57194509259927,
629
+ "learning_rate": 4.100284495021337e-07,
630
+ "logits/chosen": -2.1163532733917236,
631
+ "logits/rejected": -2.5131845474243164,
632
+ "logps/chosen": -142.27735900878906,
633
+ "logps/rejected": -264.2358703613281,
634
+ "loss": 0.0048,
635
+ "rewards/accuracies": 1.0,
636
+ "rewards/chosen": -2.5816712379455566,
637
+ "rewards/margins": 15.512285232543945,
638
+ "rewards/rejected": -18.093957901000977,
639
+ "step": 410
640
+ },
641
+ {
642
+ "epoch": 0.27,
643
+ "grad_norm": 14.91037080313429,
644
+ "learning_rate": 4.064722617354196e-07,
645
+ "logits/chosen": -2.1361889839172363,
646
+ "logits/rejected": -2.529857635498047,
647
+ "logps/chosen": -167.25698852539062,
648
+ "logps/rejected": -278.71826171875,
649
+ "loss": 0.0029,
650
+ "rewards/accuracies": 1.0,
651
+ "rewards/chosen": -2.816511631011963,
652
+ "rewards/margins": 16.648120880126953,
653
+ "rewards/rejected": -19.46463394165039,
654
+ "step": 420
655
+ },
656
+ {
657
+ "epoch": 0.28,
658
+ "grad_norm": 0.5762867478889705,
659
+ "learning_rate": 4.0291607396870553e-07,
660
+ "logits/chosen": -2.1066999435424805,
661
+ "logits/rejected": -2.456259250640869,
662
+ "logps/chosen": -163.7415008544922,
663
+ "logps/rejected": -305.5865783691406,
664
+ "loss": 0.0023,
665
+ "rewards/accuracies": 1.0,
666
+ "rewards/chosen": -5.017763137817383,
667
+ "rewards/margins": 17.317230224609375,
668
+ "rewards/rejected": -22.33499526977539,
669
+ "step": 430
670
+ },
671
+ {
672
+ "epoch": 0.28,
673
+ "grad_norm": 3.706715513864746,
674
+ "learning_rate": 3.993598862019915e-07,
675
+ "logits/chosen": -2.2640228271484375,
676
+ "logits/rejected": -2.544032096862793,
677
+ "logps/chosen": -191.6356964111328,
678
+ "logps/rejected": -306.35443115234375,
679
+ "loss": 0.002,
680
+ "rewards/accuracies": 1.0,
681
+ "rewards/chosen": -5.6402459144592285,
682
+ "rewards/margins": 16.29918670654297,
683
+ "rewards/rejected": -21.939434051513672,
684
+ "step": 440
685
+ },
686
+ {
687
+ "epoch": 0.29,
688
+ "grad_norm": 0.4884505823214583,
689
+ "learning_rate": 3.9580369843527737e-07,
690
+ "logits/chosen": -2.258357048034668,
691
+ "logits/rejected": -2.5526187419891357,
692
+ "logps/chosen": -217.8997039794922,
693
+ "logps/rejected": -372.7241516113281,
694
+ "loss": 0.0018,
695
+ "rewards/accuracies": 1.0,
696
+ "rewards/chosen": -9.196186065673828,
697
+ "rewards/margins": 19.7209529876709,
698
+ "rewards/rejected": -28.917139053344727,
699
+ "step": 450
700
+ },
701
+ {
702
+ "epoch": 0.29,
703
+ "grad_norm": 0.045591560325134,
704
+ "learning_rate": 3.9224751066856327e-07,
705
+ "logits/chosen": -1.883296251296997,
706
+ "logits/rejected": -2.3051819801330566,
707
+ "logps/chosen": -215.2701416015625,
708
+ "logps/rejected": -373.6488342285156,
709
+ "loss": 0.0023,
710
+ "rewards/accuracies": 1.0,
711
+ "rewards/chosen": -8.807767868041992,
712
+ "rewards/margins": 20.2095947265625,
713
+ "rewards/rejected": -29.017364501953125,
714
+ "step": 460
715
+ },
716
+ {
717
+ "epoch": 0.3,
718
+ "grad_norm": 0.12470490893983527,
719
+ "learning_rate": 3.886913229018492e-07,
720
+ "logits/chosen": -1.5328924655914307,
721
+ "logits/rejected": -2.0810158252716064,
722
+ "logps/chosen": -184.50637817382812,
723
+ "logps/rejected": -393.2000732421875,
724
+ "loss": 0.0017,
725
+ "rewards/accuracies": 1.0,
726
+ "rewards/chosen": -5.6133012771606445,
727
+ "rewards/margins": 24.581329345703125,
728
+ "rewards/rejected": -30.194625854492188,
729
+ "step": 470
730
+ },
731
+ {
732
+ "epoch": 0.31,
733
+ "grad_norm": 3.452822600566543,
734
+ "learning_rate": 3.851351351351351e-07,
735
+ "logits/chosen": -1.7886974811553955,
736
+ "logits/rejected": -2.2504382133483887,
737
+ "logps/chosen": -184.6663360595703,
738
+ "logps/rejected": -360.86578369140625,
739
+ "loss": 0.0015,
740
+ "rewards/accuracies": 1.0,
741
+ "rewards/chosen": -4.6061811447143555,
742
+ "rewards/margins": 22.705333709716797,
743
+ "rewards/rejected": -27.311513900756836,
744
+ "step": 480
745
+ },
746
+ {
747
+ "epoch": 0.31,
748
+ "grad_norm": 0.08494215992540369,
749
+ "learning_rate": 3.8157894736842105e-07,
750
+ "logits/chosen": -1.8703672885894775,
751
+ "logits/rejected": -2.233332872390747,
752
+ "logps/chosen": -182.80052185058594,
753
+ "logps/rejected": -391.44683837890625,
754
+ "loss": 0.0002,
755
+ "rewards/accuracies": 1.0,
756
+ "rewards/chosen": -3.9865665435791016,
757
+ "rewards/margins": 26.53458023071289,
758
+ "rewards/rejected": -30.521148681640625,
759
+ "step": 490
760
+ },
761
+ {
762
+ "epoch": 0.32,
763
+ "grad_norm": 0.4084224053393073,
764
+ "learning_rate": 3.7802275960170695e-07,
765
+ "logits/chosen": -1.906399130821228,
766
+ "logits/rejected": -2.272561550140381,
767
+ "logps/chosen": -149.4335479736328,
768
+ "logps/rejected": -365.5450439453125,
769
+ "loss": 0.0032,
770
+ "rewards/accuracies": 1.0,
771
+ "rewards/chosen": -2.793308734893799,
772
+ "rewards/margins": 25.208282470703125,
773
+ "rewards/rejected": -28.0015926361084,
774
+ "step": 500
775
+ },
776
+ {
777
+ "epoch": 0.33,
778
+ "grad_norm": 0.8910605143175322,
779
+ "learning_rate": 3.7446657183499284e-07,
780
+ "logits/chosen": -1.7485355138778687,
781
+ "logits/rejected": -2.077558755874634,
782
+ "logps/chosen": -166.2713623046875,
783
+ "logps/rejected": -369.4629821777344,
784
+ "loss": 0.0014,
785
+ "rewards/accuracies": 1.0,
786
+ "rewards/chosen": -3.770512342453003,
787
+ "rewards/margins": 24.904991149902344,
788
+ "rewards/rejected": -28.67550277709961,
789
+ "step": 510
790
+ },
791
+ {
792
+ "epoch": 0.33,
793
+ "grad_norm": 1.7685911190779986,
794
+ "learning_rate": 3.709103840682788e-07,
795
+ "logits/chosen": -1.7549102306365967,
796
+ "logits/rejected": -2.1138863563537598,
797
+ "logps/chosen": -170.63426208496094,
798
+ "logps/rejected": -404.83123779296875,
799
+ "loss": 0.0006,
800
+ "rewards/accuracies": 1.0,
801
+ "rewards/chosen": -4.2706170082092285,
802
+ "rewards/margins": 27.674823760986328,
803
+ "rewards/rejected": -31.9454402923584,
804
+ "step": 520
805
+ },
806
+ {
807
+ "epoch": 0.34,
808
+ "grad_norm": 4.243301477338939,
809
+ "learning_rate": 3.6735419630156474e-07,
810
+ "logits/chosen": -1.8727014064788818,
811
+ "logits/rejected": -2.1714062690734863,
812
+ "logps/chosen": -178.07815551757812,
813
+ "logps/rejected": -372.9120788574219,
814
+ "loss": 0.001,
815
+ "rewards/accuracies": 1.0,
816
+ "rewards/chosen": -3.480609893798828,
817
+ "rewards/margins": 25.141685485839844,
818
+ "rewards/rejected": -28.622295379638672,
819
+ "step": 530
820
+ },
821
+ {
822
+ "epoch": 0.35,
823
+ "grad_norm": 1.7633763298373188,
824
+ "learning_rate": 3.637980085348506e-07,
825
+ "logits/chosen": -2.259653091430664,
826
+ "logits/rejected": -2.5058536529541016,
827
+ "logps/chosen": -166.86276245117188,
828
+ "logps/rejected": -310.83026123046875,
829
+ "loss": 0.0087,
830
+ "rewards/accuracies": 1.0,
831
+ "rewards/chosen": -3.225189685821533,
832
+ "rewards/margins": 19.310260772705078,
833
+ "rewards/rejected": -22.535449981689453,
834
+ "step": 540
835
+ },
836
+ {
837
+ "epoch": 0.35,
838
+ "grad_norm": 0.5027205143527536,
839
+ "learning_rate": 3.602418207681365e-07,
840
+ "logits/chosen": -2.3244869709014893,
841
+ "logits/rejected": -2.5319793224334717,
842
+ "logps/chosen": -170.75979614257812,
843
+ "logps/rejected": -300.6156311035156,
844
+ "loss": 0.0019,
845
+ "rewards/accuracies": 1.0,
846
+ "rewards/chosen": -4.17783260345459,
847
+ "rewards/margins": 17.45343017578125,
848
+ "rewards/rejected": -21.631261825561523,
849
+ "step": 550
850
+ },
851
+ {
852
+ "epoch": 0.36,
853
+ "grad_norm": 15.947782616592031,
854
+ "learning_rate": 3.5668563300142247e-07,
855
+ "logits/chosen": -2.2984859943389893,
856
+ "logits/rejected": -2.4519248008728027,
857
+ "logps/chosen": -189.85989379882812,
858
+ "logps/rejected": -333.6773681640625,
859
+ "loss": 0.0039,
860
+ "rewards/accuracies": 1.0,
861
+ "rewards/chosen": -5.824889183044434,
862
+ "rewards/margins": 19.22607421875,
863
+ "rewards/rejected": -25.050960540771484,
864
+ "step": 560
865
+ },
866
+ {
867
+ "epoch": 0.36,
868
+ "grad_norm": 6.26719943945485,
869
+ "learning_rate": 3.5312944523470837e-07,
870
+ "logits/chosen": -2.2542061805725098,
871
+ "logits/rejected": -2.432546615600586,
872
+ "logps/chosen": -136.48843383789062,
873
+ "logps/rejected": -301.9817810058594,
874
+ "loss": 0.0023,
875
+ "rewards/accuracies": 1.0,
876
+ "rewards/chosen": -2.9576592445373535,
877
+ "rewards/margins": 19.037410736083984,
878
+ "rewards/rejected": -21.995067596435547,
879
+ "step": 570
880
+ },
881
+ {
882
+ "epoch": 0.37,
883
+ "grad_norm": 0.11415543320354572,
884
+ "learning_rate": 3.495732574679943e-07,
885
+ "logits/chosen": -2.215407609939575,
886
+ "logits/rejected": -2.4386544227600098,
887
+ "logps/chosen": -152.66677856445312,
888
+ "logps/rejected": -342.9513244628906,
889
+ "loss": 0.0008,
890
+ "rewards/accuracies": 1.0,
891
+ "rewards/chosen": -2.211902618408203,
892
+ "rewards/margins": 23.44979476928711,
893
+ "rewards/rejected": -25.661701202392578,
894
+ "step": 580
895
+ },
896
+ {
897
+ "epoch": 0.38,
898
+ "grad_norm": 0.22610570945942704,
899
+ "learning_rate": 3.460170697012802e-07,
900
+ "logits/chosen": -2.1291861534118652,
901
+ "logits/rejected": -2.409412384033203,
902
+ "logps/chosen": -160.59091186523438,
903
+ "logps/rejected": -361.2604064941406,
904
+ "loss": 0.0011,
905
+ "rewards/accuracies": 1.0,
906
+ "rewards/chosen": -4.499822616577148,
907
+ "rewards/margins": 22.70589828491211,
908
+ "rewards/rejected": -27.205724716186523,
909
+ "step": 590
910
+ },
911
+ {
912
+ "epoch": 0.38,
913
+ "grad_norm": 0.7406158964206918,
914
+ "learning_rate": 3.424608819345661e-07,
915
+ "logits/chosen": -2.1191792488098145,
916
+ "logits/rejected": -2.3479321002960205,
917
+ "logps/chosen": -197.30201721191406,
918
+ "logps/rejected": -371.0838317871094,
919
+ "loss": 0.0025,
920
+ "rewards/accuracies": 1.0,
921
+ "rewards/chosen": -6.3380818367004395,
922
+ "rewards/margins": 21.930988311767578,
923
+ "rewards/rejected": -28.269067764282227,
924
+ "step": 600
925
+ },
926
+ {
927
+ "epoch": 0.39,
928
+ "grad_norm": 0.32062082605886466,
929
+ "learning_rate": 3.3890469416785205e-07,
930
+ "logits/chosen": -2.0577409267425537,
931
+ "logits/rejected": -2.3198914527893066,
932
+ "logps/chosen": -195.90158081054688,
933
+ "logps/rejected": -384.8281555175781,
934
+ "loss": 0.0014,
935
+ "rewards/accuracies": 1.0,
936
+ "rewards/chosen": -5.974644660949707,
937
+ "rewards/margins": 23.86060905456543,
938
+ "rewards/rejected": -29.835254669189453,
939
+ "step": 610
940
+ },
941
+ {
942
+ "epoch": 0.4,
943
+ "grad_norm": 1.6177298432916143,
944
+ "learning_rate": 3.35348506401138e-07,
945
+ "logits/chosen": -1.9579404592514038,
946
+ "logits/rejected": -2.269507646560669,
947
+ "logps/chosen": -160.80618286132812,
948
+ "logps/rejected": -329.6241455078125,
949
+ "loss": 0.005,
950
+ "rewards/accuracies": 1.0,
951
+ "rewards/chosen": -4.555537223815918,
952
+ "rewards/margins": 20.43609619140625,
953
+ "rewards/rejected": -24.99163246154785,
954
+ "step": 620
955
+ },
956
+ {
957
+ "epoch": 0.4,
958
+ "grad_norm": 0.187358584161895,
959
+ "learning_rate": 3.3179231863442384e-07,
960
+ "logits/chosen": -2.113788366317749,
961
+ "logits/rejected": -2.3290324211120605,
962
+ "logps/chosen": -185.5271759033203,
963
+ "logps/rejected": -375.39453125,
964
+ "loss": 0.0027,
965
+ "rewards/accuracies": 1.0,
966
+ "rewards/chosen": -4.738511085510254,
967
+ "rewards/margins": 24.537811279296875,
968
+ "rewards/rejected": -29.276325225830078,
969
+ "step": 630
970
+ },
971
+ {
972
+ "epoch": 0.41,
973
+ "grad_norm": 0.010996930190111638,
974
+ "learning_rate": 3.282361308677098e-07,
975
+ "logits/chosen": -2.0769455432891846,
976
+ "logits/rejected": -2.347015142440796,
977
+ "logps/chosen": -175.67599487304688,
978
+ "logps/rejected": -431.5035705566406,
979
+ "loss": 0.0014,
980
+ "rewards/accuracies": 1.0,
981
+ "rewards/chosen": -4.6663079261779785,
982
+ "rewards/margins": 29.519245147705078,
983
+ "rewards/rejected": -34.18555450439453,
984
+ "step": 640
985
+ },
986
+ {
987
+ "epoch": 0.42,
988
+ "grad_norm": 0.21277301345388175,
989
+ "learning_rate": 3.2467994310099573e-07,
990
+ "logits/chosen": -2.198702335357666,
991
+ "logits/rejected": -2.4985084533691406,
992
+ "logps/chosen": -149.29171752929688,
993
+ "logps/rejected": -294.216552734375,
994
+ "loss": 0.0169,
995
+ "rewards/accuracies": 0.987500011920929,
996
+ "rewards/chosen": -3.3964667320251465,
997
+ "rewards/margins": 17.87206268310547,
998
+ "rewards/rejected": -21.26852798461914,
999
+ "step": 650
1000
+ },
1001
+ {
1002
+ "epoch": 0.42,
1003
+ "grad_norm": 0.02451780480832141,
1004
+ "learning_rate": 3.211237553342817e-07,
1005
+ "logits/chosen": -2.237781047821045,
1006
+ "logits/rejected": -2.509354591369629,
1007
+ "logps/chosen": -160.78524780273438,
1008
+ "logps/rejected": -333.3505554199219,
1009
+ "loss": 0.0009,
1010
+ "rewards/accuracies": 1.0,
1011
+ "rewards/chosen": -3.4144351482391357,
1012
+ "rewards/margins": 20.91831398010254,
1013
+ "rewards/rejected": -24.332752227783203,
1014
+ "step": 660
1015
+ },
1016
+ {
1017
+ "epoch": 0.43,
1018
+ "grad_norm": 0.5799365179235819,
1019
+ "learning_rate": 3.175675675675675e-07,
1020
+ "logits/chosen": -2.2056496143341064,
1021
+ "logits/rejected": -2.4211344718933105,
1022
+ "logps/chosen": -162.71096801757812,
1023
+ "logps/rejected": -334.28765869140625,
1024
+ "loss": 0.0024,
1025
+ "rewards/accuracies": 1.0,
1026
+ "rewards/chosen": -3.9299492835998535,
1027
+ "rewards/margins": 21.33687973022461,
1028
+ "rewards/rejected": -25.266826629638672,
1029
+ "step": 670
1030
+ },
1031
+ {
1032
+ "epoch": 0.44,
1033
+ "grad_norm": 0.1914899774779342,
1034
+ "learning_rate": 3.1401137980085347e-07,
1035
+ "logits/chosen": -2.2409019470214844,
1036
+ "logits/rejected": -2.518660068511963,
1037
+ "logps/chosen": -164.3127899169922,
1038
+ "logps/rejected": -349.6018371582031,
1039
+ "loss": 0.0045,
1040
+ "rewards/accuracies": 1.0,
1041
+ "rewards/chosen": -3.266920566558838,
1042
+ "rewards/margins": 22.99898338317871,
1043
+ "rewards/rejected": -26.265905380249023,
1044
+ "step": 680
1045
+ },
1046
+ {
1047
+ "epoch": 0.44,
1048
+ "grad_norm": 1.3949951852806233,
1049
+ "learning_rate": 3.104551920341394e-07,
1050
+ "logits/chosen": -2.3294060230255127,
1051
+ "logits/rejected": -2.461803674697876,
1052
+ "logps/chosen": -169.67840576171875,
1053
+ "logps/rejected": -362.62786865234375,
1054
+ "loss": 0.0009,
1055
+ "rewards/accuracies": 1.0,
1056
+ "rewards/chosen": -3.1314125061035156,
1057
+ "rewards/margins": 23.983592987060547,
1058
+ "rewards/rejected": -27.115009307861328,
1059
+ "step": 690
1060
+ },
1061
+ {
1062
+ "epoch": 0.45,
1063
+ "grad_norm": 0.5539714398553305,
1064
+ "learning_rate": 3.068990042674253e-07,
1065
+ "logits/chosen": -2.197174072265625,
1066
+ "logits/rejected": -2.44199275970459,
1067
+ "logps/chosen": -160.1177520751953,
1068
+ "logps/rejected": -375.48504638671875,
1069
+ "loss": 0.0013,
1070
+ "rewards/accuracies": 1.0,
1071
+ "rewards/chosen": -3.5987319946289062,
1072
+ "rewards/margins": 25.49329948425293,
1073
+ "rewards/rejected": -29.0920352935791,
1074
+ "step": 700
1075
+ },
1076
+ {
1077
+ "epoch": 0.45,
1078
+ "grad_norm": 0.06116625860123878,
1079
+ "learning_rate": 3.033428165007112e-07,
1080
+ "logits/chosen": -2.296790361404419,
1081
+ "logits/rejected": -2.4758524894714355,
1082
+ "logps/chosen": -194.8965301513672,
1083
+ "logps/rejected": -405.22930908203125,
1084
+ "loss": 0.0009,
1085
+ "rewards/accuracies": 1.0,
1086
+ "rewards/chosen": -3.60103178024292,
1087
+ "rewards/margins": 27.618661880493164,
1088
+ "rewards/rejected": -31.219696044921875,
1089
+ "step": 710
1090
+ },
1091
+ {
1092
+ "epoch": 0.46,
1093
+ "grad_norm": 0.101166393207417,
1094
+ "learning_rate": 2.9978662873399715e-07,
1095
+ "logits/chosen": -2.209062099456787,
1096
+ "logits/rejected": -2.4515223503112793,
1097
+ "logps/chosen": -161.6198272705078,
1098
+ "logps/rejected": -389.5179138183594,
1099
+ "loss": 0.0005,
1100
+ "rewards/accuracies": 1.0,
1101
+ "rewards/chosen": -4.203439235687256,
1102
+ "rewards/margins": 26.09065818786621,
1103
+ "rewards/rejected": -30.29409408569336,
1104
+ "step": 720
1105
+ },
1106
+ {
1107
+ "epoch": 0.47,
1108
+ "grad_norm": 0.12317613401454625,
1109
+ "learning_rate": 2.9623044096728305e-07,
1110
+ "logits/chosen": -2.3108105659484863,
1111
+ "logits/rejected": -2.4695842266082764,
1112
+ "logps/chosen": -167.56552124023438,
1113
+ "logps/rejected": -386.9646301269531,
1114
+ "loss": 0.0058,
1115
+ "rewards/accuracies": 1.0,
1116
+ "rewards/chosen": -3.244081974029541,
1117
+ "rewards/margins": 26.8917236328125,
1118
+ "rewards/rejected": -30.135807037353516,
1119
+ "step": 730
1120
+ },
1121
+ {
1122
+ "epoch": 0.47,
1123
+ "grad_norm": 0.16193920318148997,
1124
+ "learning_rate": 2.92674253200569e-07,
1125
+ "logits/chosen": -2.24593186378479,
1126
+ "logits/rejected": -2.3808069229125977,
1127
+ "logps/chosen": -166.65890502929688,
1128
+ "logps/rejected": -371.7357482910156,
1129
+ "loss": 0.0011,
1130
+ "rewards/accuracies": 1.0,
1131
+ "rewards/chosen": -4.039949417114258,
1132
+ "rewards/margins": 25.134775161743164,
1133
+ "rewards/rejected": -29.174724578857422,
1134
+ "step": 740
1135
+ },
1136
+ {
1137
+ "epoch": 0.48,
1138
+ "grad_norm": 0.24139035561225006,
1139
+ "learning_rate": 2.8911806543385494e-07,
1140
+ "logits/chosen": -2.2541756629943848,
1141
+ "logits/rejected": -2.4730725288391113,
1142
+ "logps/chosen": -153.8241729736328,
1143
+ "logps/rejected": -393.72454833984375,
1144
+ "loss": 0.0011,
1145
+ "rewards/accuracies": 1.0,
1146
+ "rewards/chosen": -3.2770607471466064,
1147
+ "rewards/margins": 27.69826889038086,
1148
+ "rewards/rejected": -30.975330352783203,
1149
+ "step": 750
1150
+ },
1151
+ {
1152
+ "epoch": 0.49,
1153
+ "grad_norm": 0.2273306730321036,
1154
+ "learning_rate": 2.855618776671408e-07,
1155
+ "logits/chosen": -2.2953999042510986,
1156
+ "logits/rejected": -2.4938910007476807,
1157
+ "logps/chosen": -156.7337646484375,
1158
+ "logps/rejected": -422.956298828125,
1159
+ "loss": 0.0009,
1160
+ "rewards/accuracies": 1.0,
1161
+ "rewards/chosen": -2.969212055206299,
1162
+ "rewards/margins": 31.034259796142578,
1163
+ "rewards/rejected": -34.00347137451172,
1164
+ "step": 760
1165
+ },
1166
+ {
1167
+ "epoch": 0.49,
1168
+ "grad_norm": 0.14372879524614457,
1169
+ "learning_rate": 2.8200568990042673e-07,
1170
+ "logits/chosen": -2.2638182640075684,
1171
+ "logits/rejected": -2.4787094593048096,
1172
+ "logps/chosen": -156.54861450195312,
1173
+ "logps/rejected": -413.94390869140625,
1174
+ "loss": 0.0003,
1175
+ "rewards/accuracies": 1.0,
1176
+ "rewards/chosen": -3.366678237915039,
1177
+ "rewards/margins": 29.10895347595215,
1178
+ "rewards/rejected": -32.47563552856445,
1179
+ "step": 770
1180
+ },
1181
+ {
1182
+ "epoch": 0.5,
1183
+ "grad_norm": 0.7791877496674509,
1184
+ "learning_rate": 2.784495021337127e-07,
1185
+ "logits/chosen": -2.0938005447387695,
1186
+ "logits/rejected": -2.353400468826294,
1187
+ "logps/chosen": -185.4242401123047,
1188
+ "logps/rejected": -443.72674560546875,
1189
+ "loss": 0.0024,
1190
+ "rewards/accuracies": 1.0,
1191
+ "rewards/chosen": -4.7681474685668945,
1192
+ "rewards/margins": 31.02960777282715,
1193
+ "rewards/rejected": -35.79775619506836,
1194
+ "step": 780
1195
+ },
1196
+ {
1197
+ "epoch": 0.51,
1198
+ "grad_norm": 0.03127157807702757,
1199
+ "learning_rate": 2.7489331436699857e-07,
1200
+ "logits/chosen": -1.8011245727539062,
1201
+ "logits/rejected": -2.1883206367492676,
1202
+ "logps/chosen": -211.3428192138672,
1203
+ "logps/rejected": -450.826904296875,
1204
+ "loss": 0.0059,
1205
+ "rewards/accuracies": 1.0,
1206
+ "rewards/chosen": -8.674280166625977,
1207
+ "rewards/margins": 28.067846298217773,
1208
+ "rewards/rejected": -36.74212646484375,
1209
+ "step": 790
1210
+ },
1211
+ {
1212
+ "epoch": 0.51,
1213
+ "grad_norm": 0.7457419151516147,
1214
+ "learning_rate": 2.7133712660028446e-07,
1215
+ "logits/chosen": -1.7626945972442627,
1216
+ "logits/rejected": -2.1461925506591797,
1217
+ "logps/chosen": -196.3053741455078,
1218
+ "logps/rejected": -414.73797607421875,
1219
+ "loss": 0.0023,
1220
+ "rewards/accuracies": 1.0,
1221
+ "rewards/chosen": -8.334815979003906,
1222
+ "rewards/margins": 25.075695037841797,
1223
+ "rewards/rejected": -33.41051483154297,
1224
+ "step": 800
1225
+ },
1226
+ {
1227
+ "epoch": 0.52,
1228
+ "grad_norm": 1.5620477401101704,
1229
+ "learning_rate": 2.677809388335704e-07,
1230
+ "logits/chosen": -1.8202106952667236,
1231
+ "logits/rejected": -2.1248602867126465,
1232
+ "logps/chosen": -214.88455200195312,
1233
+ "logps/rejected": -442.2652893066406,
1234
+ "loss": 0.0006,
1235
+ "rewards/accuracies": 1.0,
1236
+ "rewards/chosen": -8.491216659545898,
1237
+ "rewards/margins": 26.589365005493164,
1238
+ "rewards/rejected": -35.08058547973633,
1239
+ "step": 810
1240
+ },
1241
+ {
1242
+ "epoch": 0.52,
1243
+ "grad_norm": 0.08613623540760612,
1244
+ "learning_rate": 2.642247510668563e-07,
1245
+ "logits/chosen": -1.7793203592300415,
1246
+ "logits/rejected": -2.0444858074188232,
1247
+ "logps/chosen": -193.2935333251953,
1248
+ "logps/rejected": -417.64971923828125,
1249
+ "loss": 0.0004,
1250
+ "rewards/accuracies": 1.0,
1251
+ "rewards/chosen": -7.92196798324585,
1252
+ "rewards/margins": 26.353992462158203,
1253
+ "rewards/rejected": -34.275962829589844,
1254
+ "step": 820
1255
+ },
1256
+ {
1257
+ "epoch": 0.53,
1258
+ "grad_norm": 0.17621240034584026,
1259
+ "learning_rate": 2.6066856330014225e-07,
1260
+ "logits/chosen": -1.7755126953125,
1261
+ "logits/rejected": -2.158407688140869,
1262
+ "logps/chosen": -204.4994354248047,
1263
+ "logps/rejected": -403.656494140625,
1264
+ "loss": 0.0003,
1265
+ "rewards/accuracies": 1.0,
1266
+ "rewards/chosen": -8.456037521362305,
1267
+ "rewards/margins": 23.214176177978516,
1268
+ "rewards/rejected": -31.670211791992188,
1269
+ "step": 830
1270
+ },
1271
+ {
1272
+ "epoch": 0.54,
1273
+ "grad_norm": 0.2755963701991524,
1274
+ "learning_rate": 2.5711237553342815e-07,
1275
+ "logits/chosen": -1.8789972066879272,
1276
+ "logits/rejected": -2.1819803714752197,
1277
+ "logps/chosen": -228.1376190185547,
1278
+ "logps/rejected": -361.107421875,
1279
+ "loss": 0.0031,
1280
+ "rewards/accuracies": 1.0,
1281
+ "rewards/chosen": -8.826920509338379,
1282
+ "rewards/margins": 18.567317962646484,
1283
+ "rewards/rejected": -27.394237518310547,
1284
+ "step": 840
1285
+ },
1286
+ {
1287
+ "epoch": 0.54,
1288
+ "grad_norm": 0.3567356300171107,
1289
+ "learning_rate": 2.5355618776671404e-07,
1290
+ "logits/chosen": -1.8468620777130127,
1291
+ "logits/rejected": -2.1529593467712402,
1292
+ "logps/chosen": -211.9832000732422,
1293
+ "logps/rejected": -359.4861755371094,
1294
+ "loss": 0.003,
1295
+ "rewards/accuracies": 1.0,
1296
+ "rewards/chosen": -8.586137771606445,
1297
+ "rewards/margins": 18.8839054107666,
1298
+ "rewards/rejected": -27.470043182373047,
1299
+ "step": 850
1300
+ },
1301
+ {
1302
+ "epoch": 0.55,
1303
+ "grad_norm": 0.8886436285985593,
1304
+ "learning_rate": 2.5e-07,
1305
+ "logits/chosen": -1.8835853338241577,
1306
+ "logits/rejected": -2.2427430152893066,
1307
+ "logps/chosen": -199.7351531982422,
1308
+ "logps/rejected": -357.59173583984375,
1309
+ "loss": 0.0007,
1310
+ "rewards/accuracies": 1.0,
1311
+ "rewards/chosen": -8.047845840454102,
1312
+ "rewards/margins": 18.773927688598633,
1313
+ "rewards/rejected": -26.8217716217041,
1314
+ "step": 860
1315
+ },
1316
+ {
1317
+ "epoch": 0.56,
1318
+ "grad_norm": 9.219078186901797,
1319
+ "learning_rate": 2.4644381223328594e-07,
1320
+ "logits/chosen": -1.9670299291610718,
1321
+ "logits/rejected": -2.3183913230895996,
1322
+ "logps/chosen": -194.19821166992188,
1323
+ "logps/rejected": -351.28363037109375,
1324
+ "loss": 0.0036,
1325
+ "rewards/accuracies": 1.0,
1326
+ "rewards/chosen": -7.383183479309082,
1327
+ "rewards/margins": 18.614337921142578,
1328
+ "rewards/rejected": -25.997522354125977,
1329
+ "step": 870
1330
+ },
1331
+ {
1332
+ "epoch": 0.56,
1333
+ "grad_norm": 1.4090438099252132,
1334
+ "learning_rate": 2.4288762446657183e-07,
1335
+ "logits/chosen": -2.102022647857666,
1336
+ "logits/rejected": -2.330972194671631,
1337
+ "logps/chosen": -205.2625274658203,
1338
+ "logps/rejected": -360.0517578125,
1339
+ "loss": 0.0007,
1340
+ "rewards/accuracies": 1.0,
1341
+ "rewards/chosen": -7.618918418884277,
1342
+ "rewards/margins": 19.44923973083496,
1343
+ "rewards/rejected": -27.068157196044922,
1344
+ "step": 880
1345
+ },
1346
+ {
1347
+ "epoch": 0.57,
1348
+ "grad_norm": 0.18211470434359084,
1349
+ "learning_rate": 2.393314366998578e-07,
1350
+ "logits/chosen": -2.056673765182495,
1351
+ "logits/rejected": -2.251537322998047,
1352
+ "logps/chosen": -206.20236206054688,
1353
+ "logps/rejected": -347.0545349121094,
1354
+ "loss": 0.0008,
1355
+ "rewards/accuracies": 1.0,
1356
+ "rewards/chosen": -7.2250471115112305,
1357
+ "rewards/margins": 19.100948333740234,
1358
+ "rewards/rejected": -26.325998306274414,
1359
+ "step": 890
1360
+ },
1361
+ {
1362
+ "epoch": 0.58,
1363
+ "grad_norm": 0.07707203124683878,
1364
+ "learning_rate": 2.3577524893314365e-07,
1365
+ "logits/chosen": -2.017874002456665,
1366
+ "logits/rejected": -2.231964588165283,
1367
+ "logps/chosen": -202.12701416015625,
1368
+ "logps/rejected": -368.251953125,
1369
+ "loss": 0.0004,
1370
+ "rewards/accuracies": 1.0,
1371
+ "rewards/chosen": -6.979673862457275,
1372
+ "rewards/margins": 21.422286987304688,
1373
+ "rewards/rejected": -28.401958465576172,
1374
+ "step": 900
1375
+ },
1376
+ {
1377
+ "epoch": 0.58,
1378
+ "grad_norm": 0.001873680906220021,
1379
+ "learning_rate": 2.322190611664296e-07,
1380
+ "logits/chosen": -2.009934902191162,
1381
+ "logits/rejected": -2.166846752166748,
1382
+ "logps/chosen": -194.57254028320312,
1383
+ "logps/rejected": -352.115234375,
1384
+ "loss": 0.0004,
1385
+ "rewards/accuracies": 1.0,
1386
+ "rewards/chosen": -6.7056145668029785,
1387
+ "rewards/margins": 20.501380920410156,
1388
+ "rewards/rejected": -27.206995010375977,
1389
+ "step": 910
1390
+ },
1391
+ {
1392
+ "epoch": 0.59,
1393
+ "grad_norm": 0.08432204531455857,
1394
+ "learning_rate": 2.2866287339971549e-07,
1395
+ "logits/chosen": -2.018400192260742,
1396
+ "logits/rejected": -2.2199556827545166,
1397
+ "logps/chosen": -194.8164520263672,
1398
+ "logps/rejected": -383.21234130859375,
1399
+ "loss": 0.0028,
1400
+ "rewards/accuracies": 1.0,
1401
+ "rewards/chosen": -6.944446563720703,
1402
+ "rewards/margins": 22.65452003479004,
1403
+ "rewards/rejected": -29.598968505859375,
1404
+ "step": 920
1405
+ },
1406
+ {
1407
+ "epoch": 0.6,
1408
+ "grad_norm": 0.018002114098924,
1409
+ "learning_rate": 2.251066856330014e-07,
1410
+ "logits/chosen": -1.9886525869369507,
1411
+ "logits/rejected": -2.1241865158081055,
1412
+ "logps/chosen": -203.24600219726562,
1413
+ "logps/rejected": -367.25384521484375,
1414
+ "loss": 0.0004,
1415
+ "rewards/accuracies": 1.0,
1416
+ "rewards/chosen": -6.765786170959473,
1417
+ "rewards/margins": 22.038408279418945,
1418
+ "rewards/rejected": -28.8041934967041,
1419
+ "step": 930
1420
+ },
1421
+ {
1422
+ "epoch": 0.6,
1423
+ "grad_norm": 0.17969240498985034,
1424
+ "learning_rate": 2.2155049786628733e-07,
1425
+ "logits/chosen": -1.9080575704574585,
1426
+ "logits/rejected": -2.2484161853790283,
1427
+ "logps/chosen": -193.9615478515625,
1428
+ "logps/rejected": -385.5549011230469,
1429
+ "loss": 0.0006,
1430
+ "rewards/accuracies": 1.0,
1431
+ "rewards/chosen": -6.898443698883057,
1432
+ "rewards/margins": 22.909486770629883,
1433
+ "rewards/rejected": -29.807926177978516,
1434
+ "step": 940
1435
+ },
1436
+ {
1437
+ "epoch": 0.61,
1438
+ "grad_norm": 0.44454200648221737,
1439
+ "learning_rate": 2.1799431009957325e-07,
1440
+ "logits/chosen": -1.9811222553253174,
1441
+ "logits/rejected": -2.146716833114624,
1442
+ "logps/chosen": -183.6280975341797,
1443
+ "logps/rejected": -362.95428466796875,
1444
+ "loss": 0.0003,
1445
+ "rewards/accuracies": 1.0,
1446
+ "rewards/chosen": -6.245885372161865,
1447
+ "rewards/margins": 21.720762252807617,
1448
+ "rewards/rejected": -27.96664810180664,
1449
+ "step": 950
1450
+ },
1451
+ {
1452
+ "epoch": 0.61,
1453
+ "grad_norm": 0.040533997029042736,
1454
+ "learning_rate": 2.1443812233285914e-07,
1455
+ "logits/chosen": -1.913124680519104,
1456
+ "logits/rejected": -2.1230170726776123,
1457
+ "logps/chosen": -185.01473999023438,
1458
+ "logps/rejected": -370.7552490234375,
1459
+ "loss": 0.0032,
1460
+ "rewards/accuracies": 1.0,
1461
+ "rewards/chosen": -4.70156192779541,
1462
+ "rewards/margins": 23.934677124023438,
1463
+ "rewards/rejected": -28.636241912841797,
1464
+ "step": 960
1465
+ },
1466
+ {
1467
+ "epoch": 0.62,
1468
+ "grad_norm": 0.008272842079264974,
1469
+ "learning_rate": 2.108819345661451e-07,
1470
+ "logits/chosen": -1.9076120853424072,
1471
+ "logits/rejected": -2.0281999111175537,
1472
+ "logps/chosen": -170.41903686523438,
1473
+ "logps/rejected": -355.3038024902344,
1474
+ "loss": 0.0004,
1475
+ "rewards/accuracies": 1.0,
1476
+ "rewards/chosen": -4.106360912322998,
1477
+ "rewards/margins": 23.382038116455078,
1478
+ "rewards/rejected": -27.488399505615234,
1479
+ "step": 970
1480
+ },
1481
+ {
1482
+ "epoch": 0.63,
1483
+ "grad_norm": 0.12698409042008263,
1484
+ "learning_rate": 2.0732574679943098e-07,
1485
+ "logits/chosen": -1.7689754962921143,
1486
+ "logits/rejected": -2.0027101039886475,
1487
+ "logps/chosen": -151.05068969726562,
1488
+ "logps/rejected": -353.2601623535156,
1489
+ "loss": 0.0004,
1490
+ "rewards/accuracies": 1.0,
1491
+ "rewards/chosen": -4.083472728729248,
1492
+ "rewards/margins": 23.301097869873047,
1493
+ "rewards/rejected": -27.384571075439453,
1494
+ "step": 980
1495
+ },
1496
+ {
1497
+ "epoch": 0.63,
1498
+ "grad_norm": 0.526750803695656,
1499
+ "learning_rate": 2.0376955903271693e-07,
1500
+ "logits/chosen": -1.8657863140106201,
1501
+ "logits/rejected": -2.1982533931732178,
1502
+ "logps/chosen": -149.32810974121094,
1503
+ "logps/rejected": -375.81890869140625,
1504
+ "loss": 0.0006,
1505
+ "rewards/accuracies": 1.0,
1506
+ "rewards/chosen": -3.5236129760742188,
1507
+ "rewards/margins": 25.390033721923828,
1508
+ "rewards/rejected": -28.913650512695312,
1509
+ "step": 990
1510
+ },
1511
+ {
1512
+ "epoch": 0.64,
1513
+ "grad_norm": 0.4630571760419437,
1514
+ "learning_rate": 2.0021337126600283e-07,
1515
+ "logits/chosen": -1.8830486536026,
1516
+ "logits/rejected": -2.1774730682373047,
1517
+ "logps/chosen": -171.07569885253906,
1518
+ "logps/rejected": -375.9671630859375,
1519
+ "loss": 0.0001,
1520
+ "rewards/accuracies": 1.0,
1521
+ "rewards/chosen": -3.6209583282470703,
1522
+ "rewards/margins": 24.981908798217773,
1523
+ "rewards/rejected": -28.60286521911621,
1524
+ "step": 1000
1525
+ },
1526
+ {
1527
+ "epoch": 0.65,
1528
+ "grad_norm": 0.9404509604551488,
1529
+ "learning_rate": 1.9665718349928875e-07,
1530
+ "logits/chosen": -1.9226410388946533,
1531
+ "logits/rejected": -2.0701773166656494,
1532
+ "logps/chosen": -169.32447814941406,
1533
+ "logps/rejected": -370.9498596191406,
1534
+ "loss": 0.0061,
1535
+ "rewards/accuracies": 0.987500011920929,
1536
+ "rewards/chosen": -4.240044593811035,
1537
+ "rewards/margins": 24.28924560546875,
1538
+ "rewards/rejected": -28.5292911529541,
1539
+ "step": 1010
1540
+ },
1541
+ {
1542
+ "epoch": 0.65,
1543
+ "grad_norm": 0.372566229137264,
1544
+ "learning_rate": 1.931009957325747e-07,
1545
+ "logits/chosen": -1.871727705001831,
1546
+ "logits/rejected": -2.173356533050537,
1547
+ "logps/chosen": -175.205810546875,
1548
+ "logps/rejected": -347.47344970703125,
1549
+ "loss": 0.0006,
1550
+ "rewards/accuracies": 1.0,
1551
+ "rewards/chosen": -4.738024711608887,
1552
+ "rewards/margins": 21.8662052154541,
1553
+ "rewards/rejected": -26.604228973388672,
1554
+ "step": 1020
1555
+ },
1556
+ {
1557
+ "epoch": 0.66,
1558
+ "grad_norm": 0.33200076373892806,
1559
+ "learning_rate": 1.895448079658606e-07,
1560
+ "logits/chosen": -1.9309930801391602,
1561
+ "logits/rejected": -2.196063280105591,
1562
+ "logps/chosen": -179.72796630859375,
1563
+ "logps/rejected": -327.9047546386719,
1564
+ "loss": 0.0012,
1565
+ "rewards/accuracies": 1.0,
1566
+ "rewards/chosen": -4.597779273986816,
1567
+ "rewards/margins": 19.917034149169922,
1568
+ "rewards/rejected": -24.514814376831055,
1569
+ "step": 1030
1570
+ },
1571
+ {
1572
+ "epoch": 0.67,
1573
+ "grad_norm": 0.06316375390092686,
1574
+ "learning_rate": 1.859886201991465e-07,
1575
+ "logits/chosen": -1.8018014430999756,
1576
+ "logits/rejected": -2.151124954223633,
1577
+ "logps/chosen": -157.43798828125,
1578
+ "logps/rejected": -364.316162109375,
1579
+ "loss": 0.0002,
1580
+ "rewards/accuracies": 1.0,
1581
+ "rewards/chosen": -4.867424011230469,
1582
+ "rewards/margins": 23.382282257080078,
1583
+ "rewards/rejected": -28.249704360961914,
1584
+ "step": 1040
1585
+ },
1586
+ {
1587
+ "epoch": 0.67,
1588
+ "grad_norm": 0.017817721309851955,
1589
+ "learning_rate": 1.8243243243243243e-07,
1590
+ "logits/chosen": -1.8452457189559937,
1591
+ "logits/rejected": -2.215014696121216,
1592
+ "logps/chosen": -171.06100463867188,
1593
+ "logps/rejected": -375.04791259765625,
1594
+ "loss": 0.0004,
1595
+ "rewards/accuracies": 1.0,
1596
+ "rewards/chosen": -4.973246097564697,
1597
+ "rewards/margins": 23.576169967651367,
1598
+ "rewards/rejected": -28.549413681030273,
1599
+ "step": 1050
1600
+ },
1601
+ {
1602
+ "epoch": 0.68,
1603
+ "grad_norm": 0.06108555524800413,
1604
+ "learning_rate": 1.7887624466571835e-07,
1605
+ "logits/chosen": -1.908624291419983,
1606
+ "logits/rejected": -2.1829276084899902,
1607
+ "logps/chosen": -194.8101348876953,
1608
+ "logps/rejected": -388.7918395996094,
1609
+ "loss": 0.0028,
1610
+ "rewards/accuracies": 1.0,
1611
+ "rewards/chosen": -6.008497714996338,
1612
+ "rewards/margins": 24.203271865844727,
1613
+ "rewards/rejected": -30.21177101135254,
1614
+ "step": 1060
1615
+ },
1616
+ {
1617
+ "epoch": 0.68,
1618
+ "grad_norm": 0.18272607275595262,
1619
+ "learning_rate": 1.7532005689900424e-07,
1620
+ "logits/chosen": -1.9170128107070923,
1621
+ "logits/rejected": -2.2122740745544434,
1622
+ "logps/chosen": -179.13600158691406,
1623
+ "logps/rejected": -377.26361083984375,
1624
+ "loss": 0.008,
1625
+ "rewards/accuracies": 1.0,
1626
+ "rewards/chosen": -6.440054416656494,
1627
+ "rewards/margins": 22.63031768798828,
1628
+ "rewards/rejected": -29.07037353515625,
1629
+ "step": 1070
1630
+ },
1631
+ {
1632
+ "epoch": 0.69,
1633
+ "grad_norm": 0.05453926680821035,
1634
+ "learning_rate": 1.717638691322902e-07,
1635
+ "logits/chosen": -1.9766355752944946,
1636
+ "logits/rejected": -2.228825569152832,
1637
+ "logps/chosen": -205.33908081054688,
1638
+ "logps/rejected": -388.71685791015625,
1639
+ "loss": 0.0003,
1640
+ "rewards/accuracies": 1.0,
1641
+ "rewards/chosen": -6.267792701721191,
1642
+ "rewards/margins": 23.473896026611328,
1643
+ "rewards/rejected": -29.741689682006836,
1644
+ "step": 1080
1645
+ },
1646
+ {
1647
+ "epoch": 0.7,
1648
+ "grad_norm": 8.450138043857905,
1649
+ "learning_rate": 1.6820768136557609e-07,
1650
+ "logits/chosen": -1.929616928100586,
1651
+ "logits/rejected": -2.1833229064941406,
1652
+ "logps/chosen": -177.30699157714844,
1653
+ "logps/rejected": -369.02569580078125,
1654
+ "loss": 0.0007,
1655
+ "rewards/accuracies": 1.0,
1656
+ "rewards/chosen": -5.677231311798096,
1657
+ "rewards/margins": 23.399320602416992,
1658
+ "rewards/rejected": -29.076553344726562,
1659
+ "step": 1090
1660
+ },
1661
+ {
1662
+ "epoch": 0.7,
1663
+ "grad_norm": 0.2629017265227386,
1664
+ "learning_rate": 1.64651493598862e-07,
1665
+ "logits/chosen": -1.8782516717910767,
1666
+ "logits/rejected": -2.1519296169281006,
1667
+ "logps/chosen": -171.37185668945312,
1668
+ "logps/rejected": -345.8654479980469,
1669
+ "loss": 0.0003,
1670
+ "rewards/accuracies": 1.0,
1671
+ "rewards/chosen": -5.002556800842285,
1672
+ "rewards/margins": 21.3153018951416,
1673
+ "rewards/rejected": -26.317859649658203,
1674
+ "step": 1100
1675
+ },
1676
+ {
1677
+ "epoch": 0.71,
1678
+ "grad_norm": 0.08681759669643914,
1679
+ "learning_rate": 1.6109530583214793e-07,
1680
+ "logits/chosen": -1.944676160812378,
1681
+ "logits/rejected": -2.2825026512145996,
1682
+ "logps/chosen": -172.86749267578125,
1683
+ "logps/rejected": -348.4911804199219,
1684
+ "loss": 0.0023,
1685
+ "rewards/accuracies": 1.0,
1686
+ "rewards/chosen": -5.514270305633545,
1687
+ "rewards/margins": 20.45355987548828,
1688
+ "rewards/rejected": -25.967830657958984,
1689
+ "step": 1110
1690
+ },
1691
+ {
1692
+ "epoch": 0.72,
1693
+ "grad_norm": 0.14772131857638982,
1694
+ "learning_rate": 1.5753911806543385e-07,
1695
+ "logits/chosen": -2.047938823699951,
1696
+ "logits/rejected": -2.198514938354492,
1697
+ "logps/chosen": -188.20437622070312,
1698
+ "logps/rejected": -351.10565185546875,
1699
+ "loss": 0.0007,
1700
+ "rewards/accuracies": 1.0,
1701
+ "rewards/chosen": -5.5845184326171875,
1702
+ "rewards/margins": 21.684741973876953,
1703
+ "rewards/rejected": -27.26926040649414,
1704
+ "step": 1120
1705
+ },
1706
+ {
1707
+ "epoch": 0.72,
1708
+ "grad_norm": 0.020583492505652848,
1709
+ "learning_rate": 1.5398293029871974e-07,
1710
+ "logits/chosen": -2.0079150199890137,
1711
+ "logits/rejected": -2.2482612133026123,
1712
+ "logps/chosen": -177.460205078125,
1713
+ "logps/rejected": -346.9248046875,
1714
+ "loss": 0.0002,
1715
+ "rewards/accuracies": 1.0,
1716
+ "rewards/chosen": -5.606606960296631,
1717
+ "rewards/margins": 20.772077560424805,
1718
+ "rewards/rejected": -26.378686904907227,
1719
+ "step": 1130
1720
+ },
1721
+ {
1722
+ "epoch": 0.73,
1723
+ "grad_norm": 1.7059121905673134,
1724
+ "learning_rate": 1.504267425320057e-07,
1725
+ "logits/chosen": -1.9664523601531982,
1726
+ "logits/rejected": -2.1823642253875732,
1727
+ "logps/chosen": -174.95205688476562,
1728
+ "logps/rejected": -388.5026550292969,
1729
+ "loss": 0.0021,
1730
+ "rewards/accuracies": 1.0,
1731
+ "rewards/chosen": -6.427298069000244,
1732
+ "rewards/margins": 23.900110244750977,
1733
+ "rewards/rejected": -30.327407836914062,
1734
+ "step": 1140
1735
+ },
1736
+ {
1737
+ "epoch": 0.74,
1738
+ "grad_norm": 0.9977733695628042,
1739
+ "learning_rate": 1.4687055476529158e-07,
1740
+ "logits/chosen": -2.050489664077759,
1741
+ "logits/rejected": -2.270904064178467,
1742
+ "logps/chosen": -186.78663635253906,
1743
+ "logps/rejected": -372.7840270996094,
1744
+ "loss": 0.0002,
1745
+ "rewards/accuracies": 1.0,
1746
+ "rewards/chosen": -5.426581382751465,
1747
+ "rewards/margins": 23.374040603637695,
1748
+ "rewards/rejected": -28.800622940063477,
1749
+ "step": 1150
1750
+ },
1751
+ {
1752
+ "epoch": 0.74,
1753
+ "grad_norm": 0.002404789660626006,
1754
+ "learning_rate": 1.4331436699857753e-07,
1755
+ "logits/chosen": -1.9210751056671143,
1756
+ "logits/rejected": -2.240872859954834,
1757
+ "logps/chosen": -171.84713745117188,
1758
+ "logps/rejected": -396.109375,
1759
+ "loss": 0.0,
1760
+ "rewards/accuracies": 1.0,
1761
+ "rewards/chosen": -5.953668117523193,
1762
+ "rewards/margins": 25.2002010345459,
1763
+ "rewards/rejected": -31.15386962890625,
1764
+ "step": 1160
1765
+ },
1766
+ {
1767
+ "epoch": 0.75,
1768
+ "grad_norm": 0.018292107983041178,
1769
+ "learning_rate": 1.3975817923186345e-07,
1770
+ "logits/chosen": -2.0512757301330566,
1771
+ "logits/rejected": -2.261383533477783,
1772
+ "logps/chosen": -198.47787475585938,
1773
+ "logps/rejected": -409.4540710449219,
1774
+ "loss": 0.0001,
1775
+ "rewards/accuracies": 1.0,
1776
+ "rewards/chosen": -5.596243858337402,
1777
+ "rewards/margins": 26.373615264892578,
1778
+ "rewards/rejected": -31.969860076904297,
1779
+ "step": 1170
1780
+ },
1781
+ {
1782
+ "epoch": 0.75,
1783
+ "grad_norm": 8.705213421112179,
1784
+ "learning_rate": 1.3620199146514935e-07,
1785
+ "logits/chosen": -1.9703617095947266,
1786
+ "logits/rejected": -2.2260239124298096,
1787
+ "logps/chosen": -185.51535034179688,
1788
+ "logps/rejected": -381.4412536621094,
1789
+ "loss": 0.0026,
1790
+ "rewards/accuracies": 1.0,
1791
+ "rewards/chosen": -5.784358978271484,
1792
+ "rewards/margins": 23.76874542236328,
1793
+ "rewards/rejected": -29.5531063079834,
1794
+ "step": 1180
1795
+ },
1796
+ {
1797
+ "epoch": 0.76,
1798
+ "grad_norm": 1.1752034294380853,
1799
+ "learning_rate": 1.326458036984353e-07,
1800
+ "logits/chosen": -2.005053758621216,
1801
+ "logits/rejected": -2.2459633350372314,
1802
+ "logps/chosen": -178.1356964111328,
1803
+ "logps/rejected": -355.98211669921875,
1804
+ "loss": 0.0006,
1805
+ "rewards/accuracies": 1.0,
1806
+ "rewards/chosen": -5.596381664276123,
1807
+ "rewards/margins": 21.09072494506836,
1808
+ "rewards/rejected": -26.687108993530273,
1809
+ "step": 1190
1810
+ },
1811
+ {
1812
+ "epoch": 0.77,
1813
+ "grad_norm": 0.05160111703860194,
1814
+ "learning_rate": 1.290896159317212e-07,
1815
+ "logits/chosen": -2.0035533905029297,
1816
+ "logits/rejected": -2.2457115650177,
1817
+ "logps/chosen": -178.15768432617188,
1818
+ "logps/rejected": -336.3076171875,
1819
+ "loss": 0.0005,
1820
+ "rewards/accuracies": 1.0,
1821
+ "rewards/chosen": -5.5091423988342285,
1822
+ "rewards/margins": 19.467090606689453,
1823
+ "rewards/rejected": -24.97623634338379,
1824
+ "step": 1200
1825
+ },
1826
+ {
1827
+ "epoch": 0.77,
1828
+ "grad_norm": 0.061931926760717695,
1829
+ "learning_rate": 1.255334281650071e-07,
1830
+ "logits/chosen": -2.0695555210113525,
1831
+ "logits/rejected": -2.266425132751465,
1832
+ "logps/chosen": -189.8621826171875,
1833
+ "logps/rejected": -356.81158447265625,
1834
+ "loss": 0.0001,
1835
+ "rewards/accuracies": 1.0,
1836
+ "rewards/chosen": -5.535902500152588,
1837
+ "rewards/margins": 21.406814575195312,
1838
+ "rewards/rejected": -26.94271469116211,
1839
+ "step": 1210
1840
+ },
1841
+ {
1842
+ "epoch": 0.78,
1843
+ "grad_norm": 0.13621183527581743,
1844
+ "learning_rate": 1.2197724039829303e-07,
1845
+ "logits/chosen": -2.019291877746582,
1846
+ "logits/rejected": -2.1981494426727295,
1847
+ "logps/chosen": -187.91079711914062,
1848
+ "logps/rejected": -356.6788330078125,
1849
+ "loss": 0.0002,
1850
+ "rewards/accuracies": 1.0,
1851
+ "rewards/chosen": -6.211094856262207,
1852
+ "rewards/margins": 21.122692108154297,
1853
+ "rewards/rejected": -27.333786010742188,
1854
+ "step": 1220
1855
+ },
1856
+ {
1857
+ "epoch": 0.79,
1858
+ "grad_norm": 0.013930629258428272,
1859
+ "learning_rate": 1.1842105263157894e-07,
1860
+ "logits/chosen": -1.9782531261444092,
1861
+ "logits/rejected": -2.2405457496643066,
1862
+ "logps/chosen": -169.53475952148438,
1863
+ "logps/rejected": -377.62298583984375,
1864
+ "loss": 0.0001,
1865
+ "rewards/accuracies": 1.0,
1866
+ "rewards/chosen": -5.386349678039551,
1867
+ "rewards/margins": 23.819812774658203,
1868
+ "rewards/rejected": -29.206165313720703,
1869
+ "step": 1230
1870
+ },
1871
+ {
1872
+ "epoch": 0.79,
1873
+ "grad_norm": 0.63025336913527,
1874
+ "learning_rate": 1.1486486486486487e-07,
1875
+ "logits/chosen": -1.9963802099227905,
1876
+ "logits/rejected": -2.239816427230835,
1877
+ "logps/chosen": -163.4903106689453,
1878
+ "logps/rejected": -334.11529541015625,
1879
+ "loss": 0.0005,
1880
+ "rewards/accuracies": 1.0,
1881
+ "rewards/chosen": -3.6800830364227295,
1882
+ "rewards/margins": 21.3232479095459,
1883
+ "rewards/rejected": -25.003332138061523,
1884
+ "step": 1240
1885
+ },
1886
+ {
1887
+ "epoch": 0.8,
1888
+ "grad_norm": 0.055495248029239844,
1889
+ "learning_rate": 1.1130867709815078e-07,
1890
+ "logits/chosen": -1.9950904846191406,
1891
+ "logits/rejected": -2.1991913318634033,
1892
+ "logps/chosen": -180.9514923095703,
1893
+ "logps/rejected": -378.4940185546875,
1894
+ "loss": 0.0002,
1895
+ "rewards/accuracies": 1.0,
1896
+ "rewards/chosen": -5.002464771270752,
1897
+ "rewards/margins": 24.19596290588379,
1898
+ "rewards/rejected": -29.19842529296875,
1899
+ "step": 1250
1900
+ },
1901
+ {
1902
+ "epoch": 0.81,
1903
+ "grad_norm": 0.23651663696164235,
1904
+ "learning_rate": 1.077524893314367e-07,
1905
+ "logits/chosen": -2.0281214714050293,
1906
+ "logits/rejected": -2.173586368560791,
1907
+ "logps/chosen": -199.83065795898438,
1908
+ "logps/rejected": -370.74591064453125,
1909
+ "loss": 0.0001,
1910
+ "rewards/accuracies": 1.0,
1911
+ "rewards/chosen": -6.204631328582764,
1912
+ "rewards/margins": 23.164012908935547,
1913
+ "rewards/rejected": -29.3686466217041,
1914
+ "step": 1260
1915
+ },
1916
+ {
1917
+ "epoch": 0.81,
1918
+ "grad_norm": 87.17809979841759,
1919
+ "learning_rate": 1.0419630156472262e-07,
1920
+ "logits/chosen": -1.8399674892425537,
1921
+ "logits/rejected": -2.0879006385803223,
1922
+ "logps/chosen": -224.9874267578125,
1923
+ "logps/rejected": -438.8709411621094,
1924
+ "loss": 0.0084,
1925
+ "rewards/accuracies": 1.0,
1926
+ "rewards/chosen": -10.09913444519043,
1927
+ "rewards/margins": 24.833473205566406,
1928
+ "rewards/rejected": -34.9326057434082,
1929
+ "step": 1270
1930
+ },
1931
+ {
1932
+ "epoch": 0.82,
1933
+ "grad_norm": 0.03529290404850911,
1934
+ "learning_rate": 1.0064011379800854e-07,
1935
+ "logits/chosen": -2.116236686706543,
1936
+ "logits/rejected": -2.2769484519958496,
1937
+ "logps/chosen": -199.0480499267578,
1938
+ "logps/rejected": -381.0927734375,
1939
+ "loss": 0.0001,
1940
+ "rewards/accuracies": 1.0,
1941
+ "rewards/chosen": -6.298292636871338,
1942
+ "rewards/margins": 23.004474639892578,
1943
+ "rewards/rejected": -29.30276870727539,
1944
+ "step": 1280
1945
+ },
1946
+ {
1947
+ "epoch": 0.83,
1948
+ "grad_norm": 0.01781110705857936,
1949
+ "learning_rate": 9.708392603129445e-08,
1950
+ "logits/chosen": -2.098355770111084,
1951
+ "logits/rejected": -2.3084280490875244,
1952
+ "logps/chosen": -185.13162231445312,
1953
+ "logps/rejected": -364.68902587890625,
1954
+ "loss": 0.0006,
1955
+ "rewards/accuracies": 1.0,
1956
+ "rewards/chosen": -6.411550998687744,
1957
+ "rewards/margins": 21.45414924621582,
1958
+ "rewards/rejected": -27.865697860717773,
1959
+ "step": 1290
1960
+ },
1961
+ {
1962
+ "epoch": 0.83,
1963
+ "grad_norm": 0.21380541746917897,
1964
+ "learning_rate": 9.352773826458037e-08,
1965
+ "logits/chosen": -2.1513562202453613,
1966
+ "logits/rejected": -2.315798282623291,
1967
+ "logps/chosen": -192.1603546142578,
1968
+ "logps/rejected": -370.1746520996094,
1969
+ "loss": 0.0004,
1970
+ "rewards/accuracies": 1.0,
1971
+ "rewards/chosen": -6.151523590087891,
1972
+ "rewards/margins": 22.305927276611328,
1973
+ "rewards/rejected": -28.457448959350586,
1974
+ "step": 1300
1975
+ },
1976
+ {
1977
+ "epoch": 0.84,
1978
+ "grad_norm": 0.053698061298588354,
1979
+ "learning_rate": 8.997155049786629e-08,
1980
+ "logits/chosen": -2.152535915374756,
1981
+ "logits/rejected": -2.2665324211120605,
1982
+ "logps/chosen": -185.61489868164062,
1983
+ "logps/rejected": -373.6748962402344,
1984
+ "loss": 0.0001,
1985
+ "rewards/accuracies": 1.0,
1986
+ "rewards/chosen": -6.086173057556152,
1987
+ "rewards/margins": 22.81212615966797,
1988
+ "rewards/rejected": -28.898296356201172,
1989
+ "step": 1310
1990
+ },
1991
+ {
1992
+ "epoch": 0.84,
1993
+ "grad_norm": 0.21920871800240746,
1994
+ "learning_rate": 8.64153627311522e-08,
1995
+ "logits/chosen": -2.11517333984375,
1996
+ "logits/rejected": -2.210675001144409,
1997
+ "logps/chosen": -184.41207885742188,
1998
+ "logps/rejected": -339.87030029296875,
1999
+ "loss": 0.0004,
2000
+ "rewards/accuracies": 1.0,
2001
+ "rewards/chosen": -5.820789337158203,
2002
+ "rewards/margins": 20.895233154296875,
2003
+ "rewards/rejected": -26.71602439880371,
2004
+ "step": 1320
2005
+ },
2006
+ {
2007
+ "epoch": 0.85,
2008
+ "grad_norm": 0.09090747369415979,
2009
+ "learning_rate": 8.285917496443812e-08,
2010
+ "logits/chosen": -2.1557207107543945,
2011
+ "logits/rejected": -2.3053669929504395,
2012
+ "logps/chosen": -193.05734252929688,
2013
+ "logps/rejected": -388.4326171875,
2014
+ "loss": 0.0021,
2015
+ "rewards/accuracies": 1.0,
2016
+ "rewards/chosen": -5.772280216217041,
2017
+ "rewards/margins": 23.938325881958008,
2018
+ "rewards/rejected": -29.71060562133789,
2019
+ "step": 1330
2020
+ },
2021
+ {
2022
+ "epoch": 0.86,
2023
+ "grad_norm": 0.018865859918979065,
2024
+ "learning_rate": 7.930298719772404e-08,
2025
+ "logits/chosen": -2.1138229370117188,
2026
+ "logits/rejected": -2.251180648803711,
2027
+ "logps/chosen": -168.01907348632812,
2028
+ "logps/rejected": -391.6059265136719,
2029
+ "loss": 0.0009,
2030
+ "rewards/accuracies": 1.0,
2031
+ "rewards/chosen": -5.083653450012207,
2032
+ "rewards/margins": 25.798690795898438,
2033
+ "rewards/rejected": -30.882349014282227,
2034
+ "step": 1340
2035
+ },
2036
+ {
2037
+ "epoch": 0.86,
2038
+ "grad_norm": 0.015984656139045435,
2039
+ "learning_rate": 7.574679943100994e-08,
2040
+ "logits/chosen": -2.174654722213745,
2041
+ "logits/rejected": -2.2935032844543457,
2042
+ "logps/chosen": -176.84854125976562,
2043
+ "logps/rejected": -357.89007568359375,
2044
+ "loss": 0.0001,
2045
+ "rewards/accuracies": 1.0,
2046
+ "rewards/chosen": -5.099386215209961,
2047
+ "rewards/margins": 22.505481719970703,
2048
+ "rewards/rejected": -27.604867935180664,
2049
+ "step": 1350
2050
+ },
2051
+ {
2052
+ "epoch": 0.87,
2053
+ "grad_norm": 0.5463133360618999,
2054
+ "learning_rate": 7.219061166429587e-08,
2055
+ "logits/chosen": -2.146613359451294,
2056
+ "logits/rejected": -2.340137004852295,
2057
+ "logps/chosen": -194.63243103027344,
2058
+ "logps/rejected": -398.6822204589844,
2059
+ "loss": 0.0001,
2060
+ "rewards/accuracies": 1.0,
2061
+ "rewards/chosen": -5.568980693817139,
2062
+ "rewards/margins": 25.14784812927246,
2063
+ "rewards/rejected": -30.71683120727539,
2064
+ "step": 1360
2065
+ },
2066
+ {
2067
+ "epoch": 0.88,
2068
+ "grad_norm": 0.5871369349452629,
2069
+ "learning_rate": 6.863442389758179e-08,
2070
+ "logits/chosen": -2.031893491744995,
2071
+ "logits/rejected": -2.307396411895752,
2072
+ "logps/chosen": -166.64334106445312,
2073
+ "logps/rejected": -362.17724609375,
2074
+ "loss": 0.0001,
2075
+ "rewards/accuracies": 1.0,
2076
+ "rewards/chosen": -5.520142555236816,
2077
+ "rewards/margins": 21.91988754272461,
2078
+ "rewards/rejected": -27.440032958984375,
2079
+ "step": 1370
2080
+ },
2081
+ {
2082
+ "epoch": 0.88,
2083
+ "grad_norm": 0.001151297675766669,
2084
+ "learning_rate": 6.507823613086771e-08,
2085
+ "logits/chosen": -2.0853652954101562,
2086
+ "logits/rejected": -2.2990689277648926,
2087
+ "logps/chosen": -166.94778442382812,
2088
+ "logps/rejected": -373.9269104003906,
2089
+ "loss": 0.0001,
2090
+ "rewards/accuracies": 1.0,
2091
+ "rewards/chosen": -5.408658981323242,
2092
+ "rewards/margins": 23.76215362548828,
2093
+ "rewards/rejected": -29.170812606811523,
2094
+ "step": 1380
2095
+ },
2096
+ {
2097
+ "epoch": 0.89,
2098
+ "grad_norm": 0.09711906493171378,
2099
+ "learning_rate": 6.152204836415363e-08,
2100
+ "logits/chosen": -2.073453426361084,
2101
+ "logits/rejected": -2.21116304397583,
2102
+ "logps/chosen": -171.67530822753906,
2103
+ "logps/rejected": -370.3066101074219,
2104
+ "loss": 0.0001,
2105
+ "rewards/accuracies": 1.0,
2106
+ "rewards/chosen": -5.4291462898254395,
2107
+ "rewards/margins": 23.211612701416016,
2108
+ "rewards/rejected": -28.640756607055664,
2109
+ "step": 1390
2110
+ },
2111
+ {
2112
+ "epoch": 0.9,
2113
+ "grad_norm": 0.12505838228275154,
2114
+ "learning_rate": 5.796586059743954e-08,
2115
+ "logits/chosen": -2.0895724296569824,
2116
+ "logits/rejected": -2.207096576690674,
2117
+ "logps/chosen": -163.5934600830078,
2118
+ "logps/rejected": -365.29693603515625,
2119
+ "loss": 0.0001,
2120
+ "rewards/accuracies": 1.0,
2121
+ "rewards/chosen": -5.118988990783691,
2122
+ "rewards/margins": 23.092578887939453,
2123
+ "rewards/rejected": -28.21156883239746,
2124
+ "step": 1400
2125
+ },
2126
+ {
2127
+ "epoch": 0.9,
2128
+ "grad_norm": 2.6436825298723967,
2129
+ "learning_rate": 5.4409672830725456e-08,
2130
+ "logits/chosen": -2.142705202102661,
2131
+ "logits/rejected": -2.2424240112304688,
2132
+ "logps/chosen": -190.54376220703125,
2133
+ "logps/rejected": -385.0908203125,
2134
+ "loss": 0.0004,
2135
+ "rewards/accuracies": 1.0,
2136
+ "rewards/chosen": -5.370065689086914,
2137
+ "rewards/margins": 24.37637710571289,
2138
+ "rewards/rejected": -29.746444702148438,
2139
+ "step": 1410
2140
+ },
2141
+ {
2142
+ "epoch": 0.91,
2143
+ "grad_norm": 0.5795739537460435,
2144
+ "learning_rate": 5.0853485064011376e-08,
2145
+ "logits/chosen": -2.0973267555236816,
2146
+ "logits/rejected": -2.2426836490631104,
2147
+ "logps/chosen": -177.80783081054688,
2148
+ "logps/rejected": -375.2071228027344,
2149
+ "loss": 0.0005,
2150
+ "rewards/accuracies": 1.0,
2151
+ "rewards/chosen": -5.52896785736084,
2152
+ "rewards/margins": 23.480552673339844,
2153
+ "rewards/rejected": -29.009521484375,
2154
+ "step": 1420
2155
+ },
2156
+ {
2157
+ "epoch": 0.91,
2158
+ "grad_norm": 2.1741346744880965,
2159
+ "learning_rate": 4.72972972972973e-08,
2160
+ "logits/chosen": -2.207512617111206,
2161
+ "logits/rejected": -2.280388832092285,
2162
+ "logps/chosen": -201.04510498046875,
2163
+ "logps/rejected": -363.89398193359375,
2164
+ "loss": 0.0003,
2165
+ "rewards/accuracies": 1.0,
2166
+ "rewards/chosen": -5.225054740905762,
2167
+ "rewards/margins": 22.87972068786621,
2168
+ "rewards/rejected": -28.10477638244629,
2169
+ "step": 1430
2170
+ },
2171
+ {
2172
+ "epoch": 0.92,
2173
+ "grad_norm": 0.20540781104484357,
2174
+ "learning_rate": 4.374110953058322e-08,
2175
+ "logits/chosen": -2.1444334983825684,
2176
+ "logits/rejected": -2.261331081390381,
2177
+ "logps/chosen": -170.37319946289062,
2178
+ "logps/rejected": -368.539794921875,
2179
+ "loss": 0.0002,
2180
+ "rewards/accuracies": 1.0,
2181
+ "rewards/chosen": -4.832165718078613,
2182
+ "rewards/margins": 23.566085815429688,
2183
+ "rewards/rejected": -28.398250579833984,
2184
+ "step": 1440
2185
+ },
2186
+ {
2187
+ "epoch": 0.93,
2188
+ "grad_norm": 0.05727460316662975,
2189
+ "learning_rate": 4.018492176386913e-08,
2190
+ "logits/chosen": -2.1009721755981445,
2191
+ "logits/rejected": -2.2630536556243896,
2192
+ "logps/chosen": -166.5227508544922,
2193
+ "logps/rejected": -385.5596923828125,
2194
+ "loss": 0.0001,
2195
+ "rewards/accuracies": 1.0,
2196
+ "rewards/chosen": -4.496529579162598,
2197
+ "rewards/margins": 25.6341552734375,
2198
+ "rewards/rejected": -30.130685806274414,
2199
+ "step": 1450
2200
+ },
2201
+ {
2202
+ "epoch": 0.93,
2203
+ "grad_norm": 0.013882423497417175,
2204
+ "learning_rate": 3.6628733997155046e-08,
2205
+ "logits/chosen": -2.139456272125244,
2206
+ "logits/rejected": -2.292842149734497,
2207
+ "logps/chosen": -169.7640838623047,
2208
+ "logps/rejected": -355.3850402832031,
2209
+ "loss": 0.001,
2210
+ "rewards/accuracies": 1.0,
2211
+ "rewards/chosen": -4.837841987609863,
2212
+ "rewards/margins": 22.442852020263672,
2213
+ "rewards/rejected": -27.28069496154785,
2214
+ "step": 1460
2215
+ },
2216
+ {
2217
+ "epoch": 0.94,
2218
+ "grad_norm": 0.4937457805533542,
2219
+ "learning_rate": 3.3072546230440967e-08,
2220
+ "logits/chosen": -2.09645414352417,
2221
+ "logits/rejected": -2.3198351860046387,
2222
+ "logps/chosen": -166.16136169433594,
2223
+ "logps/rejected": -372.30316162109375,
2224
+ "loss": 0.0002,
2225
+ "rewards/accuracies": 1.0,
2226
+ "rewards/chosen": -4.320296764373779,
2227
+ "rewards/margins": 23.720216751098633,
2228
+ "rewards/rejected": -28.040512084960938,
2229
+ "step": 1470
2230
+ },
2231
+ {
2232
+ "epoch": 0.95,
2233
+ "grad_norm": 0.00827921495481582,
2234
+ "learning_rate": 2.9516358463726884e-08,
2235
+ "logits/chosen": -2.13936185836792,
2236
+ "logits/rejected": -2.291224241256714,
2237
+ "logps/chosen": -171.7252960205078,
2238
+ "logps/rejected": -388.84356689453125,
2239
+ "loss": 0.0002,
2240
+ "rewards/accuracies": 1.0,
2241
+ "rewards/chosen": -4.682975769042969,
2242
+ "rewards/margins": 25.828012466430664,
2243
+ "rewards/rejected": -30.5109920501709,
2244
+ "step": 1480
2245
+ },
2246
+ {
2247
+ "epoch": 0.95,
2248
+ "grad_norm": 0.004370788618868951,
2249
+ "learning_rate": 2.59601706970128e-08,
2250
+ "logits/chosen": -2.1469688415527344,
2251
+ "logits/rejected": -2.289008617401123,
2252
+ "logps/chosen": -170.2058868408203,
2253
+ "logps/rejected": -381.54656982421875,
2254
+ "loss": 0.0001,
2255
+ "rewards/accuracies": 1.0,
2256
+ "rewards/chosen": -4.3197221755981445,
2257
+ "rewards/margins": 25.29262924194336,
2258
+ "rewards/rejected": -29.612350463867188,
2259
+ "step": 1490
2260
+ },
2261
+ {
2262
+ "epoch": 0.96,
2263
+ "grad_norm": 0.19726476194562312,
2264
+ "learning_rate": 2.240398293029872e-08,
2265
+ "logits/chosen": -2.1768240928649902,
2266
+ "logits/rejected": -2.3078079223632812,
2267
+ "logps/chosen": -173.7339630126953,
2268
+ "logps/rejected": -385.33099365234375,
2269
+ "loss": 0.0001,
2270
+ "rewards/accuracies": 1.0,
2271
+ "rewards/chosen": -3.993574619293213,
2272
+ "rewards/margins": 25.351411819458008,
2273
+ "rewards/rejected": -29.344989776611328,
2274
+ "step": 1500
2275
+ },
2276
+ {
2277
+ "epoch": 0.97,
2278
+ "grad_norm": 0.00858791476540018,
2279
+ "learning_rate": 1.8847795163584636e-08,
2280
+ "logits/chosen": -2.18503737449646,
2281
+ "logits/rejected": -2.308922529220581,
2282
+ "logps/chosen": -176.88064575195312,
2283
+ "logps/rejected": -373.804443359375,
2284
+ "loss": 0.0003,
2285
+ "rewards/accuracies": 1.0,
2286
+ "rewards/chosen": -4.562438011169434,
2287
+ "rewards/margins": 24.36642837524414,
2288
+ "rewards/rejected": -28.928863525390625,
2289
+ "step": 1510
2290
+ },
2291
+ {
2292
+ "epoch": 0.97,
2293
+ "grad_norm": 3.138354599232436,
2294
+ "learning_rate": 1.5291607396870554e-08,
2295
+ "logits/chosen": -2.116145610809326,
2296
+ "logits/rejected": -2.2737691402435303,
2297
+ "logps/chosen": -175.29978942871094,
2298
+ "logps/rejected": -384.5267639160156,
2299
+ "loss": 0.0004,
2300
+ "rewards/accuracies": 1.0,
2301
+ "rewards/chosen": -5.154788017272949,
2302
+ "rewards/margins": 24.504108428955078,
2303
+ "rewards/rejected": -29.658893585205078,
2304
+ "step": 1520
2305
+ },
2306
+ {
2307
+ "epoch": 0.98,
2308
+ "grad_norm": 0.03067506561400629,
2309
+ "learning_rate": 1.1735419630156473e-08,
2310
+ "logits/chosen": -2.0927538871765137,
2311
+ "logits/rejected": -2.2137911319732666,
2312
+ "logps/chosen": -163.97552490234375,
2313
+ "logps/rejected": -366.1586608886719,
2314
+ "loss": 0.002,
2315
+ "rewards/accuracies": 1.0,
2316
+ "rewards/chosen": -4.887188911437988,
2317
+ "rewards/margins": 23.64038848876953,
2318
+ "rewards/rejected": -28.527576446533203,
2319
+ "step": 1530
2320
+ },
2321
+ {
2322
+ "epoch": 0.99,
2323
+ "grad_norm": 0.44417234512279563,
2324
+ "learning_rate": 8.179231863442388e-09,
2325
+ "logits/chosen": -2.15065860748291,
2326
+ "logits/rejected": -2.25085711479187,
2327
+ "logps/chosen": -174.73416137695312,
2328
+ "logps/rejected": -349.6435241699219,
2329
+ "loss": 0.0005,
2330
+ "rewards/accuracies": 1.0,
2331
+ "rewards/chosen": -4.77908992767334,
2332
+ "rewards/margins": 21.8181209564209,
2333
+ "rewards/rejected": -26.597209930419922,
2334
+ "step": 1540
2335
+ },
2336
+ {
2337
+ "epoch": 0.99,
2338
+ "grad_norm": 0.02198506573420274,
2339
+ "learning_rate": 4.623044096728307e-09,
2340
+ "logits/chosen": -2.1306891441345215,
2341
+ "logits/rejected": -2.3441672325134277,
2342
+ "logps/chosen": -183.6136016845703,
2343
+ "logps/rejected": -383.9253234863281,
2344
+ "loss": 0.0002,
2345
+ "rewards/accuracies": 1.0,
2346
+ "rewards/chosen": -4.792498588562012,
2347
+ "rewards/margins": 24.779081344604492,
2348
+ "rewards/rejected": -29.571582794189453,
2349
+ "step": 1550
2350
+ },
2351
+ {
2352
+ "epoch": 1.0,
2353
+ "grad_norm": 0.01071067202668103,
2354
+ "learning_rate": 1.0668563300142248e-09,
2355
+ "logits/chosen": -2.1825332641601562,
2356
+ "logits/rejected": -2.2997748851776123,
2357
+ "logps/chosen": -168.47274780273438,
2358
+ "logps/rejected": -385.79119873046875,
2359
+ "loss": 0.0001,
2360
+ "rewards/accuracies": 1.0,
2361
+ "rewards/chosen": -4.786942481994629,
2362
+ "rewards/margins": 25.29857635498047,
2363
+ "rewards/rejected": -30.085519790649414,
2364
+ "step": 1560
2365
+ },
2366
+ {
2367
+ "epoch": 1.0,
2368
+ "step": 1563,
2369
+ "total_flos": 0.0,
2370
+ "train_loss": 0.013828172848911992,
2371
+ "train_runtime": 8967.2368,
2372
+ "train_samples_per_second": 5.576,
2373
+ "train_steps_per_second": 0.174
2374
+ }
2375
+ ],
2376
+ "logging_steps": 10,
2377
+ "max_steps": 1563,
2378
+ "num_input_tokens_seen": 0,
2379
+ "num_train_epochs": 1,
2380
+ "save_steps": 100,
2381
+ "total_flos": 0.0,
2382
+ "train_batch_size": 8,
2383
+ "trial_name": null,
2384
+ "trial_params": null
2385
+ }