jcmei commited on
Commit
efaa6a3
·
verified ·
1 Parent(s): 85aaf23

Model save

Browse files
README.md CHANGED
@@ -2,16 +2,10 @@
2
  license: mit
3
  base_model: HuggingFaceH4/mistral-7b-sft-beta
4
  tags:
5
- - alignment-handbook
6
- - trl
7
- - dpo
8
- - generated_from_trainer
9
  - trl
10
  - dpo
11
  - alignment-handbook
12
  - generated_from_trainer
13
- datasets:
14
- - HuggingFaceH4/ultrafeedback_binarized
15
  model-index:
16
  - name: DPO-Zephyr-7B
17
  results: []
@@ -22,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
22
 
23
  # DPO-Zephyr-7B
24
 
25
- This model is a fine-tuned version of [HuggingFaceH4/mistral-7b-sft-beta](https://huggingface.co/HuggingFaceH4/mistral-7b-sft-beta) on the HuggingFaceH4/ultrafeedback_binarized dataset.
26
 
27
  ## Model description
28
 
@@ -47,8 +41,8 @@ The following hyperparameters were used during training:
47
  - seed: 42
48
  - distributed_type: multi-GPU
49
  - num_devices: 8
50
- - gradient_accumulation_steps: 4
51
- - total_train_batch_size: 128
52
  - total_eval_batch_size: 64
53
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
54
  - lr_scheduler_type: cosine
 
2
  license: mit
3
  base_model: HuggingFaceH4/mistral-7b-sft-beta
4
  tags:
 
 
 
 
5
  - trl
6
  - dpo
7
  - alignment-handbook
8
  - generated_from_trainer
 
 
9
  model-index:
10
  - name: DPO-Zephyr-7B
11
  results: []
 
16
 
17
  # DPO-Zephyr-7B
18
 
19
+ This model is a fine-tuned version of [HuggingFaceH4/mistral-7b-sft-beta](https://huggingface.co/HuggingFaceH4/mistral-7b-sft-beta) on the None dataset.
20
 
21
  ## Model description
22
 
 
41
  - seed: 42
42
  - distributed_type: multi-GPU
43
  - num_devices: 8
44
+ - gradient_accumulation_steps: 8
45
+ - total_train_batch_size: 256
46
  - total_eval_batch_size: 64
47
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
48
  - lr_scheduler_type: cosine
all_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 0.99581589958159,
3
  "total_flos": 0.0,
4
- "train_loss": 0.09238386955581793,
5
- "train_runtime": 386.322,
6
  "train_samples": 15283,
7
- "train_samples_per_second": 39.56,
8
- "train_steps_per_second": 0.308
9
  }
 
1
  {
2
+ "epoch": 0.8368200836820083,
3
  "total_flos": 0.0,
4
+ "train_loss": 0.0,
5
+ "train_runtime": 172.1278,
6
  "train_samples": 15283,
7
+ "train_samples_per_second": 88.789,
8
+ "train_steps_per_second": 0.343
9
  }
config.json CHANGED
@@ -21,6 +21,6 @@
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
  "transformers_version": "4.40.2",
24
- "use_cache": true,
25
  "vocab_size": 32000
26
  }
 
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
  "transformers_version": "4.40.2",
24
+ "use_cache": false,
25
  "vocab_size": 32000
26
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:af488688f56cda56ff87ad1eec035fa9ae3a85bca93bdbe59211aa9ae3eae187
3
  size 4943162336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed828d1d3747d787beccc5e05e82a27cf053394733010c14dce37af5e1b14050
3
  size 4943162336
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3768b0b7c81899c59e9afb844018c1b1f295d77c8c47e5c118e3562a776ae5a7
3
  size 4999819336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e1be7f1462cc851f1af5cb4395cd8721ff70220bf4f51337337e3d64ef195b4
3
  size 4999819336
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:16f3d87f94195cc6683db7aeafd3e916dad7e931f54a05872d55a851ad2d0ed0
3
  size 4540516344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65324868754ad5cbed72f6dce5c6497a4e143e70c2c101b69cdd69df2e3d31cf
3
  size 4540516344
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 0.99581589958159,
3
  "total_flos": 0.0,
4
- "train_loss": 0.09238386955581793,
5
- "train_runtime": 386.322,
6
  "train_samples": 15283,
7
- "train_samples_per_second": 39.56,
8
- "train_steps_per_second": 0.308
9
  }
 
1
  {
2
+ "epoch": 0.8368200836820083,
3
  "total_flos": 0.0,
4
+ "train_loss": 0.0,
5
+ "train_runtime": 172.1278,
6
  "train_samples": 15283,
7
+ "train_samples_per_second": 88.789,
8
+ "train_steps_per_second": 0.343
9
  }
trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.99581589958159,
5
  "eval_steps": 500,
6
- "global_step": 119,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -174,32 +174,17 @@
174
  "step": 100
175
  },
176
  {
177
- "epoch": 0.9205020920502092,
178
- "grad_norm": 15.746470079783288,
179
- "learning_rate": 8.677580722139671e-09,
180
- "logits/chosen": -2.7727127075195312,
181
- "logits/rejected": -2.7130322456359863,
182
- "logps/chosen": -367.272216796875,
183
- "logps/rejected": -343.4369201660156,
184
- "loss": 0.5945,
185
- "rewards/accuracies": 0.699999988079071,
186
- "rewards/chosen": -0.46789541840553284,
187
- "rewards/margins": 0.3372054696083069,
188
- "rewards/rejected": -0.8051007986068726,
189
- "step": 110
190
- },
191
- {
192
- "epoch": 0.99581589958159,
193
- "step": 119,
194
  "total_flos": 0.0,
195
- "train_loss": 0.09238386955581793,
196
- "train_runtime": 386.322,
197
- "train_samples_per_second": 39.56,
198
- "train_steps_per_second": 0.308
199
  }
200
  ],
201
  "logging_steps": 10,
202
- "max_steps": 119,
203
  "num_input_tokens_seen": 0,
204
  "num_train_epochs": 1,
205
  "save_steps": 100,
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.8368200836820083,
5
  "eval_steps": 500,
6
+ "global_step": 100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
174
  "step": 100
175
  },
176
  {
177
+ "epoch": 0.8368200836820083,
178
+ "step": 100,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
179
  "total_flos": 0.0,
180
+ "train_loss": 0.0,
181
+ "train_runtime": 172.1278,
182
+ "train_samples_per_second": 88.789,
183
+ "train_steps_per_second": 0.343
184
  }
185
  ],
186
  "logging_steps": 10,
187
+ "max_steps": 59,
188
  "num_input_tokens_seen": 0,
189
  "num_train_epochs": 1,
190
  "save_steps": 100,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:128ab99ba4e1911f9502d9e9cba3a105bb6a758aa83704d9f3e92a77614e5773
3
  size 6264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26d85afdfb7f05f00d4e8812ab20799311b4401986345249d4734fdf1972cd0b
3
  size 6264