DongfuJiang commited on
Commit
57177f9
·
verified ·
1 Parent(s): 8593ae3

End of training

Browse files
README.md CHANGED
@@ -4,6 +4,7 @@ library_name: peft
4
  license: mit
5
  tags:
6
  - llama-factory
 
7
  - full
8
  - generated_from_trainer
9
  model-index:
@@ -17,9 +18,9 @@ should probably proofread and complete it, then remove this comment. -->
17
  [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/dongfu/huggingface/runs/vdittf2b)
18
  # PairRM-V2-phi3-3-mini-ultra-feedback-binarized-lora
19
 
20
- This model is a fine-tuned version of [microsoft/Phi-3-mini-128k-instruct](https://huggingface.co/microsoft/Phi-3-mini-128k-instruct) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 0.2640
23
 
24
  ## Model description
25
 
 
4
  license: mit
5
  tags:
6
  - llama-factory
7
+ - lora
8
  - full
9
  - generated_from_trainer
10
  model-index:
 
18
  [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/dongfu/huggingface/runs/vdittf2b)
19
  # PairRM-V2-phi3-3-mini-ultra-feedback-binarized-lora
20
 
21
+ This model is a fine-tuned version of [microsoft/Phi-3-mini-128k-instruct](https://huggingface.co/microsoft/Phi-3-mini-128k-instruct) on the ultra-feedback-binarized dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.2605
24
 
25
  ## Model description
26
 
all_results.json CHANGED
@@ -1,12 +1,12 @@
1
  {
2
- "epoch": 1.9994796635157401,
3
- "eval_loss": 0.4639037847518921,
4
- "eval_runtime": 767.9829,
5
- "eval_samples_per_second": 12.644,
6
- "eval_steps_per_second": 1.581,
7
- "total_flos": 420396754255872.0,
8
- "train_loss": 0.3626429720482505,
9
- "train_runtime": 106139.7498,
10
- "train_samples_per_second": 3.476,
11
- "train_steps_per_second": 0.027
12
  }
 
1
  {
2
+ "epoch": 0.9992793658419409,
3
+ "eval_loss": 0.2604576349258423,
4
+ "eval_runtime": 270.7958,
5
+ "eval_samples_per_second": 19.421,
6
+ "eval_steps_per_second": 2.43,
7
+ "total_flos": 2.0573294793064448e+18,
8
+ "train_loss": 0.29625688539101525,
9
+ "train_runtime": 10334.3449,
10
+ "train_samples_per_second": 9.668,
11
+ "train_steps_per_second": 0.075
12
  }
eval_results.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "epoch": 1.9994796635157401,
3
- "eval_loss": 0.4639037847518921,
4
- "eval_runtime": 767.9829,
5
- "eval_samples_per_second": 12.644,
6
- "eval_steps_per_second": 1.581
7
  }
 
1
  {
2
+ "epoch": 0.9992793658419409,
3
+ "eval_loss": 0.2604576349258423,
4
+ "eval_runtime": 270.7958,
5
+ "eval_samples_per_second": 19.421,
6
+ "eval_steps_per_second": 2.43
7
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 1.9994796635157401,
3
- "total_flos": 420396754255872.0,
4
- "train_loss": 0.3626429720482505,
5
- "train_runtime": 106139.7498,
6
- "train_samples_per_second": 3.476,
7
- "train_steps_per_second": 0.027
8
  }
 
1
  {
2
+ "epoch": 0.9992793658419409,
3
+ "total_flos": 2.0573294793064448e+18,
4
+ "train_loss": 0.29625688539101525,
5
+ "train_runtime": 10334.3449,
6
+ "train_samples_per_second": 9.668,
7
+ "train_steps_per_second": 0.075
8
  }
trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff
 
training_eval_loss.png CHANGED
training_loss.png CHANGED