martimfasantos commited on
Commit
2089607
·
verified ·
1 Parent(s): ea4ba97

End of training

Browse files
README.md CHANGED
@@ -2,11 +2,13 @@
2
  license: apache-2.0
3
  library_name: peft
4
  tags:
 
5
  - trl
6
  - dpo
7
- - alignment-handbook
8
  - generated_from_trainer
9
  base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
 
 
10
  model-index:
11
  - name: tinyllama-1.1b-sum-dpo-qlora
12
  results: []
@@ -17,17 +19,17 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  # tinyllama-1.1b-sum-dpo-qlora
19
 
20
- This model is a fine-tuned version of [TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T](https://huggingface.co/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
- - Logits/chosen: -3.0239
23
- - Logits/rejected: -3.0176
24
- - Logps/chosen: -166.7881
25
- - Logps/rejected: -187.0472
26
  - Loss: 0.6482
27
- - Rewards/accuracies: 0.6171
28
  - Rewards/chosen: -0.9538
29
- - Rewards/margins: 0.1656
30
  - Rewards/rejected: -1.1194
 
 
 
 
 
 
31
 
32
  ## Model description
33
 
 
2
  license: apache-2.0
3
  library_name: peft
4
  tags:
5
+ - alignment-handbook
6
  - trl
7
  - dpo
 
8
  - generated_from_trainer
9
  base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
10
+ datasets:
11
+ - openai/summarize_from_feedback
12
  model-index:
13
  - name: tinyllama-1.1b-sum-dpo-qlora
14
  results: []
 
19
 
20
  # tinyllama-1.1b-sum-dpo-qlora
21
 
22
+ This model is a fine-tuned version of [martimfasantos/tinyllama-1.1b-sum-sft-qlora](https://huggingface.co/martimfasantos/tinyllama-1.1b-sum-sft-qlora) on the openai/summarize_from_feedback dataset.
23
  It achieves the following results on the evaluation set:
 
 
 
 
24
  - Loss: 0.6482
 
25
  - Rewards/chosen: -0.9538
 
26
  - Rewards/rejected: -1.1194
27
+ - Rewards/accuracies: 0.6171
28
+ - Rewards/margins: 0.1656
29
+ - Logps/rejected: -187.0472
30
+ - Logps/chosen: -166.7881
31
+ - Logits/rejected: -3.0176
32
+ - Logits/chosen: -3.0239
33
 
34
  ## Model description
35
 
all_results.json CHANGED
@@ -1,5 +1,18 @@
1
  {
2
  "epoch": 1.0,
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  "train_loss": 0.0003277428618961422,
4
  "train_runtime": 17.7068,
5
  "train_samples": 92858,
 
1
  {
2
  "epoch": 1.0,
3
+ "eval_logits/chosen": -3.023923397064209,
4
+ "eval_logits/rejected": -3.017603874206543,
5
+ "eval_logps/chosen": -166.78807067871094,
6
+ "eval_logps/rejected": -187.0471954345703,
7
+ "eval_loss": 0.6481729745864868,
8
+ "eval_rewards/accuracies": 0.6171003580093384,
9
+ "eval_rewards/chosen": -0.9537805914878845,
10
+ "eval_rewards/margins": 0.1656205952167511,
11
+ "eval_rewards/rejected": -1.1194013357162476,
12
+ "eval_runtime": 483.9291,
13
+ "eval_samples": 4304,
14
+ "eval_samples_per_second": 8.894,
15
+ "eval_steps_per_second": 1.112,
16
  "train_loss": 0.0003277428618961422,
17
  "train_runtime": 17.7068,
18
  "train_samples": 92858,
eval_results.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_logits/chosen": -3.023923397064209,
4
+ "eval_logits/rejected": -3.017603874206543,
5
+ "eval_logps/chosen": -166.78807067871094,
6
+ "eval_logps/rejected": -187.0471954345703,
7
+ "eval_loss": 0.6481729745864868,
8
+ "eval_rewards/accuracies": 0.6171003580093384,
9
+ "eval_rewards/chosen": -0.9537805914878845,
10
+ "eval_rewards/margins": 0.1656205952167511,
11
+ "eval_rewards/rejected": -1.1194013357162476,
12
+ "eval_runtime": 483.9291,
13
+ "eval_samples": 4304,
14
+ "eval_samples_per_second": 8.894,
15
+ "eval_steps_per_second": 1.112
16
+ }
runs/May10_17-24-35_poseidon/events.out.tfevents.1715362408.poseidon.2825808.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17351d92b525e2982b03f12e406dede84638b2c14162f66f4796dfa5070a83b8
3
+ size 828