CharlesLi commited on
Commit
82d7cd6
1 Parent(s): 60b4bb5

Model save

Browse files
README.md CHANGED
@@ -3,6 +3,7 @@ library_name: transformers
3
  tags:
4
  - trl
5
  - dpo
 
6
  - generated_from_trainer
7
  model-index:
8
  - name: OpenELM-1_1B-SLiC
@@ -16,15 +17,15 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model was trained from scratch on an unknown dataset.
18
  It achieves the following results on the evaluation set:
 
 
 
 
19
  - Loss: 0.6883
20
- - Rewards/chosen: -4.3438
21
- - Rewards/rejected: -5.3438
22
  - Rewards/accuracies: 0.7344
 
23
  - Rewards/margins: 0.9922
24
- - Logps/rejected: -824.0
25
- - Logps/chosen: -752.0
26
- - Logits/rejected: -8.75
27
- - Logits/chosen: -10.0625
28
 
29
  ## Model description
30
 
@@ -59,36 +60,36 @@ The following hyperparameters were used during training:
59
 
60
  ### Training results
61
 
62
- | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
63
- |:-------------:|:------:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
64
- | 0.7634 | 0.1047 | 100 | 0.7878 | -0.7461 | -1.0312 | 0.6406 | 0.2832 | -392.0 | -392.0 | -12.9375 | -13.0625 |
65
- | 0.7498 | 0.2093 | 200 | 0.7468 | -1.1719 | -1.5547 | 0.6719 | 0.3809 | -444.0 | -436.0 | -12.4375 | -12.75 |
66
- | 0.8142 | 0.3140 | 300 | 0.7466 | -1.8594 | -2.2812 | 0.6914 | 0.4141 | -516.0 | -504.0 | -14.75 | -14.8125 |
67
- | 0.7764 | 0.4186 | 400 | 0.7499 | -1.9688 | -2.4062 | 0.6699 | 0.4316 | -528.0 | -516.0 | -14.4375 | -14.5625 |
68
- | 0.731 | 0.5233 | 500 | 0.7240 | -2.4219 | -2.8594 | 0.6914 | 0.4375 | -576.0 | -560.0 | -10.5 | -11.0 |
69
- | 0.665 | 0.6279 | 600 | 0.7045 | -3.4062 | -4.0625 | 0.6973 | 0.6680 | -696.0 | -660.0 | -10.0625 | -10.75 |
70
- | 0.6806 | 0.7326 | 700 | 0.6912 | -2.5156 | -3.1562 | 0.7070 | 0.6523 | -604.0 | -568.0 | -13.4375 | -13.875 |
71
- | 0.6597 | 0.8373 | 800 | 0.7087 | -2.2969 | -2.8594 | 0.6777 | 0.5664 | -576.0 | -548.0 | -13.3125 | -13.5 |
72
- | 0.7325 | 0.9419 | 900 | 0.6838 | -2.6875 | -3.3594 | 0.7090 | 0.6602 | -624.0 | -588.0 | -13.25 | -14.0 |
73
- | 0.2677 | 1.0466 | 1000 | 0.6726 | -3.2344 | -4.0 | 0.7070 | 0.7734 | -688.0 | -640.0 | -11.0625 | -12.1875 |
74
- | 0.2256 | 1.1512 | 1100 | 0.6992 | -3.5938 | -4.375 | 0.7090 | 0.7969 | -728.0 | -676.0 | -10.0625 | -11.125 |
75
- | 0.1954 | 1.2559 | 1200 | 0.7033 | -3.4688 | -4.3125 | 0.7051 | 0.8477 | -720.0 | -664.0 | -10.125 | -11.3125 |
76
- | 0.2289 | 1.3605 | 1300 | 0.6722 | -3.7344 | -4.5 | 0.7344 | 0.7852 | -740.0 | -692.0 | -9.9375 | -11.0 |
77
- | 0.2227 | 1.4652 | 1400 | 0.6925 | -3.5781 | -4.3125 | 0.6953 | 0.7383 | -720.0 | -676.0 | -11.8125 | -12.5 |
78
- | 0.1902 | 1.5699 | 1500 | 0.6758 | -4.1875 | -5.0312 | 0.7148 | 0.8320 | -792.0 | -736.0 | -11.125 | -12.0625 |
79
- | 0.2192 | 1.6745 | 1600 | 0.6833 | -3.8438 | -4.625 | 0.7148 | 0.7695 | -748.0 | -704.0 | -12.875 | -13.625 |
80
- | 0.2137 | 1.7792 | 1700 | 0.6734 | -3.9688 | -4.7812 | 0.7207 | 0.8008 | -764.0 | -716.0 | -11.0 | -11.9375 |
81
- | 0.2001 | 1.8838 | 1800 | 0.6734 | -3.7344 | -4.5 | 0.7207 | 0.7617 | -740.0 | -692.0 | -11.3125 | -12.125 |
82
- | 0.1713 | 1.9885 | 1900 | 0.6680 | -3.9375 | -4.8125 | 0.7383 | 0.8789 | -768.0 | -712.0 | -9.25 | -10.4375 |
83
- | 0.0184 | 2.0931 | 2000 | 0.6845 | -3.8594 | -4.8125 | 0.7305 | 0.9453 | -768.0 | -704.0 | -9.875 | -11.0625 |
84
- | 0.0313 | 2.1978 | 2100 | 0.6798 | -4.0 | -4.9688 | 0.7402 | 0.9570 | -784.0 | -720.0 | -10.125 | -11.25 |
85
- | 0.0401 | 2.3025 | 2200 | 0.6865 | -4.1562 | -5.0938 | 0.7363 | 0.9492 | -800.0 | -732.0 | -9.375 | -10.6875 |
86
- | 0.0211 | 2.4071 | 2300 | 0.6874 | -4.2188 | -5.2188 | 0.7383 | 1.0078 | -812.0 | -740.0 | -8.75 | -10.125 |
87
- | 0.0239 | 2.5118 | 2400 | 0.6858 | -4.1562 | -5.125 | 0.7383 | 0.9766 | -800.0 | -736.0 | -8.875 | -10.1875 |
88
- | 0.0188 | 2.6164 | 2500 | 0.6902 | -4.2812 | -5.25 | 0.7324 | 0.9883 | -816.0 | -744.0 | -8.8125 | -10.125 |
89
- | 0.0145 | 2.7211 | 2600 | 0.6874 | -4.2812 | -5.2812 | 0.7383 | 0.9844 | -816.0 | -748.0 | -8.8125 | -10.125 |
90
- | 0.0229 | 2.8257 | 2700 | 0.6883 | -4.3438 | -5.3438 | 0.7344 | 0.9922 | -824.0 | -752.0 | -8.75 | -10.0625 |
91
- | 0.0298 | 2.9304 | 2800 | 0.6883 | -4.3438 | -5.3438 | 0.7344 | 0.9922 | -824.0 | -752.0 | -8.75 | -10.0625 |
92
 
93
 
94
  ### Framework versions
 
3
  tags:
4
  - trl
5
  - dpo
6
+ - alignment-handbook
7
  - generated_from_trainer
8
  model-index:
9
  - name: OpenELM-1_1B-SLiC
 
17
 
18
  This model was trained from scratch on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Logits/chosen: -10.0625
21
+ - Logits/rejected: -8.75
22
+ - Logps/chosen: -752.0
23
+ - Logps/rejected: -824.0
24
  - Loss: 0.6883
 
 
25
  - Rewards/accuracies: 0.7344
26
+ - Rewards/chosen: -4.3438
27
  - Rewards/margins: 0.9922
28
+ - Rewards/rejected: -5.3438
 
 
 
29
 
30
  ## Model description
31
 
 
60
 
61
  ### Training results
62
 
63
+ | Training Loss | Epoch | Step | Logits/chosen | Logits/rejected | Logps/chosen | Logps/rejected | Validation Loss | Rewards/accuracies | Rewards/chosen | Rewards/margins | Rewards/rejected |
64
+ |:-------------:|:------:|:----:|:-------------:|:---------------:|:------------:|:--------------:|:---------------:|:------------------:|:--------------:|:---------------:|:----------------:|
65
+ | 0.7634 | 0.1047 | 100 | -13.0625 | -12.9375 | -392.0 | -392.0 | 0.7878 | 0.6406 | -0.7461 | 0.2832 | -1.0312 |
66
+ | 0.7498 | 0.2093 | 200 | -12.75 | -12.4375 | -436.0 | -444.0 | 0.7468 | 0.6719 | -1.1719 | 0.3809 | -1.5547 |
67
+ | 0.8142 | 0.3140 | 300 | -14.8125 | -14.75 | -504.0 | -516.0 | 0.7466 | 0.6914 | -1.8594 | 0.4141 | -2.2812 |
68
+ | 0.7764 | 0.4186 | 400 | -14.5625 | -14.4375 | -516.0 | -528.0 | 0.7499 | 0.6699 | -1.9688 | 0.4316 | -2.4062 |
69
+ | 0.731 | 0.5233 | 500 | -11.0 | -10.5 | -560.0 | -576.0 | 0.7240 | 0.6914 | -2.4219 | 0.4375 | -2.8594 |
70
+ | 0.665 | 0.6279 | 600 | -10.75 | -10.0625 | -660.0 | -696.0 | 0.7045 | 0.6973 | -3.4062 | 0.6680 | -4.0625 |
71
+ | 0.6806 | 0.7326 | 700 | -13.875 | -13.4375 | -568.0 | -604.0 | 0.6912 | 0.7070 | -2.5156 | 0.6523 | -3.1562 |
72
+ | 0.6597 | 0.8373 | 800 | -13.5 | -13.3125 | -548.0 | -576.0 | 0.7087 | 0.6777 | -2.2969 | 0.5664 | -2.8594 |
73
+ | 0.7325 | 0.9419 | 900 | -14.0 | -13.25 | -588.0 | -624.0 | 0.6838 | 0.7090 | -2.6875 | 0.6602 | -3.3594 |
74
+ | 0.2677 | 1.0466 | 1000 | -12.1875 | -11.0625 | -640.0 | -688.0 | 0.6726 | 0.7070 | -3.2344 | 0.7734 | -4.0 |
75
+ | 0.2256 | 1.1512 | 1100 | -11.125 | -10.0625 | -676.0 | -728.0 | 0.6992 | 0.7090 | -3.5938 | 0.7969 | -4.375 |
76
+ | 0.1954 | 1.2559 | 1200 | -11.3125 | -10.125 | -664.0 | -720.0 | 0.7033 | 0.7051 | -3.4688 | 0.8477 | -4.3125 |
77
+ | 0.2289 | 1.3605 | 1300 | -11.0 | -9.9375 | -692.0 | -740.0 | 0.6722 | 0.7344 | -3.7344 | 0.7852 | -4.5 |
78
+ | 0.2227 | 1.4652 | 1400 | -12.5 | -11.8125 | -676.0 | -720.0 | 0.6925 | 0.6953 | -3.5781 | 0.7383 | -4.3125 |
79
+ | 0.1902 | 1.5699 | 1500 | -12.0625 | -11.125 | -736.0 | -792.0 | 0.6758 | 0.7148 | -4.1875 | 0.8320 | -5.0312 |
80
+ | 0.2192 | 1.6745 | 1600 | -13.625 | -12.875 | -704.0 | -748.0 | 0.6833 | 0.7148 | -3.8438 | 0.7695 | -4.625 |
81
+ | 0.2137 | 1.7792 | 1700 | -11.9375 | -11.0 | -716.0 | -764.0 | 0.6734 | 0.7207 | -3.9688 | 0.8008 | -4.7812 |
82
+ | 0.2001 | 1.8838 | 1800 | -12.125 | -11.3125 | -692.0 | -740.0 | 0.6734 | 0.7207 | -3.7344 | 0.7617 | -4.5 |
83
+ | 0.1713 | 1.9885 | 1900 | -10.4375 | -9.25 | -712.0 | -768.0 | 0.6680 | 0.7383 | -3.9375 | 0.8789 | -4.8125 |
84
+ | 0.0184 | 2.0931 | 2000 | -11.0625 | -9.875 | -704.0 | -768.0 | 0.6845 | 0.7305 | -3.8594 | 0.9453 | -4.8125 |
85
+ | 0.0313 | 2.1978 | 2100 | -11.25 | -10.125 | -720.0 | -784.0 | 0.6798 | 0.7402 | -4.0 | 0.9570 | -4.9688 |
86
+ | 0.0401 | 2.3025 | 2200 | -10.6875 | -9.375 | -732.0 | -800.0 | 0.6865 | 0.7363 | -4.1562 | 0.9492 | -5.0938 |
87
+ | 0.0211 | 2.4071 | 2300 | -10.125 | -8.75 | -740.0 | -812.0 | 0.6874 | 0.7383 | -4.2188 | 1.0078 | -5.2188 |
88
+ | 0.0239 | 2.5118 | 2400 | -10.1875 | -8.875 | -736.0 | -800.0 | 0.6858 | 0.7383 | -4.1562 | 0.9766 | -5.125 |
89
+ | 0.0188 | 2.6164 | 2500 | -10.125 | -8.8125 | -744.0 | -816.0 | 0.6902 | 0.7324 | -4.2812 | 0.9883 | -5.25 |
90
+ | 0.0145 | 2.7211 | 2600 | -10.125 | -8.8125 | -748.0 | -816.0 | 0.6874 | 0.7383 | -4.2812 | 0.9844 | -5.2812 |
91
+ | 0.0229 | 2.8257 | 2700 | -10.0625 | -8.75 | -752.0 | -824.0 | 0.6883 | 0.7344 | -4.3438 | 0.9922 | -5.3438 |
92
+ | 0.0298 | 2.9304 | 2800 | -10.0625 | -8.75 | -752.0 | -824.0 | 0.6883 | 0.7344 | -4.3438 | 0.9922 | -5.3438 |
93
 
94
 
95
  ### Framework versions
all_results.json CHANGED
@@ -1,9 +1,22 @@
1
  {
2
  "epoch": 2.998430141287284,
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  "total_flos": 0.0,
4
- "train_loss": 0.32237146187827226,
5
- "train_runtime": 12289.6357,
6
  "train_samples": 61134,
7
- "train_samples_per_second": 14.923,
8
- "train_steps_per_second": 0.233
9
  }
 
1
  {
2
  "epoch": 2.998430141287284,
3
+ "eval_logits/chosen": -10.0625,
4
+ "eval_logits/rejected": -8.75,
5
+ "eval_logps/chosen": -752.0,
6
+ "eval_logps/rejected": -824.0,
7
+ "eval_loss": 0.6881640553474426,
8
+ "eval_rewards/accuracies": 0.73828125,
9
+ "eval_rewards/chosen": -4.34375,
10
+ "eval_rewards/margins": 0.9921875,
11
+ "eval_rewards/rejected": -5.34375,
12
+ "eval_runtime": 46.6028,
13
+ "eval_samples": 2000,
14
+ "eval_samples_per_second": 42.916,
15
+ "eval_steps_per_second": 0.687,
16
  "total_flos": 0.0,
17
+ "train_loss": 0.0,
18
+ "train_runtime": 0.0555,
19
  "train_samples": 61134,
20
+ "train_samples_per_second": 3306285.721,
21
+ "train_steps_per_second": 51648.884
22
  }
eval_results.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.998430141287284,
3
+ "eval_logits/chosen": -10.0625,
4
+ "eval_logits/rejected": -8.75,
5
+ "eval_logps/chosen": -752.0,
6
+ "eval_logps/rejected": -824.0,
7
+ "eval_loss": 0.6881640553474426,
8
+ "eval_rewards/accuracies": 0.73828125,
9
+ "eval_rewards/chosen": -4.34375,
10
+ "eval_rewards/margins": 0.9921875,
11
+ "eval_rewards/rejected": -5.34375,
12
+ "eval_runtime": 46.6028,
13
+ "eval_samples": 2000,
14
+ "eval_samples_per_second": 42.916,
15
+ "eval_steps_per_second": 0.687
16
+ }
runs/Sep20_12-36-35_xe8545-a100-29/events.out.tfevents.1726841470.xe8545-a100-29.130585.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbb2566e4ffc9e692b5751b33f944f88f09052197134d466860c9d7a5bfd59bf
3
+ size 828
runs/Sep20_23-38-34_xe8545-a100-31/events.out.tfevents.1726868855.xe8545-a100-31.112058.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bede5ef115179935360a3b44dd357286ee954e8ba37f6d3ee858075608fece52
3
+ size 7584
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 2.998430141287284,
3
  "total_flos": 0.0,
4
- "train_loss": 0.32237146187827226,
5
- "train_runtime": 12289.6357,
6
  "train_samples": 61134,
7
- "train_samples_per_second": 14.923,
8
- "train_steps_per_second": 0.233
9
  }
 
1
  {
2
  "epoch": 2.998430141287284,
3
  "total_flos": 0.0,
4
+ "train_loss": 0.0,
5
+ "train_runtime": 0.0555,
6
  "train_samples": 61134,
7
+ "train_samples_per_second": 3306285.721,
8
+ "train_steps_per_second": 51648.884
9
  }
trainer_state.json CHANGED
@@ -4765,10 +4765,10 @@
4765
  "epoch": 2.998430141287284,
4766
  "step": 2865,
4767
  "total_flos": 0.0,
4768
- "train_loss": 0.32237146187827226,
4769
- "train_runtime": 12289.6357,
4770
- "train_samples_per_second": 14.923,
4771
- "train_steps_per_second": 0.233
4772
  }
4773
  ],
4774
  "logging_steps": 10,
 
4765
  "epoch": 2.998430141287284,
4766
  "step": 2865,
4767
  "total_flos": 0.0,
4768
+ "train_loss": 0.0,
4769
+ "train_runtime": 0.0555,
4770
+ "train_samples_per_second": 3306285.721,
4771
+ "train_steps_per_second": 51648.884
4772
  }
4773
  ],
4774
  "logging_steps": 10,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:770139a9cd07aee62daec75582c947b74da28832436bcc45ad8457b862177472
3
  size 7544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fde72b2e3187ad24ebbb60938e3b6c6ba097afcac66749731a5c1b5a8f27722
3
  size 7544