fpadovani commited on
Commit
2568548
·
verified ·
1 Parent(s): a3ee290

Training in progress, step 20000, checkpoint

Browse files
checkpoint-20000/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:82fd1c1df42995174911c316d58566d182a000491bb2c6dc4d5513854de66d3f
3
  size 51007160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71b6ef8cca4c6614bb6734b2dd3640350fc1eb3ae53a5ef74c1dc70c325f70a1
3
  size 51007160
checkpoint-20000/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:51f3ba6493e6418c34d757d055d741760e93f33dc179c22240e7fc3a94761de6
3
  size 102078202
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9dc2a91d67c7130566c43df1f0ac053c9461513fcc7573d918f05e074a02bee
3
  size 102078202
checkpoint-20000/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d0f5c66975498941bc1c143cde1c364129ddcc429019efeffe49329fb69d81c9
3
  size 14308
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe3593036612a9e72e49d67a4e0e919b703bdeeab50bebe3fb40f9a8b48b6b5f
3
  size 14308
checkpoint-20000/tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
checkpoint-20000/trainer_state.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "best_metric": 3.9675967693328857,
3
- "best_model_checkpoint": "/home/p318482/babyLM_controlled/models_trained/fr_clm/childes_30/checkpoint-20000",
4
- "epoch": 31.446540880503143,
5
  "eval_steps": 2000,
6
  "global_step": 20000,
7
  "is_hyper_param_search": false,
@@ -9,125 +9,125 @@
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 3.1446540880503147,
13
- "eval_loss": 6.67229700088501,
14
- "eval_runtime": 0.6568,
15
- "eval_samples_per_second": 1386.938,
16
- "eval_steps_per_second": 86.779,
17
  "step": 2000
18
  },
19
  {
20
- "epoch": 6.289308176100629,
21
- "grad_norm": 1.4630517959594727,
22
  "learning_rate": 1e-05,
23
- "loss": 6.5359,
24
  "step": 4000
25
  },
26
  {
27
- "epoch": 6.289308176100629,
28
- "eval_loss": 5.102311611175537,
29
- "eval_runtime": 0.6383,
30
- "eval_samples_per_second": 1427.13,
31
- "eval_steps_per_second": 89.294,
32
  "step": 4000
33
  },
34
  {
35
- "epoch": 9.433962264150944,
36
- "eval_loss": 4.675341606140137,
37
- "eval_runtime": 0.6396,
38
- "eval_samples_per_second": 1424.427,
39
- "eval_steps_per_second": 89.124,
40
  "step": 6000
41
  },
42
  {
43
- "epoch": 12.578616352201259,
44
- "grad_norm": 2.3778915405273438,
45
  "learning_rate": 2e-05,
46
- "loss": 4.34,
47
  "step": 8000
48
  },
49
  {
50
- "epoch": 12.578616352201259,
51
- "eval_loss": 4.42986536026001,
52
- "eval_runtime": 0.646,
53
- "eval_samples_per_second": 1410.229,
54
- "eval_steps_per_second": 88.236,
55
  "step": 8000
56
  },
57
  {
58
- "epoch": 15.723270440251572,
59
- "eval_loss": 4.2789154052734375,
60
- "eval_runtime": 0.6485,
61
- "eval_samples_per_second": 1404.856,
62
- "eval_steps_per_second": 87.9,
63
  "step": 10000
64
  },
65
  {
66
- "epoch": 18.867924528301888,
67
- "grad_norm": 2.512312889099121,
68
- "learning_rate": 2.9995e-05,
69
- "loss": 3.8692,
70
  "step": 12000
71
  },
72
  {
73
- "epoch": 18.867924528301888,
74
- "eval_loss": 4.1580729484558105,
75
- "eval_runtime": 0.6403,
76
- "eval_samples_per_second": 1422.769,
77
- "eval_steps_per_second": 89.021,
78
  "step": 12000
79
  },
80
  {
81
- "epoch": 22.0125786163522,
82
- "eval_loss": 4.082061767578125,
83
- "eval_runtime": 0.6415,
84
- "eval_samples_per_second": 1420.041,
85
- "eval_steps_per_second": 88.85,
86
  "step": 14000
87
  },
88
  {
89
- "epoch": 25.157232704402517,
90
- "grad_norm": 2.4205257892608643,
91
  "learning_rate": 3.999e-05,
92
- "loss": 3.5905,
93
  "step": 16000
94
  },
95
  {
96
- "epoch": 25.157232704402517,
97
- "eval_loss": 4.033827781677246,
98
- "eval_runtime": 0.6416,
99
- "eval_samples_per_second": 1419.823,
100
- "eval_steps_per_second": 88.836,
101
  "step": 16000
102
  },
103
  {
104
- "epoch": 28.30188679245283,
105
- "eval_loss": 3.9935402870178223,
106
- "eval_runtime": 0.6614,
107
- "eval_samples_per_second": 1377.343,
108
- "eval_steps_per_second": 86.178,
109
  "step": 18000
110
  },
111
  {
112
- "epoch": 31.446540880503143,
113
- "grad_norm": 2.5674731731414795,
114
  "learning_rate": 4.9985e-05,
115
- "loss": 3.3778,
116
  "step": 20000
117
  },
118
  {
119
- "epoch": 31.446540880503143,
120
- "eval_loss": 3.9675967693328857,
121
- "eval_runtime": 0.6406,
122
- "eval_samples_per_second": 1422.059,
123
- "eval_steps_per_second": 88.976,
124
  "step": 20000
125
  }
126
  ],
127
  "logging_steps": 4000,
128
  "max_steps": 100000,
129
  "num_input_tokens_seen": 0,
130
- "num_train_epochs": 158,
131
  "save_steps": 4000,
132
  "stateful_callbacks": {
133
  "TrainerControl": {
@@ -141,7 +141,7 @@
141
  "attributes": {}
142
  }
143
  },
144
- "total_flos": 5170824201633792.0,
145
  "train_batch_size": 16,
146
  "trial_name": null,
147
  "trial_params": null
 
1
  {
2
+ "best_metric": 4.690371513366699,
3
+ "best_model_checkpoint": "/home/p318482/babyLM_controlled/models_trained/de_clm/childes_30/checkpoint-20000",
4
+ "epoch": 20.964360587002098,
5
  "eval_steps": 2000,
6
  "global_step": 20000,
7
  "is_hyper_param_search": false,
 
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 2.0964360587002098,
13
+ "eval_loss": 7.102903366088867,
14
+ "eval_runtime": 0.9708,
15
+ "eval_samples_per_second": 1416.286,
16
+ "eval_steps_per_second": 88.582,
17
  "step": 2000
18
  },
19
  {
20
+ "epoch": 4.1928721174004195,
21
+ "grad_norm": 1.3964662551879883,
22
  "learning_rate": 1e-05,
23
+ "loss": 6.9987,
24
  "step": 4000
25
  },
26
  {
27
+ "epoch": 4.1928721174004195,
28
+ "eval_loss": 5.884151935577393,
29
+ "eval_runtime": 0.966,
30
+ "eval_samples_per_second": 1423.408,
31
+ "eval_steps_per_second": 89.028,
32
  "step": 4000
33
  },
34
  {
35
+ "epoch": 6.289308176100629,
36
+ "eval_loss": 5.54873514175415,
37
+ "eval_runtime": 0.9657,
38
+ "eval_samples_per_second": 1423.84,
39
+ "eval_steps_per_second": 89.055,
40
  "step": 6000
41
  },
42
  {
43
+ "epoch": 8.385744234800839,
44
+ "grad_norm": 2.7172107696533203,
45
  "learning_rate": 2e-05,
46
+ "loss": 5.2204,
47
  "step": 8000
48
  },
49
  {
50
+ "epoch": 8.385744234800839,
51
+ "eval_loss": 5.2793288230896,
52
+ "eval_runtime": 0.9644,
53
+ "eval_samples_per_second": 1425.779,
54
+ "eval_steps_per_second": 89.176,
55
  "step": 8000
56
  },
57
  {
58
+ "epoch": 10.482180293501049,
59
+ "eval_loss": 5.10486364364624,
60
+ "eval_runtime": 0.9641,
61
+ "eval_samples_per_second": 1426.204,
62
+ "eval_steps_per_second": 89.203,
63
  "step": 10000
64
  },
65
  {
66
+ "epoch": 12.578616352201259,
67
+ "grad_norm": 2.500443458557129,
68
+ "learning_rate": 2.99925e-05,
69
+ "loss": 4.7358,
70
  "step": 12000
71
  },
72
  {
73
+ "epoch": 12.578616352201259,
74
+ "eval_loss": 4.983631134033203,
75
+ "eval_runtime": 0.9644,
76
+ "eval_samples_per_second": 1425.809,
77
+ "eval_steps_per_second": 89.178,
78
  "step": 12000
79
  },
80
  {
81
+ "epoch": 14.675052410901468,
82
+ "eval_loss": 4.882917404174805,
83
+ "eval_runtime": 0.9686,
84
+ "eval_samples_per_second": 1419.612,
85
+ "eval_steps_per_second": 88.79,
86
  "step": 14000
87
  },
88
  {
89
+ "epoch": 16.771488469601678,
90
+ "grad_norm": 2.400749444961548,
91
  "learning_rate": 3.999e-05,
92
+ "loss": 4.4216,
93
  "step": 16000
94
  },
95
  {
96
+ "epoch": 16.771488469601678,
97
+ "eval_loss": 4.802889823913574,
98
+ "eval_runtime": 0.9763,
99
+ "eval_samples_per_second": 1408.393,
100
+ "eval_steps_per_second": 88.089,
101
  "step": 16000
102
  },
103
  {
104
+ "epoch": 18.867924528301888,
105
+ "eval_loss": 4.74226188659668,
106
+ "eval_runtime": 0.976,
107
+ "eval_samples_per_second": 1408.882,
108
+ "eval_steps_per_second": 88.119,
109
  "step": 18000
110
  },
111
  {
112
+ "epoch": 20.964360587002098,
113
+ "grad_norm": 2.2613022327423096,
114
  "learning_rate": 4.9985e-05,
115
+ "loss": 4.1842,
116
  "step": 20000
117
  },
118
  {
119
+ "epoch": 20.964360587002098,
120
+ "eval_loss": 4.690371513366699,
121
+ "eval_runtime": 0.9669,
122
+ "eval_samples_per_second": 1422.037,
123
+ "eval_steps_per_second": 88.942,
124
  "step": 20000
125
  }
126
  ],
127
  "logging_steps": 4000,
128
  "max_steps": 100000,
129
  "num_input_tokens_seen": 0,
130
+ "num_train_epochs": 105,
131
  "save_steps": 4000,
132
  "stateful_callbacks": {
133
  "TrainerControl": {
 
141
  "attributes": {}
142
  }
143
  },
144
+ "total_flos": 5171323153711104.0,
145
  "train_batch_size": 16,
146
  "trial_name": null,
147
  "trial_params": null
checkpoint-20000/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6af9de70859d62cbd3b61f71e8a5bc95702dbc6ddb62d2994641a31953e4ea9b
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c779122676b54107edc62ae0b9293c062733193c5f82a36a2bc097bca192814
3
  size 5368