Training in progress, step 4000, checkpoint
Browse files
checkpoint-4000/model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 51007160
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b7774ea055ba209b0622ca46eba2a3108ce70c38abba893c84e6fb6ddc434c85
|
3 |
size 51007160
|
checkpoint-4000/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 102078202
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:775ec6fece3677d36b35721bcf6caa91f79fc3b10f92eb9821f9b349fbfd6160
|
3 |
size 102078202
|
checkpoint-4000/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14308
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d441e627a17e600f38f437f3b8bc81e1298392560003a20a62b75cbab2905880
|
3 |
size 14308
|
checkpoint-4000/tokenizer.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-4000/trainer_state.json
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
{
|
2 |
-
"best_metric": 5.
|
3 |
-
"best_model_checkpoint": "/home/p318482/babyLM_controlled/models_trained/
|
4 |
-
"epoch":
|
5 |
"eval_steps": 2000,
|
6 |
"global_step": 4000,
|
7 |
"is_hyper_param_search": false,
|
@@ -9,33 +9,33 @@
|
|
9 |
"is_world_process_zero": true,
|
10 |
"log_history": [
|
11 |
{
|
12 |
-
"epoch":
|
13 |
-
"eval_loss":
|
14 |
-
"eval_runtime": 0.
|
15 |
-
"eval_samples_per_second":
|
16 |
-
"eval_steps_per_second":
|
17 |
"step": 2000
|
18 |
},
|
19 |
{
|
20 |
-
"epoch":
|
21 |
-
"grad_norm": 1.
|
22 |
"learning_rate": 1e-05,
|
23 |
-
"loss": 6.
|
24 |
"step": 4000
|
25 |
},
|
26 |
{
|
27 |
-
"epoch":
|
28 |
-
"eval_loss": 5.
|
29 |
-
"eval_runtime": 0.
|
30 |
-
"eval_samples_per_second":
|
31 |
-
"eval_steps_per_second": 89.
|
32 |
"step": 4000
|
33 |
}
|
34 |
],
|
35 |
"logging_steps": 4000,
|
36 |
"max_steps": 100000,
|
37 |
"num_input_tokens_seen": 0,
|
38 |
-
"num_train_epochs":
|
39 |
"save_steps": 4000,
|
40 |
"stateful_callbacks": {
|
41 |
"TrainerControl": {
|
@@ -49,7 +49,7 @@
|
|
49 |
"attributes": {}
|
50 |
}
|
51 |
},
|
52 |
-
"total_flos":
|
53 |
"train_batch_size": 16,
|
54 |
"trial_name": null,
|
55 |
"trial_params": null
|
|
|
1 |
{
|
2 |
+
"best_metric": 5.884151935577393,
|
3 |
+
"best_model_checkpoint": "/home/p318482/babyLM_controlled/models_trained/de_clm/childes_30/checkpoint-4000",
|
4 |
+
"epoch": 4.1928721174004195,
|
5 |
"eval_steps": 2000,
|
6 |
"global_step": 4000,
|
7 |
"is_hyper_param_search": false,
|
|
|
9 |
"is_world_process_zero": true,
|
10 |
"log_history": [
|
11 |
{
|
12 |
+
"epoch": 2.0964360587002098,
|
13 |
+
"eval_loss": 7.102903366088867,
|
14 |
+
"eval_runtime": 0.9708,
|
15 |
+
"eval_samples_per_second": 1416.286,
|
16 |
+
"eval_steps_per_second": 88.582,
|
17 |
"step": 2000
|
18 |
},
|
19 |
{
|
20 |
+
"epoch": 4.1928721174004195,
|
21 |
+
"grad_norm": 1.3964662551879883,
|
22 |
"learning_rate": 1e-05,
|
23 |
+
"loss": 6.9987,
|
24 |
"step": 4000
|
25 |
},
|
26 |
{
|
27 |
+
"epoch": 4.1928721174004195,
|
28 |
+
"eval_loss": 5.884151935577393,
|
29 |
+
"eval_runtime": 0.966,
|
30 |
+
"eval_samples_per_second": 1423.408,
|
31 |
+
"eval_steps_per_second": 89.028,
|
32 |
"step": 4000
|
33 |
}
|
34 |
],
|
35 |
"logging_steps": 4000,
|
36 |
"max_steps": 100000,
|
37 |
"num_input_tokens_seen": 0,
|
38 |
+
"num_train_epochs": 105,
|
39 |
"save_steps": 4000,
|
40 |
"stateful_callbacks": {
|
41 |
"TrainerControl": {
|
|
|
49 |
"attributes": {}
|
50 |
}
|
51 |
},
|
52 |
+
"total_flos": 1034264024727552.0,
|
53 |
"train_batch_size": 16,
|
54 |
"trial_name": null,
|
55 |
"trial_params": null
|
checkpoint-4000/training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5368
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0c779122676b54107edc62ae0b9293c062733193c5f82a36a2bc097bca192814
|
3 |
size 5368
|