akahana commited on
Commit
86b39d6
·
verified ·
1 Parent(s): 9b2b1e8

Upload folder using huggingface_hub

Browse files
all_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "total_flos": 2.5377466023936e+16,
4
+ "train_loss": 5.757713494451266,
5
+ "train_runtime": 5411.3319,
6
+ "train_samples": 97123,
7
+ "train_samples_per_second": 17.948,
8
+ "train_steps_per_second": 2.244
9
+ }
checkpoint-10000/config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "./gpt2/config.json",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPT2LMHeadModel"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "bos_token_id": 50256,
9
+ "embd_pdrop": 0.1,
10
+ "eos_token_id": 50256,
11
+ "initializer_range": 0.02,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_ctx": 1024,
15
+ "n_embd": 768,
16
+ "n_head": 6,
17
+ "n_inner": null,
18
+ "n_layer": 12,
19
+ "n_positions": 1024,
20
+ "reorder_and_upcast_attn": false,
21
+ "resid_pdrop": 0.1,
22
+ "scale_attn_by_inverse_layer_idx": false,
23
+ "scale_attn_weights": true,
24
+ "summary_activation": null,
25
+ "summary_first_dropout": 0.1,
26
+ "summary_proj_to_labels": true,
27
+ "summary_type": "cls_index",
28
+ "summary_use_proj": true,
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.47.1",
31
+ "use_cache": true,
32
+ "vocab_size": 50257
33
+ }
checkpoint-10000/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "eos_token_id": 50256,
5
+ "transformers_version": "4.47.1"
6
+ }
checkpoint-10000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-10000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ac5051c5733efec770dd14dc85b0511539f257ae3ec666430834d19ed6b958a
3
+ size 497774208
checkpoint-10000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d2a10dd40cc005c6f61574bc63727ddeb4d91f57ece5df84fc3c772b392ec90
3
+ size 995642298
checkpoint-10000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86265bc3b5bdd99cf8b09c33ce545311a548aa60acf41313375be0f2221b884c
3
+ size 14244
checkpoint-10000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e95c05e607b1866ea12ab07e69e308cf16182f4b085df3507b86b2f6e88c8eed
3
+ size 1064
checkpoint-10000/special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|endoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<|endoftext|>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
checkpoint-10000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-10000/tokenizer_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ }
13
+ },
14
+ "bos_token": "<|endoftext|>",
15
+ "clean_up_tokenization_spaces": false,
16
+ "eos_token": "<|endoftext|>",
17
+ "errors": "replace",
18
+ "extra_special_tokens": {},
19
+ "max_len": 512,
20
+ "model_max_length": 512,
21
+ "pad_token": null,
22
+ "tokenizer_class": "GPT2Tokenizer",
23
+ "unk_token": "<|endoftext|>"
24
+ }
checkpoint-10000/trainer_state.json ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.8236553825879253,
5
+ "eval_steps": 1000,
6
+ "global_step": 10000,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.08236553825879252,
13
+ "grad_norm": 2.806833267211914,
14
+ "learning_rate": 1.835268923482415e-05,
15
+ "loss": 7.3233,
16
+ "step": 1000
17
+ },
18
+ {
19
+ "epoch": 0.16473107651758503,
20
+ "grad_norm": 2.414405345916748,
21
+ "learning_rate": 1.67053784696483e-05,
22
+ "loss": 6.428,
23
+ "step": 2000
24
+ },
25
+ {
26
+ "epoch": 0.24709661477637757,
27
+ "grad_norm": 2.7336931228637695,
28
+ "learning_rate": 1.505806770447245e-05,
29
+ "loss": 6.121,
30
+ "step": 3000
31
+ },
32
+ {
33
+ "epoch": 0.32946215303517007,
34
+ "grad_norm": 2.707063674926758,
35
+ "learning_rate": 1.3410756939296599e-05,
36
+ "loss": 5.8761,
37
+ "step": 4000
38
+ },
39
+ {
40
+ "epoch": 0.4118276912939626,
41
+ "grad_norm": 2.918213129043579,
42
+ "learning_rate": 1.176344617412075e-05,
43
+ "loss": 5.7252,
44
+ "step": 5000
45
+ },
46
+ {
47
+ "epoch": 0.49419322955275513,
48
+ "grad_norm": 2.722219944000244,
49
+ "learning_rate": 1.0116135408944898e-05,
50
+ "loss": 5.5951,
51
+ "step": 6000
52
+ },
53
+ {
54
+ "epoch": 0.5765587678115477,
55
+ "grad_norm": 3.1989521980285645,
56
+ "learning_rate": 8.470471954534224e-06,
57
+ "loss": 5.4935,
58
+ "step": 7000
59
+ },
60
+ {
61
+ "epoch": 0.6589243060703401,
62
+ "grad_norm": 3.0108842849731445,
63
+ "learning_rate": 6.823161189358373e-06,
64
+ "loss": 5.4158,
65
+ "step": 8000
66
+ },
67
+ {
68
+ "epoch": 0.7412898443291327,
69
+ "grad_norm": 2.9963932037353516,
70
+ "learning_rate": 5.177497734947698e-06,
71
+ "loss": 5.3553,
72
+ "step": 9000
73
+ },
74
+ {
75
+ "epoch": 0.8236553825879253,
76
+ "grad_norm": 3.4173614978790283,
77
+ "learning_rate": 3.530186969771848e-06,
78
+ "loss": 5.2982,
79
+ "step": 10000
80
+ }
81
+ ],
82
+ "logging_steps": 1000,
83
+ "max_steps": 12141,
84
+ "num_input_tokens_seen": 0,
85
+ "num_train_epochs": 1,
86
+ "save_steps": 5000,
87
+ "stateful_callbacks": {
88
+ "TrainerControl": {
89
+ "args": {
90
+ "should_epoch_stop": false,
91
+ "should_evaluate": false,
92
+ "should_log": false,
93
+ "should_save": true,
94
+ "should_training_stop": false
95
+ },
96
+ "attributes": {}
97
+ }
98
+ },
99
+ "total_flos": 2.090336256e+16,
100
+ "train_batch_size": 8,
101
+ "trial_name": null,
102
+ "trial_params": null
103
+ }
checkpoint-10000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36ef3c13c88151db8e45a0b5710a309567ce4c7d0ec21fe92d39500047d43726
3
+ size 5368
checkpoint-10000/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-12141/config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "./gpt2/config.json",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPT2LMHeadModel"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "bos_token_id": 50256,
9
+ "embd_pdrop": 0.1,
10
+ "eos_token_id": 50256,
11
+ "initializer_range": 0.02,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_ctx": 1024,
15
+ "n_embd": 768,
16
+ "n_head": 6,
17
+ "n_inner": null,
18
+ "n_layer": 12,
19
+ "n_positions": 1024,
20
+ "reorder_and_upcast_attn": false,
21
+ "resid_pdrop": 0.1,
22
+ "scale_attn_by_inverse_layer_idx": false,
23
+ "scale_attn_weights": true,
24
+ "summary_activation": null,
25
+ "summary_first_dropout": 0.1,
26
+ "summary_proj_to_labels": true,
27
+ "summary_type": "cls_index",
28
+ "summary_use_proj": true,
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.47.1",
31
+ "use_cache": true,
32
+ "vocab_size": 50257
33
+ }
checkpoint-12141/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "eos_token_id": 50256,
5
+ "transformers_version": "4.47.1"
6
+ }
checkpoint-12141/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-12141/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:484497fd81aa8274235e37459d5194c43f2e078e52b81d56bb57a29edf21817b
3
+ size 497774208
checkpoint-12141/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87d14857c0b728b89df20385d432fcf48f6840795e77e21dd167987a60ec3a7c
3
+ size 995642298
checkpoint-12141/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0622c513c9f4497e3c3c3583048de7eb5a3c8c797f78f267464f88929b3e30b9
3
+ size 14244
checkpoint-12141/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ec524885c31fa20b26ef6f9fff97de491dba4aad844145b65ae384ff09dbddb
3
+ size 1064
checkpoint-12141/special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|endoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<|endoftext|>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
checkpoint-12141/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-12141/tokenizer_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ }
13
+ },
14
+ "bos_token": "<|endoftext|>",
15
+ "clean_up_tokenization_spaces": false,
16
+ "eos_token": "<|endoftext|>",
17
+ "errors": "replace",
18
+ "extra_special_tokens": {},
19
+ "max_len": 512,
20
+ "model_max_length": 512,
21
+ "pad_token": null,
22
+ "tokenizer_class": "GPT2Tokenizer",
23
+ "unk_token": "<|endoftext|>"
24
+ }
checkpoint-12141/trainer_state.json ADDED
@@ -0,0 +1,117 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
+ "eval_steps": 1000,
6
+ "global_step": 12141,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.08236553825879252,
13
+ "grad_norm": 2.806833267211914,
14
+ "learning_rate": 1.835268923482415e-05,
15
+ "loss": 7.3233,
16
+ "step": 1000
17
+ },
18
+ {
19
+ "epoch": 0.16473107651758503,
20
+ "grad_norm": 2.414405345916748,
21
+ "learning_rate": 1.67053784696483e-05,
22
+ "loss": 6.428,
23
+ "step": 2000
24
+ },
25
+ {
26
+ "epoch": 0.24709661477637757,
27
+ "grad_norm": 2.7336931228637695,
28
+ "learning_rate": 1.505806770447245e-05,
29
+ "loss": 6.121,
30
+ "step": 3000
31
+ },
32
+ {
33
+ "epoch": 0.32946215303517007,
34
+ "grad_norm": 2.707063674926758,
35
+ "learning_rate": 1.3410756939296599e-05,
36
+ "loss": 5.8761,
37
+ "step": 4000
38
+ },
39
+ {
40
+ "epoch": 0.4118276912939626,
41
+ "grad_norm": 2.918213129043579,
42
+ "learning_rate": 1.176344617412075e-05,
43
+ "loss": 5.7252,
44
+ "step": 5000
45
+ },
46
+ {
47
+ "epoch": 0.49419322955275513,
48
+ "grad_norm": 2.722219944000244,
49
+ "learning_rate": 1.0116135408944898e-05,
50
+ "loss": 5.5951,
51
+ "step": 6000
52
+ },
53
+ {
54
+ "epoch": 0.5765587678115477,
55
+ "grad_norm": 3.1989521980285645,
56
+ "learning_rate": 8.470471954534224e-06,
57
+ "loss": 5.4935,
58
+ "step": 7000
59
+ },
60
+ {
61
+ "epoch": 0.6589243060703401,
62
+ "grad_norm": 3.0108842849731445,
63
+ "learning_rate": 6.823161189358373e-06,
64
+ "loss": 5.4158,
65
+ "step": 8000
66
+ },
67
+ {
68
+ "epoch": 0.7412898443291327,
69
+ "grad_norm": 2.9963932037353516,
70
+ "learning_rate": 5.177497734947698e-06,
71
+ "loss": 5.3553,
72
+ "step": 9000
73
+ },
74
+ {
75
+ "epoch": 0.8236553825879253,
76
+ "grad_norm": 3.4173614978790283,
77
+ "learning_rate": 3.530186969771848e-06,
78
+ "loss": 5.2982,
79
+ "step": 10000
80
+ },
81
+ {
82
+ "epoch": 0.9060209208467177,
83
+ "grad_norm": 2.9771502017974854,
84
+ "learning_rate": 1.884523515361173e-06,
85
+ "loss": 5.2726,
86
+ "step": 11000
87
+ },
88
+ {
89
+ "epoch": 0.9883864591055103,
90
+ "grad_norm": 3.5910282135009766,
91
+ "learning_rate": 2.3721275018532246e-07,
92
+ "loss": 5.2688,
93
+ "step": 12000
94
+ }
95
+ ],
96
+ "logging_steps": 1000,
97
+ "max_steps": 12141,
98
+ "num_input_tokens_seen": 0,
99
+ "num_train_epochs": 1,
100
+ "save_steps": 5000,
101
+ "stateful_callbacks": {
102
+ "TrainerControl": {
103
+ "args": {
104
+ "should_epoch_stop": false,
105
+ "should_evaluate": false,
106
+ "should_log": false,
107
+ "should_save": true,
108
+ "should_training_stop": true
109
+ },
110
+ "attributes": {}
111
+ }
112
+ },
113
+ "total_flos": 2.5377466023936e+16,
114
+ "train_batch_size": 8,
115
+ "trial_name": null,
116
+ "trial_params": null
117
+ }
checkpoint-12141/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36ef3c13c88151db8e45a0b5710a309567ce4c7d0ec21fe92d39500047d43726
3
+ size 5368
checkpoint-12141/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "./gpt2/config.json",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPT2LMHeadModel"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "bos_token_id": 50256,
9
+ "embd_pdrop": 0.1,
10
+ "eos_token_id": 50256,
11
+ "initializer_range": 0.02,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_ctx": 1024,
15
+ "n_embd": 768,
16
+ "n_head": 6,
17
+ "n_inner": null,
18
+ "n_layer": 12,
19
+ "n_positions": 1024,
20
+ "reorder_and_upcast_attn": false,
21
+ "resid_pdrop": 0.1,
22
+ "scale_attn_by_inverse_layer_idx": false,
23
+ "scale_attn_weights": true,
24
+ "summary_activation": null,
25
+ "summary_first_dropout": 0.1,
26
+ "summary_proj_to_labels": true,
27
+ "summary_type": "cls_index",
28
+ "summary_use_proj": true,
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.47.1",
31
+ "use_cache": true,
32
+ "vocab_size": 50257
33
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "eos_token_id": 50256,
5
+ "transformers_version": "4.47.1"
6
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:484497fd81aa8274235e37459d5194c43f2e078e52b81d56bb57a29edf21817b
3
+ size 497774208
runs/Dec30_04-43-15_52a4379daf8e/events.out.tfevents.1735533808.52a4379daf8e.1729.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22217bff33d82cf46396d0c1dc825c4e1def6d4721b3e11c72cc291af78674cf
3
+ size 8089
special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|endoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<|endoftext|>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ }
13
+ },
14
+ "bos_token": "<|endoftext|>",
15
+ "clean_up_tokenization_spaces": false,
16
+ "eos_token": "<|endoftext|>",
17
+ "errors": "replace",
18
+ "extra_special_tokens": {},
19
+ "max_len": 512,
20
+ "model_max_length": 512,
21
+ "pad_token": null,
22
+ "tokenizer_class": "GPT2Tokenizer",
23
+ "unk_token": "<|endoftext|>"
24
+ }
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "total_flos": 2.5377466023936e+16,
4
+ "train_loss": 5.757713494451266,
5
+ "train_runtime": 5411.3319,
6
+ "train_samples": 97123,
7
+ "train_samples_per_second": 17.948,
8
+ "train_steps_per_second": 2.244
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,126 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
+ "eval_steps": 1000,
6
+ "global_step": 12141,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.08236553825879252,
13
+ "grad_norm": 2.806833267211914,
14
+ "learning_rate": 1.835268923482415e-05,
15
+ "loss": 7.3233,
16
+ "step": 1000
17
+ },
18
+ {
19
+ "epoch": 0.16473107651758503,
20
+ "grad_norm": 2.414405345916748,
21
+ "learning_rate": 1.67053784696483e-05,
22
+ "loss": 6.428,
23
+ "step": 2000
24
+ },
25
+ {
26
+ "epoch": 0.24709661477637757,
27
+ "grad_norm": 2.7336931228637695,
28
+ "learning_rate": 1.505806770447245e-05,
29
+ "loss": 6.121,
30
+ "step": 3000
31
+ },
32
+ {
33
+ "epoch": 0.32946215303517007,
34
+ "grad_norm": 2.707063674926758,
35
+ "learning_rate": 1.3410756939296599e-05,
36
+ "loss": 5.8761,
37
+ "step": 4000
38
+ },
39
+ {
40
+ "epoch": 0.4118276912939626,
41
+ "grad_norm": 2.918213129043579,
42
+ "learning_rate": 1.176344617412075e-05,
43
+ "loss": 5.7252,
44
+ "step": 5000
45
+ },
46
+ {
47
+ "epoch": 0.49419322955275513,
48
+ "grad_norm": 2.722219944000244,
49
+ "learning_rate": 1.0116135408944898e-05,
50
+ "loss": 5.5951,
51
+ "step": 6000
52
+ },
53
+ {
54
+ "epoch": 0.5765587678115477,
55
+ "grad_norm": 3.1989521980285645,
56
+ "learning_rate": 8.470471954534224e-06,
57
+ "loss": 5.4935,
58
+ "step": 7000
59
+ },
60
+ {
61
+ "epoch": 0.6589243060703401,
62
+ "grad_norm": 3.0108842849731445,
63
+ "learning_rate": 6.823161189358373e-06,
64
+ "loss": 5.4158,
65
+ "step": 8000
66
+ },
67
+ {
68
+ "epoch": 0.7412898443291327,
69
+ "grad_norm": 2.9963932037353516,
70
+ "learning_rate": 5.177497734947698e-06,
71
+ "loss": 5.3553,
72
+ "step": 9000
73
+ },
74
+ {
75
+ "epoch": 0.8236553825879253,
76
+ "grad_norm": 3.4173614978790283,
77
+ "learning_rate": 3.530186969771848e-06,
78
+ "loss": 5.2982,
79
+ "step": 10000
80
+ },
81
+ {
82
+ "epoch": 0.9060209208467177,
83
+ "grad_norm": 2.9771502017974854,
84
+ "learning_rate": 1.884523515361173e-06,
85
+ "loss": 5.2726,
86
+ "step": 11000
87
+ },
88
+ {
89
+ "epoch": 0.9883864591055103,
90
+ "grad_norm": 3.5910282135009766,
91
+ "learning_rate": 2.3721275018532246e-07,
92
+ "loss": 5.2688,
93
+ "step": 12000
94
+ },
95
+ {
96
+ "epoch": 1.0,
97
+ "step": 12141,
98
+ "total_flos": 2.5377466023936e+16,
99
+ "train_loss": 5.757713494451266,
100
+ "train_runtime": 5411.3319,
101
+ "train_samples_per_second": 17.948,
102
+ "train_steps_per_second": 2.244
103
+ }
104
+ ],
105
+ "logging_steps": 1000,
106
+ "max_steps": 12141,
107
+ "num_input_tokens_seen": 0,
108
+ "num_train_epochs": 1,
109
+ "save_steps": 5000,
110
+ "stateful_callbacks": {
111
+ "TrainerControl": {
112
+ "args": {
113
+ "should_epoch_stop": false,
114
+ "should_evaluate": false,
115
+ "should_log": false,
116
+ "should_save": true,
117
+ "should_training_stop": true
118
+ },
119
+ "attributes": {}
120
+ }
121
+ },
122
+ "total_flos": 2.5377466023936e+16,
123
+ "train_batch_size": 8,
124
+ "trial_name": null,
125
+ "trial_params": null
126
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36ef3c13c88151db8e45a0b5710a309567ce4c7d0ec21fe92d39500047d43726
3
+ size 5368
vocab.json ADDED
The diff for this file is too large to render. See raw diff