Slava commited on
Commit
baa99d8
·
verified ·
1 Parent(s): f7444f8

Training in progress, epoch 1

Browse files
Files changed (39) hide show
  1. logs/events.out.tfevents.1706261444.A-FVFFG2C4Q05P.18450.21 +2 -2
  2. logs/events.out.tfevents.1706265079.A-FVFFG2C4Q05P.18450.22 +3 -0
  3. logs/events.out.tfevents.1706265130.A-FVFFG2C4Q05P.18450.23 +3 -0
  4. logs/events.out.tfevents.1706265238.A-FVFFG2C4Q05P.18450.24 +3 -0
  5. logs/events.out.tfevents.1706267654.A-FVFFG2C4Q05P.18450.25 +3 -0
  6. model.safetensors +1 -1
  7. run-0/checkpoint-1054/model.safetensors +1 -1
  8. run-0/checkpoint-1054/optimizer.pt +1 -1
  9. run-0/checkpoint-1054/scheduler.pt +1 -1
  10. run-0/checkpoint-1054/trainer_state.json +21 -21
  11. run-0/checkpoint-1054/training_args.bin +1 -1
  12. run-0/checkpoint-1581/config.json +34 -0
  13. run-0/checkpoint-1581/model.safetensors +3 -0
  14. run-0/checkpoint-1581/optimizer.pt +3 -0
  15. run-0/checkpoint-1581/rng_state.pth +3 -0
  16. run-0/checkpoint-1581/scheduler.pt +3 -0
  17. run-0/checkpoint-1581/special_tokens_map.json +7 -0
  18. run-0/checkpoint-1581/tokenizer.json +0 -0
  19. run-0/checkpoint-1581/tokenizer_config.json +57 -0
  20. run-0/checkpoint-1581/trainer_state.json +71 -0
  21. run-0/checkpoint-1581/training_args.bin +3 -0
  22. run-0/checkpoint-1581/vocab.txt +0 -0
  23. run-1/checkpoint-527/model.safetensors +1 -1
  24. run-1/checkpoint-527/optimizer.pt +1 -1
  25. run-1/checkpoint-527/scheduler.pt +1 -1
  26. run-1/checkpoint-527/trainer_state.json +14 -14
  27. run-1/checkpoint-527/training_args.bin +1 -1
  28. run-20/checkpoint-2108/config.json +34 -0
  29. run-20/checkpoint-2108/model.safetensors +3 -0
  30. run-20/checkpoint-2108/optimizer.pt +3 -0
  31. run-20/checkpoint-2108/rng_state.pth +3 -0
  32. run-20/checkpoint-2108/scheduler.pt +3 -0
  33. run-20/checkpoint-2108/special_tokens_map.json +7 -0
  34. run-20/checkpoint-2108/tokenizer.json +0 -0
  35. run-20/checkpoint-2108/tokenizer_config.json +57 -0
  36. run-20/checkpoint-2108/trainer_state.json +86 -0
  37. run-20/checkpoint-2108/training_args.bin +3 -0
  38. run-20/checkpoint-2108/vocab.txt +0 -0
  39. training_args.bin +1 -1
logs/events.out.tfevents.1706261444.A-FVFFG2C4Q05P.18450.21 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:043d076a72417eb40faf963884d7a46845e8e57a8796f830bd5ebd888be89b43
3
- size 5396
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70b3f81981b99a22a2b4dfdc336bf0a355c612880099fe159bf5942bb7a1d708
3
+ size 6710
logs/events.out.tfevents.1706265079.A-FVFFG2C4Q05P.18450.22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a403755184bdee42a0aea585b652f2679e4c4a5f415c71cea5d865001de5735
3
+ size 4436
logs/events.out.tfevents.1706265130.A-FVFFG2C4Q05P.18450.23 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a802463db9f947701e76361445bf6812f86090831c1eef4aefda8c1608ded33
3
+ size 4435
logs/events.out.tfevents.1706265238.A-FVFFG2C4Q05P.18450.24 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:163f4c6ab12aa9ac3b9e3fdf1076137852ff3da1a7c34f63654266de25c8e7e3
3
+ size 6229
logs/events.out.tfevents.1706267654.A-FVFFG2C4Q05P.18450.25 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52e8a45c581fb5b3516f4cb4889b959907d751513b182ded83870e5a483005e9
3
+ size 4916
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6919625a06013ebf0b5d09d591d72059ecaacfd149aa6039620747ce130d8c89
3
  size 17549312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87568c0afc5ae5e4ff718507f68ad4943997f13f2f2353ed7480c5872143ba54
3
  size 17549312
run-0/checkpoint-1054/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:908638c4664a16b2be1e4378ceae2501bd6c5c8980ac2e8e1c440c2a13417b84
3
  size 17549312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a961afa01da16a24f2c49a0338f504c6969c49dea807a0163eb712ec19e5d2a5
3
  size 17549312
run-0/checkpoint-1054/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2820bc83db3bc05fcedc9b2e3c0814ef2ae544eaacfc2e20c898ca87db47f3a3
3
  size 35122373
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7081d3ab29e701c70aee56961c84534d54087a4a96d123288ab8300123271256
3
  size 35122373
run-0/checkpoint-1054/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:16e276f2a6209f1be03f0c62afe640f64e06f172ce4dd0006b8d39aabb5c7bee
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af1bf4840b29f03f6ccc15b1112ac510b2f94f820d4116b07a73af95409537ee
3
  size 627
run-0/checkpoint-1054/trainer_state.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "best_metric": 0.8245412844036697,
3
  "best_model_checkpoint": "tiny-bert-sst2-distilled/run-0/checkpoint-1054",
4
  "epoch": 2.0,
5
  "eval_steps": 500,
@@ -10,47 +10,47 @@
10
  "log_history": [
11
  {
12
  "epoch": 1.0,
13
- "learning_rate": 9.507440967713177e-05,
14
- "loss": 1.6043,
15
  "step": 527
16
  },
17
  {
18
  "epoch": 1.0,
19
- "eval_accuracy": 0.8107798165137615,
20
- "eval_loss": 1.3382474184036255,
21
- "eval_runtime": 7.7477,
22
- "eval_samples_per_second": 112.55,
23
- "eval_steps_per_second": 0.903,
24
  "step": 527
25
  },
26
  {
27
  "epoch": 2.0,
28
- "learning_rate": 6.338293978475452e-05,
29
- "loss": 0.7901,
30
  "step": 1054
31
  },
32
  {
33
  "epoch": 2.0,
34
- "eval_accuracy": 0.8245412844036697,
35
- "eval_loss": 1.3008999824523926,
36
- "eval_runtime": 7.5885,
37
- "eval_samples_per_second": 114.91,
38
- "eval_steps_per_second": 0.922,
39
  "step": 1054
40
  }
41
  ],
42
  "logging_steps": 500,
43
- "max_steps": 2108,
44
  "num_input_tokens_seen": 0,
45
- "num_train_epochs": 4,
46
  "save_steps": 500,
47
  "total_flos": 16162868129520.0,
48
  "train_batch_size": 128,
49
  "trial_name": null,
50
  "trial_params": {
51
- "alpha": 0.46638270752374766,
52
- "learning_rate": 0.00012676587956950903,
53
- "num_train_epochs": 4,
54
- "temperature": 20
55
  }
56
  }
 
1
  {
2
+ "best_metric": 0.8348623853211009,
3
  "best_model_checkpoint": "tiny-bert-sst2-distilled/run-0/checkpoint-1054",
4
  "epoch": 2.0,
5
  "eval_steps": 500,
 
10
  "log_history": [
11
  {
12
  "epoch": 1.0,
13
+ "learning_rate": 0.00017521452501048488,
14
+ "loss": 1.4337,
15
  "step": 527
16
  },
17
  {
18
  "epoch": 1.0,
19
+ "eval_accuracy": 0.8211009174311926,
20
+ "eval_loss": 1.309804081916809,
21
+ "eval_runtime": 7.7796,
22
+ "eval_samples_per_second": 112.088,
23
+ "eval_steps_per_second": 0.9,
24
  "step": 527
25
  },
26
  {
27
  "epoch": 2.0,
28
+ "learning_rate": 8.760726250524244e-05,
29
+ "loss": 0.6599,
30
  "step": 1054
31
  },
32
  {
33
  "epoch": 2.0,
34
+ "eval_accuracy": 0.8348623853211009,
35
+ "eval_loss": 1.4746311902999878,
36
+ "eval_runtime": 10.2236,
37
+ "eval_samples_per_second": 85.293,
38
+ "eval_steps_per_second": 0.685,
39
  "step": 1054
40
  }
41
  ],
42
  "logging_steps": 500,
43
+ "max_steps": 1581,
44
  "num_input_tokens_seen": 0,
45
+ "num_train_epochs": 3,
46
  "save_steps": 500,
47
  "total_flos": 16162868129520.0,
48
  "train_batch_size": 128,
49
  "trial_name": null,
50
  "trial_params": {
51
+ "alpha": 0.3828930992000458,
52
+ "learning_rate": 0.0002628217875157273,
53
+ "num_train_epochs": 3,
54
+ "temperature": 27
55
  }
56
  }
run-0/checkpoint-1054/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b149b247816d0d72efcc0fd576feb131fc41a6f49009ffa81023f2608190ff20
3
  size 4283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:816583fc783f300fde35f96ba82eb430b7b6945ff6b056fdd16b0cdcde5ee72d
3
  size 4283
run-0/checkpoint-1581/config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/bert_uncased_L-2_H-128_A-2",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 128,
11
+ "id2label": {
12
+ "0": "negative",
13
+ "1": "positive"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 512,
17
+ "label2id": {
18
+ "negative": "0",
19
+ "positive": "1"
20
+ },
21
+ "layer_norm_eps": 1e-12,
22
+ "max_position_embeddings": 512,
23
+ "model_type": "bert",
24
+ "num_attention_heads": 2,
25
+ "num_hidden_layers": 2,
26
+ "pad_token_id": 0,
27
+ "position_embedding_type": "absolute",
28
+ "problem_type": "single_label_classification",
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.37.1",
31
+ "type_vocab_size": 2,
32
+ "use_cache": true,
33
+ "vocab_size": 30522
34
+ }
run-0/checkpoint-1581/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f15d1272d95ede3fcd11cc383d65d589c95b329d53df77804d4b17b3294f11ff
3
+ size 17549312
run-0/checkpoint-1581/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f8914ab56eff118e58441cdca24f376e5bb1b78ae89f796ed69a1f0b804169a
3
+ size 35122373
run-0/checkpoint-1581/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8f1004ae2205b2fd82b8662eddc31a84c64c086e45e25c12d4a402a3d6a6e4d
3
+ size 13617
run-0/checkpoint-1581/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a223f5e1fbd77cbb00143de3f5585b5cf5790cbe69b7bbadea1af2c81ba6dabc
3
+ size 627
run-0/checkpoint-1581/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
run-0/checkpoint-1581/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
run-0/checkpoint-1581/tokenizer_config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 512,
50
+ "never_split": null,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "strip_accents": null,
54
+ "tokenize_chinese_chars": true,
55
+ "tokenizer_class": "BertTokenizer",
56
+ "unk_token": "[UNK]"
57
+ }
run-0/checkpoint-1581/trainer_state.json ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.8348623853211009,
3
+ "best_model_checkpoint": "tiny-bert-sst2-distilled/run-0/checkpoint-1054",
4
+ "epoch": 3.0,
5
+ "eval_steps": 500,
6
+ "global_step": 1581,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "learning_rate": 0.00017521452501048488,
14
+ "loss": 1.4337,
15
+ "step": 527
16
+ },
17
+ {
18
+ "epoch": 1.0,
19
+ "eval_accuracy": 0.8211009174311926,
20
+ "eval_loss": 1.309804081916809,
21
+ "eval_runtime": 7.7796,
22
+ "eval_samples_per_second": 112.088,
23
+ "eval_steps_per_second": 0.9,
24
+ "step": 527
25
+ },
26
+ {
27
+ "epoch": 2.0,
28
+ "learning_rate": 8.760726250524244e-05,
29
+ "loss": 0.6599,
30
+ "step": 1054
31
+ },
32
+ {
33
+ "epoch": 2.0,
34
+ "eval_accuracy": 0.8348623853211009,
35
+ "eval_loss": 1.4746311902999878,
36
+ "eval_runtime": 10.2236,
37
+ "eval_samples_per_second": 85.293,
38
+ "eval_steps_per_second": 0.685,
39
+ "step": 1054
40
+ },
41
+ {
42
+ "epoch": 3.0,
43
+ "learning_rate": 0.0,
44
+ "loss": 0.4906,
45
+ "step": 1581
46
+ },
47
+ {
48
+ "epoch": 3.0,
49
+ "eval_accuracy": 0.8348623853211009,
50
+ "eval_loss": 1.4973729848861694,
51
+ "eval_runtime": 8.919,
52
+ "eval_samples_per_second": 97.769,
53
+ "eval_steps_per_second": 0.785,
54
+ "step": 1581
55
+ }
56
+ ],
57
+ "logging_steps": 500,
58
+ "max_steps": 1581,
59
+ "num_input_tokens_seen": 0,
60
+ "num_train_epochs": 3,
61
+ "save_steps": 500,
62
+ "total_flos": 24238892698680.0,
63
+ "train_batch_size": 128,
64
+ "trial_name": null,
65
+ "trial_params": {
66
+ "alpha": 0.3828930992000458,
67
+ "learning_rate": 0.0002628217875157273,
68
+ "num_train_epochs": 3,
69
+ "temperature": 27
70
+ }
71
+ }
run-0/checkpoint-1581/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:816583fc783f300fde35f96ba82eb430b7b6945ff6b056fdd16b0cdcde5ee72d
3
+ size 4283
run-0/checkpoint-1581/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
run-1/checkpoint-527/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7bcd7d0aec1de757ac2cdfe0bc6f7c2b14a91f1a613b624a2f0093d80ca6355a
3
  size 17549312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87568c0afc5ae5e4ff718507f68ad4943997f13f2f2353ed7480c5872143ba54
3
  size 17549312
run-1/checkpoint-527/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d52e5de6bd8997ddf8ca3a2a1d76c2c6d88bf42dbd60546f3891044f28d441fa
3
  size 35122373
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e25ef891fdd6e5ede4e2fedcfccced541dbf26d70ac600869bb89500a94077d
3
  size 35122373
run-1/checkpoint-527/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4593c184c19c11a2f817520c5f5fdeb4ee376ac8d1c8d9f9c8b43128595f77d5
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b5b16144d365fbd5d227d61884860fc872eaa13bdf9a8f761164c135d3e8bd5
3
  size 627
run-1/checkpoint-527/trainer_state.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "best_metric": 0.6513761467889908,
3
  "best_model_checkpoint": "tiny-bert-sst2-distilled/run-1/checkpoint-527",
4
  "epoch": 1.0,
5
  "eval_steps": 500,
@@ -10,32 +10,32 @@
10
  "log_history": [
11
  {
12
  "epoch": 1.0,
13
- "learning_rate": 5.34694014828574e-06,
14
- "loss": 4.485,
15
  "step": 527
16
  },
17
  {
18
  "epoch": 1.0,
19
- "eval_accuracy": 0.6513761467889908,
20
- "eval_loss": 3.9686293601989746,
21
- "eval_runtime": 9.1719,
22
- "eval_samples_per_second": 95.073,
23
- "eval_steps_per_second": 0.763,
24
  "step": 527
25
  }
26
  ],
27
  "logging_steps": 500,
28
- "max_steps": 1054,
29
  "num_input_tokens_seen": 0,
30
- "num_train_epochs": 2,
31
  "save_steps": 500,
32
  "total_flos": 8069051778960.0,
33
  "train_batch_size": 128,
34
  "trial_name": null,
35
  "trial_params": {
36
- "alpha": 0.10705118382199963,
37
- "learning_rate": 1.069388029657148e-05,
38
- "num_train_epochs": 2,
39
- "temperature": 16
40
  }
41
  }
 
1
  {
2
+ "best_metric": 0.801605504587156,
3
  "best_model_checkpoint": "tiny-bert-sst2-distilled/run-1/checkpoint-527",
4
  "epoch": 1.0,
5
  "eval_steps": 500,
 
10
  "log_history": [
11
  {
12
  "epoch": 1.0,
13
+ "learning_rate": 0.0006875917567735723,
14
+ "loss": 0.9069,
15
  "step": 527
16
  },
17
  {
18
  "epoch": 1.0,
19
+ "eval_accuracy": 0.801605504587156,
20
+ "eval_loss": 1.1160032749176025,
21
+ "eval_runtime": 9.4507,
22
+ "eval_samples_per_second": 92.269,
23
+ "eval_steps_per_second": 0.741,
24
  "step": 527
25
  }
26
  ],
27
  "logging_steps": 500,
28
+ "max_steps": 5270,
29
  "num_input_tokens_seen": 0,
30
+ "num_train_epochs": 10,
31
  "save_steps": 500,
32
  "total_flos": 8069051778960.0,
33
  "train_batch_size": 128,
34
  "trial_name": null,
35
  "trial_params": {
36
+ "alpha": 0.5682681228180307,
37
+ "learning_rate": 0.0007639908408595248,
38
+ "num_train_epochs": 10,
39
+ "temperature": 8
40
  }
41
  }
run-1/checkpoint-527/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:70d09d03f84ec5e6f07e718b975ff8effb1d82efe0ca25ef5c8642902ec419c8
3
  size 4283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d99816c97dcbe7250039b2fe72e8bb2c24840867dcd296d9eb6598b2afd2670d
3
  size 4283
run-20/checkpoint-2108/config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/bert_uncased_L-2_H-128_A-2",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 128,
11
+ "id2label": {
12
+ "0": "negative",
13
+ "1": "positive"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 512,
17
+ "label2id": {
18
+ "negative": "0",
19
+ "positive": "1"
20
+ },
21
+ "layer_norm_eps": 1e-12,
22
+ "max_position_embeddings": 512,
23
+ "model_type": "bert",
24
+ "num_attention_heads": 2,
25
+ "num_hidden_layers": 2,
26
+ "pad_token_id": 0,
27
+ "position_embedding_type": "absolute",
28
+ "problem_type": "single_label_classification",
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.37.1",
31
+ "type_vocab_size": 2,
32
+ "use_cache": true,
33
+ "vocab_size": 30522
34
+ }
run-20/checkpoint-2108/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0633e4058de509cac3a60d0a6d60e05ab3aa4011607a39e5a464cfdc3f3758a8
3
+ size 17549312
run-20/checkpoint-2108/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96a4f3c21ead4b862b6583f29c96e128c3cae8817b7d70f2e2dac9f1b90b3d0d
3
+ size 35122373
run-20/checkpoint-2108/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2aae3bffb74208d13dfae4d8b66eb00427742bd84752ee081ae0f135d90f179b
3
+ size 13617
run-20/checkpoint-2108/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a5a8a2f6f77232a378bfada54063827d1b729af8eed15d1758da3d54babde72
3
+ size 627
run-20/checkpoint-2108/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
run-20/checkpoint-2108/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
run-20/checkpoint-2108/tokenizer_config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 512,
50
+ "never_split": null,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "strip_accents": null,
54
+ "tokenize_chinese_chars": true,
55
+ "tokenizer_class": "BertTokenizer",
56
+ "unk_token": "[UNK]"
57
+ }
run-20/checkpoint-2108/trainer_state.json ADDED
@@ -0,0 +1,86 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.8302752293577982,
3
+ "best_model_checkpoint": "tiny-bert-sst2-distilled/run-20/checkpoint-527",
4
+ "epoch": 4.0,
5
+ "eval_steps": 500,
6
+ "global_step": 2108,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "learning_rate": 0.0003127843891708784,
14
+ "loss": 0.4678,
15
+ "step": 527
16
+ },
17
+ {
18
+ "epoch": 1.0,
19
+ "eval_accuracy": 0.8302752293577982,
20
+ "eval_loss": 0.5433254241943359,
21
+ "eval_runtime": 7.3448,
22
+ "eval_samples_per_second": 118.723,
23
+ "eval_steps_per_second": 0.953,
24
+ "step": 527
25
+ },
26
+ {
27
+ "epoch": 2.0,
28
+ "learning_rate": 0.00020852292611391892,
29
+ "loss": 0.244,
30
+ "step": 1054
31
+ },
32
+ {
33
+ "epoch": 2.0,
34
+ "eval_accuracy": 0.8107798165137615,
35
+ "eval_loss": 0.7215615510940552,
36
+ "eval_runtime": 8.2206,
37
+ "eval_samples_per_second": 106.075,
38
+ "eval_steps_per_second": 0.852,
39
+ "step": 1054
40
+ },
41
+ {
42
+ "epoch": 3.0,
43
+ "learning_rate": 0.00010426146305695946,
44
+ "loss": 0.1704,
45
+ "step": 1581
46
+ },
47
+ {
48
+ "epoch": 3.0,
49
+ "eval_accuracy": 0.8107798165137615,
50
+ "eval_loss": 0.7814552783966064,
51
+ "eval_runtime": 7.9178,
52
+ "eval_samples_per_second": 110.132,
53
+ "eval_steps_per_second": 0.884,
54
+ "step": 1581
55
+ },
56
+ {
57
+ "epoch": 4.0,
58
+ "learning_rate": 0.0,
59
+ "loss": 0.1314,
60
+ "step": 2108
61
+ },
62
+ {
63
+ "epoch": 4.0,
64
+ "eval_accuracy": 0.8142201834862385,
65
+ "eval_loss": 0.8089592456817627,
66
+ "eval_runtime": 7.5447,
67
+ "eval_samples_per_second": 115.578,
68
+ "eval_steps_per_second": 0.928,
69
+ "step": 2108
70
+ }
71
+ ],
72
+ "logging_steps": 500,
73
+ "max_steps": 2108,
74
+ "num_input_tokens_seen": 0,
75
+ "num_train_epochs": 4,
76
+ "save_steps": 500,
77
+ "total_flos": 32344049138640.0,
78
+ "train_batch_size": 128,
79
+ "trial_name": null,
80
+ "trial_params": {
81
+ "alpha": 0.7544223674187817,
82
+ "learning_rate": 0.00041704585222783784,
83
+ "num_train_epochs": 4,
84
+ "temperature": 2
85
+ }
86
+ }
run-20/checkpoint-2108/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57025044228698d343c4abcfe17b305ab3bdbf218d07f2cabb4a08f29fbffe25
3
+ size 4283
run-20/checkpoint-2108/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:57025044228698d343c4abcfe17b305ab3bdbf218d07f2cabb4a08f29fbffe25
3
  size 4283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d99816c97dcbe7250039b2fe72e8bb2c24840867dcd296d9eb6598b2afd2670d
3
  size 4283