hemanth-kj commited on
Commit
12bb5b3
Β·
1 Parent(s): 909bd22

Training in progress, step 1875

Browse files
Files changed (32) hide show
  1. adapter_model.safetensors +1 -1
  2. config.json +3 -3
  3. last-checkpoint/adapter_model.safetensors +1 -1
  4. last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_0_mp_rank_00_model_states.pt +1 -1
  5. last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_0_mp_rank_00_optim_states.pt +1 -1
  6. last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_1_mp_rank_00_model_states.pt +1 -1
  7. last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_1_mp_rank_00_optim_states.pt +1 -1
  8. last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_2_mp_rank_00_model_states.pt +1 -1
  9. last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_2_mp_rank_00_optim_states.pt +1 -1
  10. last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_3_mp_rank_00_model_states.pt +1 -1
  11. last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_3_mp_rank_00_optim_states.pt +1 -1
  12. last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_4_mp_rank_00_model_states.pt +1 -1
  13. last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_4_mp_rank_00_optim_states.pt +1 -1
  14. last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_5_mp_rank_00_model_states.pt +1 -1
  15. last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_5_mp_rank_00_optim_states.pt +1 -1
  16. last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_6_mp_rank_00_model_states.pt +1 -1
  17. last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_6_mp_rank_00_optim_states.pt +1 -1
  18. last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_7_mp_rank_00_model_states.pt +1 -1
  19. last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_7_mp_rank_00_optim_states.pt +1 -1
  20. last-checkpoint/latest +1 -1
  21. last-checkpoint/rng_state_0.pth +1 -1
  22. last-checkpoint/rng_state_1.pth +1 -1
  23. last-checkpoint/rng_state_2.pth +1 -1
  24. last-checkpoint/rng_state_3.pth +1 -1
  25. last-checkpoint/rng_state_4.pth +1 -1
  26. last-checkpoint/rng_state_5.pth +1 -1
  27. last-checkpoint/rng_state_6.pth +1 -1
  28. last-checkpoint/rng_state_7.pth +1 -1
  29. last-checkpoint/trainer_state.json +1885 -373
  30. last-checkpoint/training_args.bin +1 -1
  31. last-checkpoint/zero_to_fp32.py +10 -6
  32. training_args.bin +1 -1
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da32d3ea1bfbd0ebd3b10a65710db858a2b52840601b85c79d7dea239a7ad777
3
  size 104900720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55936da28ccaa45e5ddca8dcdc226991b73479c6b56eb976c62b4594f394ab98
3
  size 104900720
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "elinas/llama-13b-hf-transformers-4.29",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -9,14 +9,14 @@
9
  "hidden_size": 5120,
10
  "initializer_range": 0.02,
11
  "intermediate_size": 13824,
12
- "max_position_embeddings": 2048,
13
  "model_type": "llama",
14
  "num_attention_heads": 40,
15
  "num_hidden_layers": 40,
16
  "num_key_value_heads": 40,
17
  "pad_token_id": 0,
18
  "pretraining_tp": 1,
19
- "rms_norm_eps": 1e-06,
20
  "rope_scaling": null,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "float32",
 
1
  {
2
+ "_name_or_path": "TheBloke/Llama-2-13B-fp16",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
9
  "hidden_size": 5120,
10
  "initializer_range": 0.02,
11
  "intermediate_size": 13824,
12
+ "max_position_embeddings": 4096,
13
  "model_type": "llama",
14
  "num_attention_heads": 40,
15
  "num_hidden_layers": 40,
16
  "num_key_value_heads": 40,
17
  "pad_token_id": 0,
18
  "pretraining_tp": 1,
19
+ "rms_norm_eps": 1e-05,
20
  "rope_scaling": null,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "float32",
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da32d3ea1bfbd0ebd3b10a65710db858a2b52840601b85c79d7dea239a7ad777
3
  size 104900720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55936da28ccaa45e5ddca8dcdc226991b73479c6b56eb976c62b4594f394ab98
3
  size 104900720
last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_0_mp_rank_00_model_states.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc7a512b9dd649ea7b8bcd0f4bfeba78db117a4d21241eb3c445f28b9b035b44
3
  size 6508524919
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b86c8546dd1666401b84d90ca0b94ef6cb6c0b73117ae6c1d687ddb1e11f648b
3
  size 6508524919
last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_0_mp_rank_00_optim_states.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c6ac434744b5670e20dbf7de0f019f547687454f89ea0280eb188b83b45f13ef
3
  size 39324734
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fee3a7fd29d3119d2b21bf40c50d75a2eebabc650ed9e0e16b2e8bcfe9449031
3
  size 39324734
last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_1_mp_rank_00_model_states.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0ae5fa3e3b1dcbb4a4e0fe8d8297960bf81e85883e8626388f2741f1256a7f42
3
  size 6508524919
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86353a275442ecc1818f29ee23975deedfc0c42b7dc80def1b26bff124460652
3
  size 6508524919
last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_1_mp_rank_00_optim_states.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8287daec126b9d2fb8e8cd51ec9da468856db2418f4582876b55fe61757ef61a
3
  size 39324734
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb9c8eabf344996ecedd94f084d70e33300335e97cdc132bc047b014185c6059
3
  size 39324734
last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_2_mp_rank_00_model_states.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5e4f6ed520b87d564d216f089f3527986451a9a7fda56fa546fd646514bd44f9
3
  size 6508524919
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25055a4966ab33becc0bcb18d051dd85b05d08342ee8e7609d2649be7f0fe9d4
3
  size 6508524919
last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_2_mp_rank_00_optim_states.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f4042166c4f941b64617579ea00f6d7f095a5be1bc6c0e8bbdf9c40550033e6b
3
  size 39324734
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1a1d5289fa7a41de645b8cb7890fb4ed7624b274892688cef21dfb5703cc831
3
  size 39324734
last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_3_mp_rank_00_model_states.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8f2f0a3215aac7e2ec10298d856832621b2db0c91916c00868411fc3298786ee
3
  size 6508524919
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da7d9d6595b20c78f8650a456533ca3ca7fd8a0d09c8b080efd314c39ab51fc7
3
  size 6508524919
last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_3_mp_rank_00_optim_states.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f92054239ef1b333a36e40ed281baf5e518dbf38687fd7707a744cc060ea4bfe
3
  size 39324734
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bef3cc71b28be9da43756109bd915e5a8693bc2e9f945c3a2c35f77d8fd5a36
3
  size 39324734
last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_4_mp_rank_00_model_states.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7807b566bad78986341cee14e7d4aa53bc8a0200ca5eb2d5f029a2beaf7ebf66
3
  size 6508524919
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4dd45dd6271af5951105ff05aa28f48e34d0aa71ad8e9374c030e8495033d274
3
  size 6508524919
last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_4_mp_rank_00_optim_states.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e0479c922106519c15b9049c6ef433091218afc9898d85ecf50d0bca0f00f1dd
3
  size 39324734
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9c7d905697ce3ebcb3108c7a8d9c377f03e508805123a11d78736db0628e3b0
3
  size 39324734
last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_5_mp_rank_00_model_states.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c1b68b41313f92d5461f5650971d0d7fc2021f831e8b209bbe7684a9fa6de036
3
  size 6508524919
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7214390febf54a55dfed233c4252b5a5e1fe4a2e5a1dce4c268ff0b8af78add1
3
  size 6508524919
last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_5_mp_rank_00_optim_states.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6d43038f68b214eaaf0de4696eac4ce4fbf8b203f0278eae4d39d6b03b2c8b5
3
  size 39324734
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d2e9a3e15eab72cb27e2ad212925c0bcc40c109704d5c12d76a00c783c0714b
3
  size 39324734
last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_6_mp_rank_00_model_states.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca1774553df058481dced20adaab498c5c060713fc551b7824be8b5b899199f0
3
  size 6508524919
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4d315b42177b90d73673c0917af534743f494256bd881f5858dd280d4432bbb
3
  size 6508524919
last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_6_mp_rank_00_optim_states.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0d9e28784bda3c89c935090afe7e9c4f28a0319586034f42c7a6ddb62cb23145
3
  size 39324734
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bb2afbc9c7cb4f347271f6e4391eb9c7956b9e8fc68c0117bf80cb52c8cba7c
3
  size 39324734
last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_7_mp_rank_00_model_states.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ad420ae915a4f4dda92d5e553c7fa52197f613dd9a32042c920b5f57993795db
3
  size 6508524919
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6a90eac5198209d06ad10d0085cb23471f4de6a307c6b2d74b5cb7067c96729
3
  size 6508524919
last-checkpoint/{global_step615 β†’ global_step1875}/zero_pp_rank_7_mp_rank_00_optim_states.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e0d966f44a532fe90a784275aa0ecbcca4250af5ac121375136da07434f1e132
3
  size 39324734
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1e2129e9ba4b98ab3bf5f0c3afc2e58206d9d375a9163c28300e85e1b38fe30
3
  size 39324734
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step615
 
1
+ global_step1875
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2650662faaa72b41472de1537f17ead9d812f09f4fb8cd2d674128a4712a4e7d
3
  size 21687
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d914bd28016f6122c43d9f2c83724d2eb3fb30e66e779af973bb2d54cc8392b
3
  size 21687
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a4627a7e7a7dfe2ee147a65e0d6015baf626890fa1b156301c335b9ad9e316e6
3
  size 21687
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6211fb1ee2fec138ceb8c47005fc9bc2418104c532ee8dfc1b97c6ff0a4d8ca6
3
  size 21687
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:26052e9c4d518cdd7d9b6100278576ee5087e52a6278f6b0143dacdb8bf18fdb
3
  size 21687
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e6e93d0e06b92d456cdf9ae8d458dc04fe62bcaff12dcfd66e34ec4d96610a3
3
  size 21687
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:979fe1e078e9bce07319930c2cf56f16954ac1c8ce9e6e4acceb06554c8cef58
3
  size 21687
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd94ec78519765c3ee8dc3b4493f42b5fd0c4cd2976fd129e1f8ac18f99ae6d2
3
  size 21687
last-checkpoint/rng_state_4.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dd01729cff1a8faf266bd665eeba2ab4f2ed3a9a830f734b2db4c2eae107dcff
3
  size 21687
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63751f591f898abcc08b990d99973e8f008e12981693a109dce30ef0cd238781
3
  size 21687
last-checkpoint/rng_state_5.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a1e8fe0589efc84bd5ff71928ffae22410544c7901ec04e9e97eb1a38525f7fe
3
  size 21687
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee6e3fa3cef42dc11afb890036889f87f425dfc577cbf9e1d5de01bb444eb373
3
  size 21687
last-checkpoint/rng_state_6.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:425f4a8de6e80c8b017685beb871dcc15ab951bf996eecf512432f3e6d9c9abf
3
  size 21687
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61d84c5c8cd6aa95d17a38e63e5b602a1ef6cf5067ed65357c055b82af7e9406
3
  size 21687
last-checkpoint/rng_state_7.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:08811a4460727a46d30b4089dffa5d5fea621645158464f2a8d71c7027d8688d
3
  size 21687
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1182410896d797ec61be8102c6e7744e694a891c867f4e04ed79c5737df19e2c
3
  size 21687
last-checkpoint/trainer_state.json CHANGED
@@ -1,754 +1,2266 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.0,
5
- "global_step": 615,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
- "epoch": 0.01,
12
- "learning_rate": 1.3979400086720373e-06,
13
- "loss": 1.8018,
14
  "step": 5
15
  },
16
  {
17
- "epoch": 0.02,
18
- "learning_rate": 1.9999999999999995e-06,
19
- "loss": 1.7649,
20
  "step": 10
21
  },
22
  {
23
- "epoch": 0.02,
24
- "learning_rate": 1.9934426229508195e-06,
25
- "loss": 1.7617,
26
  "step": 15
27
  },
28
  {
29
- "epoch": 0.03,
30
- "learning_rate": 1.9852459016393443e-06,
31
- "loss": 1.9239,
32
  "step": 20
33
  },
34
  {
35
- "epoch": 0.04,
36
- "learning_rate": 1.9770491803278686e-06,
37
- "loss": 1.7685,
38
  "step": 25
39
  },
40
  {
41
- "epoch": 0.05,
42
- "learning_rate": 1.9688524590163934e-06,
43
- "loss": 1.7113,
44
  "step": 30
45
  },
46
  {
47
- "epoch": 0.06,
48
- "learning_rate": 1.960655737704918e-06,
49
- "loss": 1.8364,
50
  "step": 35
51
  },
52
  {
53
- "epoch": 0.07,
54
- "learning_rate": 1.9524590163934425e-06,
55
- "loss": 1.7904,
56
  "step": 40
57
  },
58
  {
59
- "epoch": 0.07,
60
- "learning_rate": 1.9442622950819673e-06,
61
- "loss": 1.7484,
62
  "step": 45
63
  },
64
  {
65
- "epoch": 0.08,
66
- "learning_rate": 1.9360655737704916e-06,
67
- "loss": 1.694,
68
  "step": 50
69
  },
70
  {
71
- "epoch": 0.09,
72
- "learning_rate": 1.9278688524590163e-06,
73
- "loss": 1.7381,
74
  "step": 55
75
  },
76
  {
77
- "epoch": 0.1,
78
- "learning_rate": 1.919672131147541e-06,
79
- "loss": 1.7264,
80
  "step": 60
81
  },
82
  {
83
- "epoch": 0.11,
84
- "learning_rate": 1.9114754098360654e-06,
85
- "loss": 1.6833,
86
  "step": 65
87
  },
88
  {
89
- "epoch": 0.11,
90
- "learning_rate": 1.9032786885245902e-06,
91
- "loss": 1.7088,
92
  "step": 70
93
  },
94
  {
95
- "epoch": 0.12,
96
- "learning_rate": 1.8950819672131145e-06,
97
- "loss": 1.6512,
98
  "step": 75
99
  },
100
  {
101
- "epoch": 0.13,
102
- "learning_rate": 1.8868852459016393e-06,
103
- "loss": 1.5752,
104
  "step": 80
105
  },
106
  {
107
- "epoch": 0.14,
108
- "learning_rate": 1.8786885245901638e-06,
109
- "loss": 1.8083,
110
  "step": 85
111
  },
112
  {
113
- "epoch": 0.15,
114
- "learning_rate": 1.8704918032786886e-06,
115
- "loss": 1.7016,
116
  "step": 90
117
  },
118
  {
119
- "epoch": 0.15,
120
- "learning_rate": 1.862295081967213e-06,
121
- "loss": 1.6425,
122
  "step": 95
123
  },
124
  {
125
- "epoch": 0.16,
126
- "learning_rate": 1.8540983606557377e-06,
127
- "loss": 1.6506,
128
  "step": 100
129
  },
130
  {
131
- "epoch": 0.17,
132
- "learning_rate": 1.845901639344262e-06,
133
- "loss": 1.6875,
134
  "step": 105
135
  },
136
  {
137
- "epoch": 0.18,
138
- "learning_rate": 1.8377049180327868e-06,
139
- "loss": 1.667,
140
  "step": 110
141
  },
142
  {
143
- "epoch": 0.19,
144
- "learning_rate": 1.8295081967213113e-06,
145
- "loss": 1.7864,
146
  "step": 115
147
  },
148
  {
149
- "epoch": 0.2,
150
- "learning_rate": 1.8213114754098359e-06,
151
- "loss": 1.7222,
152
  "step": 120
153
  },
154
  {
155
- "epoch": 0.2,
156
- "learning_rate": 1.8131147540983606e-06,
157
- "loss": 1.6307,
158
  "step": 125
159
  },
160
  {
161
- "epoch": 0.21,
162
- "learning_rate": 1.8049180327868852e-06,
163
- "loss": 1.678,
164
  "step": 130
165
  },
166
  {
167
- "epoch": 0.22,
168
- "learning_rate": 1.7967213114754097e-06,
169
- "loss": 1.6302,
170
  "step": 135
171
  },
172
  {
173
- "epoch": 0.23,
174
- "learning_rate": 1.7885245901639343e-06,
175
- "loss": 1.7003,
176
  "step": 140
177
  },
178
  {
179
- "epoch": 0.24,
180
- "learning_rate": 1.780327868852459e-06,
181
- "loss": 1.6147,
182
  "step": 145
183
  },
184
  {
185
- "epoch": 0.24,
186
- "learning_rate": 1.7721311475409834e-06,
187
- "loss": 1.6948,
188
  "step": 150
189
  },
190
  {
191
- "epoch": 0.25,
192
- "learning_rate": 1.7639344262295081e-06,
193
- "loss": 1.6458,
194
  "step": 155
195
  },
196
  {
197
- "epoch": 0.26,
198
- "learning_rate": 1.7557377049180327e-06,
199
- "loss": 1.6142,
200
  "step": 160
201
  },
202
  {
203
- "epoch": 0.27,
204
- "learning_rate": 1.7475409836065572e-06,
205
- "loss": 1.5277,
206
  "step": 165
207
  },
208
  {
209
- "epoch": 0.28,
210
- "learning_rate": 1.7393442622950818e-06,
211
- "loss": 1.5558,
212
  "step": 170
213
  },
214
  {
215
- "epoch": 0.28,
216
- "learning_rate": 1.7311475409836065e-06,
217
- "loss": 1.56,
218
  "step": 175
219
  },
220
  {
221
- "epoch": 0.29,
222
- "learning_rate": 1.7229508196721309e-06,
223
- "loss": 1.6358,
224
  "step": 180
225
  },
226
  {
227
- "epoch": 0.3,
228
- "learning_rate": 1.7147540983606556e-06,
229
- "loss": 1.5567,
230
  "step": 185
231
  },
232
  {
233
- "epoch": 0.31,
234
- "learning_rate": 1.7065573770491804e-06,
235
- "loss": 1.4662,
236
  "step": 190
237
  },
238
  {
239
- "epoch": 0.32,
240
- "learning_rate": 1.6983606557377047e-06,
241
- "loss": 1.5566,
242
  "step": 195
243
  },
244
  {
245
- "epoch": 0.33,
246
- "learning_rate": 1.6901639344262295e-06,
247
- "loss": 1.5082,
248
  "step": 200
249
  },
250
  {
251
- "epoch": 0.33,
252
- "learning_rate": 1.681967213114754e-06,
253
- "loss": 1.5863,
254
  "step": 205
255
  },
256
  {
257
- "epoch": 0.34,
258
- "learning_rate": 1.6737704918032786e-06,
259
- "loss": 1.4774,
260
  "step": 210
261
  },
262
  {
263
- "epoch": 0.35,
264
- "learning_rate": 1.6655737704918031e-06,
265
- "loss": 1.4835,
266
  "step": 215
267
  },
268
  {
269
- "epoch": 0.36,
270
- "learning_rate": 1.6573770491803279e-06,
271
- "loss": 1.5662,
272
  "step": 220
273
  },
274
  {
275
- "epoch": 0.37,
276
- "learning_rate": 1.6491803278688522e-06,
277
- "loss": 1.5638,
278
  "step": 225
279
  },
280
  {
281
- "epoch": 0.37,
282
- "learning_rate": 1.640983606557377e-06,
283
- "loss": 1.5964,
284
  "step": 230
285
  },
286
  {
287
- "epoch": 0.38,
288
- "learning_rate": 1.6327868852459015e-06,
289
- "loss": 1.4099,
290
  "step": 235
291
  },
292
  {
293
- "epoch": 0.39,
294
- "learning_rate": 1.624590163934426e-06,
295
- "loss": 1.4246,
296
  "step": 240
297
  },
298
  {
299
- "epoch": 0.4,
300
- "learning_rate": 1.6163934426229508e-06,
301
- "loss": 1.4101,
302
  "step": 245
303
  },
304
  {
305
- "epoch": 0.41,
306
- "learning_rate": 1.6081967213114754e-06,
307
- "loss": 1.4219,
308
  "step": 250
309
  },
310
  {
311
- "epoch": 0.41,
312
- "learning_rate": 1.6e-06,
313
- "loss": 1.4839,
314
  "step": 255
315
  },
316
  {
317
- "epoch": 0.42,
318
- "learning_rate": 1.5918032786885245e-06,
319
- "loss": 1.4891,
320
  "step": 260
321
  },
322
  {
323
- "epoch": 0.43,
324
- "learning_rate": 1.5836065573770492e-06,
325
- "loss": 1.466,
326
  "step": 265
327
  },
328
  {
329
- "epoch": 0.44,
330
- "learning_rate": 1.5754098360655736e-06,
331
- "loss": 1.3792,
332
  "step": 270
333
  },
334
  {
335
- "epoch": 0.45,
336
- "learning_rate": 1.5672131147540983e-06,
337
- "loss": 1.4344,
338
  "step": 275
339
  },
340
  {
341
- "epoch": 0.46,
342
- "learning_rate": 1.5590163934426229e-06,
343
- "loss": 1.3924,
344
  "step": 280
345
  },
346
  {
347
- "epoch": 0.46,
348
- "learning_rate": 1.5508196721311474e-06,
349
- "loss": 1.3575,
350
  "step": 285
351
  },
352
  {
353
- "epoch": 0.47,
354
- "learning_rate": 1.542622950819672e-06,
355
- "loss": 1.3742,
356
  "step": 290
357
  },
358
  {
359
- "epoch": 0.48,
360
- "learning_rate": 1.5344262295081967e-06,
361
- "loss": 1.4584,
362
  "step": 295
363
  },
364
  {
365
- "epoch": 0.49,
366
- "learning_rate": 1.526229508196721e-06,
367
- "loss": 1.3866,
368
  "step": 300
369
  },
370
  {
371
- "epoch": 0.5,
372
- "learning_rate": 1.5180327868852458e-06,
373
- "loss": 1.4555,
374
  "step": 305
375
  },
376
  {
377
- "epoch": 0.5,
378
- "learning_rate": 1.5098360655737706e-06,
379
- "loss": 1.3399,
380
  "step": 310
381
  },
382
  {
383
- "epoch": 0.51,
384
- "learning_rate": 1.501639344262295e-06,
385
- "loss": 1.3791,
386
  "step": 315
387
  },
388
  {
389
- "epoch": 0.52,
390
- "learning_rate": 1.4934426229508197e-06,
391
- "loss": 1.3861,
392
  "step": 320
393
  },
394
  {
395
- "epoch": 0.53,
396
- "learning_rate": 1.4852459016393442e-06,
397
- "loss": 1.4207,
398
  "step": 325
399
  },
400
  {
401
- "epoch": 0.54,
402
- "learning_rate": 1.4770491803278688e-06,
403
- "loss": 1.424,
404
  "step": 330
405
  },
406
  {
407
- "epoch": 0.54,
408
- "learning_rate": 1.4688524590163933e-06,
409
- "loss": 1.408,
410
  "step": 335
411
  },
412
  {
413
- "epoch": 0.55,
414
- "learning_rate": 1.460655737704918e-06,
415
- "loss": 1.3347,
416
  "step": 340
417
  },
418
  {
419
- "epoch": 0.56,
420
- "learning_rate": 1.4524590163934424e-06,
421
- "loss": 1.4363,
422
  "step": 345
423
  },
424
  {
425
- "epoch": 0.57,
426
- "learning_rate": 1.4442622950819672e-06,
427
- "loss": 1.4262,
428
  "step": 350
429
  },
430
  {
431
- "epoch": 0.58,
432
- "learning_rate": 1.4360655737704917e-06,
433
- "loss": 1.4408,
434
  "step": 355
435
  },
436
  {
437
- "epoch": 0.59,
438
- "learning_rate": 1.4278688524590163e-06,
439
- "loss": 1.4739,
440
  "step": 360
441
  },
442
  {
443
- "epoch": 0.59,
444
- "learning_rate": 1.419672131147541e-06,
445
- "loss": 1.3784,
446
  "step": 365
447
  },
448
  {
449
- "epoch": 0.6,
450
- "learning_rate": 1.4114754098360656e-06,
451
- "loss": 1.3516,
452
  "step": 370
453
  },
454
  {
455
- "epoch": 0.61,
456
- "learning_rate": 1.4032786885245901e-06,
457
- "loss": 1.2991,
458
  "step": 375
459
  },
460
  {
461
- "epoch": 0.62,
462
- "learning_rate": 1.3950819672131147e-06,
463
- "loss": 1.3686,
464
  "step": 380
465
  },
466
  {
467
- "epoch": 0.63,
468
- "learning_rate": 1.3868852459016394e-06,
469
- "loss": 1.5018,
470
  "step": 385
471
  },
472
  {
473
- "epoch": 0.63,
474
- "learning_rate": 1.3786885245901638e-06,
475
- "loss": 1.3487,
476
  "step": 390
477
  },
478
  {
479
- "epoch": 0.64,
480
- "learning_rate": 1.3704918032786885e-06,
481
- "loss": 1.4386,
482
  "step": 395
483
  },
484
  {
485
- "epoch": 0.65,
486
- "learning_rate": 1.362295081967213e-06,
487
- "loss": 1.4099,
488
  "step": 400
489
  },
490
  {
491
- "epoch": 0.66,
492
- "learning_rate": 1.3540983606557376e-06,
493
- "loss": 1.3353,
494
  "step": 405
495
  },
496
  {
497
- "epoch": 0.67,
498
- "learning_rate": 1.3459016393442622e-06,
499
- "loss": 1.4481,
500
  "step": 410
501
  },
502
  {
503
- "epoch": 0.67,
504
- "learning_rate": 1.337704918032787e-06,
505
- "loss": 1.3743,
506
  "step": 415
507
  },
508
  {
509
- "epoch": 0.68,
510
- "learning_rate": 1.3295081967213113e-06,
511
- "loss": 1.3592,
512
  "step": 420
513
  },
514
  {
515
- "epoch": 0.69,
516
- "learning_rate": 1.321311475409836e-06,
517
- "loss": 1.3307,
518
  "step": 425
519
  },
520
  {
521
- "epoch": 0.7,
522
- "learning_rate": 1.3131147540983608e-06,
523
- "loss": 1.4015,
524
  "step": 430
525
  },
526
  {
527
- "epoch": 0.71,
528
- "learning_rate": 1.3049180327868851e-06,
529
- "loss": 1.4114,
530
  "step": 435
531
  },
532
  {
533
- "epoch": 0.72,
534
- "learning_rate": 1.2967213114754099e-06,
535
- "loss": 1.4648,
536
  "step": 440
537
  },
538
  {
539
- "epoch": 0.72,
540
- "learning_rate": 1.2885245901639344e-06,
541
- "loss": 1.4291,
542
  "step": 445
543
  },
544
  {
545
- "epoch": 0.73,
546
- "learning_rate": 1.280327868852459e-06,
547
- "loss": 1.3383,
548
  "step": 450
549
  },
550
  {
551
- "epoch": 0.74,
552
- "learning_rate": 1.2721311475409835e-06,
553
- "loss": 1.3256,
554
  "step": 455
555
  },
556
  {
557
- "epoch": 0.75,
558
- "learning_rate": 1.2639344262295083e-06,
559
- "loss": 1.347,
560
  "step": 460
561
  },
562
  {
563
- "epoch": 0.76,
564
- "learning_rate": 1.2557377049180326e-06,
565
- "loss": 1.4265,
566
  "step": 465
567
  },
568
  {
569
- "epoch": 0.76,
570
- "learning_rate": 1.2475409836065574e-06,
571
- "loss": 1.3616,
572
  "step": 470
573
  },
574
  {
575
- "epoch": 0.77,
576
- "learning_rate": 1.239344262295082e-06,
577
- "loss": 1.347,
578
  "step": 475
579
  },
580
  {
581
- "epoch": 0.78,
582
- "learning_rate": 1.2311475409836065e-06,
583
- "loss": 1.3326,
584
  "step": 480
585
  },
586
  {
587
- "epoch": 0.79,
588
- "learning_rate": 1.222950819672131e-06,
589
- "loss": 1.3019,
590
  "step": 485
591
  },
592
  {
593
- "epoch": 0.8,
594
- "learning_rate": 1.2147540983606558e-06,
595
- "loss": 1.4361,
596
  "step": 490
597
  },
598
  {
599
- "epoch": 0.8,
600
- "learning_rate": 1.2065573770491803e-06,
601
- "loss": 1.4145,
602
  "step": 495
603
  },
604
  {
605
- "epoch": 0.81,
606
- "learning_rate": 1.1983606557377049e-06,
607
- "loss": 1.381,
608
  "step": 500
609
  },
610
  {
611
- "epoch": 0.82,
612
- "learning_rate": 1.1901639344262296e-06,
613
- "loss": 1.422,
614
  "step": 505
615
  },
616
  {
617
- "epoch": 0.83,
618
- "learning_rate": 1.181967213114754e-06,
619
- "loss": 1.3907,
620
  "step": 510
621
  },
622
  {
623
- "epoch": 0.84,
624
- "learning_rate": 1.1737704918032787e-06,
625
- "loss": 1.4062,
626
  "step": 515
627
  },
628
  {
629
- "epoch": 0.85,
630
- "learning_rate": 1.1655737704918033e-06,
631
- "loss": 1.4302,
632
  "step": 520
633
  },
634
  {
635
- "epoch": 0.85,
636
- "learning_rate": 1.1573770491803278e-06,
637
- "loss": 1.4916,
638
  "step": 525
639
  },
640
  {
641
- "epoch": 0.86,
642
- "learning_rate": 1.1491803278688523e-06,
643
- "loss": 1.3817,
644
  "step": 530
645
  },
646
  {
647
- "epoch": 0.87,
648
- "learning_rate": 1.1409836065573771e-06,
649
- "loss": 1.3041,
650
  "step": 535
651
  },
652
  {
653
- "epoch": 0.88,
654
- "learning_rate": 1.1327868852459014e-06,
655
- "loss": 1.415,
656
  "step": 540
657
  },
658
  {
659
- "epoch": 0.89,
660
- "learning_rate": 1.1245901639344262e-06,
661
- "loss": 1.3001,
662
  "step": 545
663
  },
664
  {
665
- "epoch": 0.89,
666
- "learning_rate": 1.116393442622951e-06,
667
- "loss": 1.4388,
668
  "step": 550
669
  },
670
  {
671
- "epoch": 0.9,
672
- "learning_rate": 1.1081967213114753e-06,
673
- "loss": 1.3848,
674
  "step": 555
675
  },
676
  {
677
- "epoch": 0.91,
678
- "learning_rate": 1.1e-06,
679
- "loss": 1.4446,
680
  "step": 560
681
  },
682
  {
683
- "epoch": 0.92,
684
- "learning_rate": 1.0918032786885246e-06,
685
- "loss": 1.4944,
686
  "step": 565
687
  },
688
  {
689
- "epoch": 0.93,
690
- "learning_rate": 1.0836065573770492e-06,
691
- "loss": 1.4,
692
  "step": 570
693
  },
694
  {
695
- "epoch": 0.93,
696
- "learning_rate": 1.0754098360655737e-06,
697
- "loss": 1.3372,
698
  "step": 575
699
  },
700
  {
701
- "epoch": 0.94,
702
- "learning_rate": 1.0672131147540985e-06,
703
- "loss": 1.3619,
704
  "step": 580
705
  },
706
  {
707
- "epoch": 0.95,
708
- "learning_rate": 1.0590163934426228e-06,
709
- "loss": 1.3584,
710
  "step": 585
711
  },
712
  {
713
- "epoch": 0.96,
714
- "learning_rate": 1.0508196721311476e-06,
715
- "loss": 1.3189,
716
  "step": 590
717
  },
718
  {
719
- "epoch": 0.97,
720
- "learning_rate": 1.042622950819672e-06,
721
- "loss": 1.3524,
722
  "step": 595
723
  },
724
  {
725
- "epoch": 0.98,
726
- "learning_rate": 1.0344262295081966e-06,
727
- "loss": 1.3713,
728
  "step": 600
729
  },
730
  {
731
- "epoch": 0.98,
732
- "learning_rate": 1.0262295081967212e-06,
733
- "loss": 1.3962,
734
  "step": 605
735
  },
736
  {
737
- "epoch": 0.99,
738
- "learning_rate": 1.018032786885246e-06,
739
- "loss": 1.4458,
740
  "step": 610
741
  },
742
  {
743
- "epoch": 1.0,
744
- "learning_rate": 1.0098360655737705e-06,
745
- "loss": 1.3579,
746
  "step": 615
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
747
  }
748
  ],
749
- "max_steps": 1230,
750
  "num_train_epochs": 2,
751
- "total_flos": 204046325514240.0,
752
  "trial_name": null,
753
  "trial_params": null
754
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.7203227045716482,
5
+ "global_step": 1875,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
+ "epoch": 0.0,
12
+ "learning_rate": 0.00013979400086720374,
13
+ "loss": 1.8118,
14
  "step": 5
15
  },
16
  {
17
+ "epoch": 0.0,
18
+ "learning_rate": 0.00019999999999999998,
19
+ "loss": 1.6525,
20
  "step": 10
21
  },
22
  {
23
+ "epoch": 0.01,
24
+ "learning_rate": 0.0001998460354118553,
25
+ "loss": 1.5208,
26
  "step": 15
27
  },
28
  {
29
+ "epoch": 0.01,
30
+ "learning_rate": 0.00019965357967667439,
31
+ "loss": 1.4179,
32
  "step": 20
33
  },
34
  {
35
+ "epoch": 0.01,
36
+ "learning_rate": 0.00019946112394149347,
37
+ "loss": 1.5179,
38
  "step": 25
39
  },
40
  {
41
+ "epoch": 0.01,
42
+ "learning_rate": 0.00019926866820631255,
43
+ "loss": 1.4164,
44
  "step": 30
45
  },
46
  {
47
+ "epoch": 0.01,
48
+ "learning_rate": 0.00019907621247113163,
49
+ "loss": 1.3599,
50
  "step": 35
51
  },
52
  {
53
+ "epoch": 0.02,
54
+ "learning_rate": 0.00019888375673595074,
55
+ "loss": 1.4631,
56
  "step": 40
57
  },
58
  {
59
+ "epoch": 0.02,
60
+ "learning_rate": 0.00019869130100076983,
61
+ "loss": 1.4569,
62
  "step": 45
63
  },
64
  {
65
+ "epoch": 0.02,
66
+ "learning_rate": 0.00019849884526558894,
67
+ "loss": 1.4491,
68
  "step": 50
69
  },
70
  {
71
+ "epoch": 0.02,
72
+ "learning_rate": 0.00019830638953040802,
73
+ "loss": 1.4852,
74
  "step": 55
75
  },
76
  {
77
+ "epoch": 0.02,
78
+ "learning_rate": 0.0001981139337952271,
79
+ "loss": 1.5018,
80
  "step": 60
81
  },
82
  {
83
+ "epoch": 0.02,
84
+ "learning_rate": 0.00019792147806004618,
85
+ "loss": 1.4777,
86
  "step": 65
87
  },
88
  {
89
+ "epoch": 0.03,
90
+ "learning_rate": 0.0001977290223248653,
91
+ "loss": 1.4531,
92
  "step": 70
93
  },
94
  {
95
+ "epoch": 0.03,
96
+ "learning_rate": 0.00019753656658968438,
97
+ "loss": 1.3477,
98
  "step": 75
99
  },
100
  {
101
+ "epoch": 0.03,
102
+ "learning_rate": 0.0001973441108545035,
103
+ "loss": 1.5621,
104
  "step": 80
105
  },
106
  {
107
+ "epoch": 0.03,
108
+ "learning_rate": 0.00019715165511932257,
109
+ "loss": 1.3823,
110
  "step": 85
111
  },
112
  {
113
+ "epoch": 0.03,
114
+ "learning_rate": 0.00019695919938414165,
115
+ "loss": 1.4521,
116
  "step": 90
117
  },
118
  {
119
+ "epoch": 0.04,
120
+ "learning_rate": 0.00019676674364896076,
121
+ "loss": 1.4822,
122
  "step": 95
123
  },
124
  {
125
+ "epoch": 0.04,
126
+ "learning_rate": 0.00019657428791377982,
127
+ "loss": 1.4313,
128
  "step": 100
129
  },
130
  {
131
+ "epoch": 0.04,
132
+ "learning_rate": 0.00019638183217859893,
133
+ "loss": 1.3581,
134
  "step": 105
135
  },
136
  {
137
+ "epoch": 0.04,
138
+ "learning_rate": 0.000196189376443418,
139
+ "loss": 1.3917,
140
  "step": 110
141
  },
142
  {
143
+ "epoch": 0.04,
144
+ "learning_rate": 0.00019599692070823712,
145
+ "loss": 1.4632,
146
  "step": 115
147
  },
148
  {
149
+ "epoch": 0.05,
150
+ "learning_rate": 0.0001958044649730562,
151
+ "loss": 1.4535,
152
  "step": 120
153
  },
154
  {
155
+ "epoch": 0.05,
156
+ "learning_rate": 0.0001956120092378753,
157
+ "loss": 1.487,
158
  "step": 125
159
  },
160
  {
161
+ "epoch": 0.05,
162
+ "learning_rate": 0.0001954195535026944,
163
+ "loss": 1.3851,
164
  "step": 130
165
  },
166
  {
167
+ "epoch": 0.05,
168
+ "learning_rate": 0.00019522709776751348,
169
+ "loss": 1.3467,
170
  "step": 135
171
  },
172
  {
173
+ "epoch": 0.05,
174
+ "learning_rate": 0.00019503464203233256,
175
+ "loss": 1.4262,
176
  "step": 140
177
  },
178
  {
179
+ "epoch": 0.06,
180
+ "learning_rate": 0.00019484218629715167,
181
+ "loss": 1.326,
182
  "step": 145
183
  },
184
  {
185
+ "epoch": 0.06,
186
+ "learning_rate": 0.00019464973056197075,
187
+ "loss": 1.471,
188
  "step": 150
189
  },
190
  {
191
+ "epoch": 0.06,
192
+ "learning_rate": 0.00019445727482678984,
193
+ "loss": 1.4928,
194
  "step": 155
195
  },
196
  {
197
+ "epoch": 0.06,
198
+ "learning_rate": 0.00019426481909160895,
199
+ "loss": 1.3059,
200
  "step": 160
201
  },
202
  {
203
+ "epoch": 0.06,
204
+ "learning_rate": 0.00019407236335642803,
205
+ "loss": 1.3448,
206
  "step": 165
207
  },
208
  {
209
+ "epoch": 0.07,
210
+ "learning_rate": 0.0001938799076212471,
211
+ "loss": 1.4605,
212
  "step": 170
213
  },
214
  {
215
+ "epoch": 0.07,
216
+ "learning_rate": 0.0001936874518860662,
217
+ "loss": 1.3755,
218
  "step": 175
219
  },
220
  {
221
+ "epoch": 0.07,
222
+ "learning_rate": 0.0001934949961508853,
223
+ "loss": 1.4118,
224
  "step": 180
225
  },
226
  {
227
+ "epoch": 0.07,
228
+ "learning_rate": 0.0001933025404157044,
229
+ "loss": 1.302,
230
  "step": 185
231
  },
232
  {
233
+ "epoch": 0.07,
234
+ "learning_rate": 0.0001931100846805235,
235
+ "loss": 1.441,
236
  "step": 190
237
  },
238
  {
239
+ "epoch": 0.07,
240
+ "learning_rate": 0.00019291762894534258,
241
+ "loss": 1.4891,
242
  "step": 195
243
  },
244
  {
245
+ "epoch": 0.08,
246
+ "learning_rate": 0.0001927251732101617,
247
+ "loss": 1.4324,
248
  "step": 200
249
  },
250
  {
251
+ "epoch": 0.08,
252
+ "learning_rate": 0.00019253271747498077,
253
+ "loss": 1.3829,
254
  "step": 205
255
  },
256
  {
257
+ "epoch": 0.08,
258
+ "learning_rate": 0.00019234026173979986,
259
+ "loss": 1.3732,
260
  "step": 210
261
  },
262
  {
263
+ "epoch": 0.08,
264
+ "learning_rate": 0.00019214780600461894,
265
+ "loss": 1.3262,
266
  "step": 215
267
  },
268
  {
269
+ "epoch": 0.08,
270
+ "learning_rate": 0.00019195535026943802,
271
+ "loss": 1.3651,
272
  "step": 220
273
  },
274
  {
275
+ "epoch": 0.09,
276
+ "learning_rate": 0.00019176289453425713,
277
+ "loss": 1.4975,
278
  "step": 225
279
  },
280
  {
281
+ "epoch": 0.09,
282
+ "learning_rate": 0.00019157043879907621,
283
+ "loss": 1.341,
284
  "step": 230
285
  },
286
  {
287
+ "epoch": 0.09,
288
+ "learning_rate": 0.00019137798306389532,
289
+ "loss": 1.4237,
290
  "step": 235
291
  },
292
  {
293
+ "epoch": 0.09,
294
+ "learning_rate": 0.0001911855273287144,
295
+ "loss": 1.444,
296
  "step": 240
297
  },
298
  {
299
+ "epoch": 0.09,
300
+ "learning_rate": 0.0001909930715935335,
301
+ "loss": 1.4071,
302
  "step": 245
303
  },
304
  {
305
+ "epoch": 0.1,
306
+ "learning_rate": 0.00019080061585835257,
307
+ "loss": 1.3715,
308
  "step": 250
309
  },
310
  {
311
+ "epoch": 0.1,
312
+ "learning_rate": 0.00019060816012317168,
313
+ "loss": 1.3385,
314
  "step": 255
315
  },
316
  {
317
+ "epoch": 0.1,
318
+ "learning_rate": 0.00019041570438799076,
319
+ "loss": 1.4171,
320
  "step": 260
321
  },
322
  {
323
+ "epoch": 0.1,
324
+ "learning_rate": 0.00019022324865280987,
325
+ "loss": 1.3596,
326
  "step": 265
327
  },
328
  {
329
+ "epoch": 0.1,
330
+ "learning_rate": 0.00019003079291762896,
331
+ "loss": 1.3444,
332
  "step": 270
333
  },
334
  {
335
+ "epoch": 0.11,
336
+ "learning_rate": 0.00018983833718244807,
337
+ "loss": 1.4197,
338
  "step": 275
339
  },
340
  {
341
+ "epoch": 0.11,
342
+ "learning_rate": 0.00018964588144726715,
343
+ "loss": 1.3405,
344
  "step": 280
345
  },
346
  {
347
+ "epoch": 0.11,
348
+ "learning_rate": 0.0001894534257120862,
349
+ "loss": 1.4371,
350
  "step": 285
351
  },
352
  {
353
+ "epoch": 0.11,
354
+ "learning_rate": 0.00018926096997690532,
355
+ "loss": 1.3825,
356
  "step": 290
357
  },
358
  {
359
+ "epoch": 0.11,
360
+ "learning_rate": 0.0001890685142417244,
361
+ "loss": 1.4089,
362
  "step": 295
363
  },
364
  {
365
+ "epoch": 0.12,
366
+ "learning_rate": 0.0001888760585065435,
367
+ "loss": 1.4502,
368
  "step": 300
369
  },
370
  {
371
+ "epoch": 0.12,
372
+ "learning_rate": 0.0001886836027713626,
373
+ "loss": 1.3097,
374
  "step": 305
375
  },
376
  {
377
+ "epoch": 0.12,
378
+ "learning_rate": 0.0001884911470361817,
379
+ "loss": 1.3422,
380
  "step": 310
381
  },
382
  {
383
+ "epoch": 0.12,
384
+ "learning_rate": 0.00018829869130100078,
385
+ "loss": 1.4088,
386
  "step": 315
387
  },
388
  {
389
+ "epoch": 0.12,
390
+ "learning_rate": 0.00018810623556581987,
391
+ "loss": 1.3372,
392
  "step": 320
393
  },
394
  {
395
+ "epoch": 0.12,
396
+ "learning_rate": 0.00018791377983063895,
397
+ "loss": 1.3307,
398
  "step": 325
399
  },
400
  {
401
+ "epoch": 0.13,
402
+ "learning_rate": 0.00018772132409545806,
403
+ "loss": 1.4299,
404
  "step": 330
405
  },
406
  {
407
+ "epoch": 0.13,
408
+ "learning_rate": 0.00018752886836027714,
409
+ "loss": 1.4264,
410
  "step": 335
411
  },
412
  {
413
+ "epoch": 0.13,
414
+ "learning_rate": 0.00018733641262509625,
415
+ "loss": 1.3151,
416
  "step": 340
417
  },
418
  {
419
+ "epoch": 0.13,
420
+ "learning_rate": 0.00018714395688991533,
421
+ "loss": 1.3675,
422
  "step": 345
423
  },
424
  {
425
+ "epoch": 0.13,
426
+ "learning_rate": 0.00018695150115473442,
427
+ "loss": 1.3794,
428
  "step": 350
429
  },
430
  {
431
+ "epoch": 0.14,
432
+ "learning_rate": 0.00018675904541955353,
433
+ "loss": 1.3892,
434
  "step": 355
435
  },
436
  {
437
+ "epoch": 0.14,
438
+ "learning_rate": 0.00018656658968437258,
439
+ "loss": 1.4612,
440
  "step": 360
441
  },
442
  {
443
+ "epoch": 0.14,
444
+ "learning_rate": 0.0001863741339491917,
445
+ "loss": 1.283,
446
  "step": 365
447
  },
448
  {
449
+ "epoch": 0.14,
450
+ "learning_rate": 0.00018618167821401078,
451
+ "loss": 1.4235,
452
  "step": 370
453
  },
454
  {
455
+ "epoch": 0.14,
456
+ "learning_rate": 0.00018598922247882988,
457
+ "loss": 1.3036,
458
  "step": 375
459
  },
460
  {
461
+ "epoch": 0.15,
462
+ "learning_rate": 0.00018579676674364897,
463
+ "loss": 1.3886,
464
  "step": 380
465
  },
466
  {
467
+ "epoch": 0.15,
468
+ "learning_rate": 0.00018560431100846808,
469
+ "loss": 1.3226,
470
  "step": 385
471
  },
472
  {
473
+ "epoch": 0.15,
474
+ "learning_rate": 0.00018541185527328716,
475
+ "loss": 1.3966,
476
  "step": 390
477
  },
478
  {
479
+ "epoch": 0.15,
480
+ "learning_rate": 0.00018521939953810624,
481
+ "loss": 1.3835,
482
  "step": 395
483
  },
484
  {
485
+ "epoch": 0.15,
486
+ "learning_rate": 0.00018502694380292533,
487
+ "loss": 1.3544,
488
  "step": 400
489
  },
490
  {
491
+ "epoch": 0.16,
492
+ "learning_rate": 0.00018483448806774444,
493
+ "loss": 1.5449,
494
  "step": 405
495
  },
496
  {
497
+ "epoch": 0.16,
498
+ "learning_rate": 0.00018464203233256352,
499
+ "loss": 1.3814,
500
  "step": 410
501
  },
502
  {
503
+ "epoch": 0.16,
504
+ "learning_rate": 0.0001844495765973826,
505
+ "loss": 1.4272,
506
  "step": 415
507
  },
508
  {
509
+ "epoch": 0.16,
510
+ "learning_rate": 0.0001842571208622017,
511
+ "loss": 1.4256,
512
  "step": 420
513
  },
514
  {
515
+ "epoch": 0.16,
516
+ "learning_rate": 0.0001840646651270208,
517
+ "loss": 1.367,
518
  "step": 425
519
  },
520
  {
521
+ "epoch": 0.17,
522
+ "learning_rate": 0.00018387220939183988,
523
+ "loss": 1.3814,
524
  "step": 430
525
  },
526
  {
527
+ "epoch": 0.17,
528
+ "learning_rate": 0.00018367975365665896,
529
+ "loss": 1.351,
530
  "step": 435
531
  },
532
  {
533
+ "epoch": 0.17,
534
+ "learning_rate": 0.00018348729792147807,
535
+ "loss": 1.3966,
536
  "step": 440
537
  },
538
  {
539
+ "epoch": 0.17,
540
+ "learning_rate": 0.00018329484218629715,
541
+ "loss": 1.4355,
542
  "step": 445
543
  },
544
  {
545
+ "epoch": 0.17,
546
+ "learning_rate": 0.00018310238645111626,
547
+ "loss": 1.3635,
548
  "step": 450
549
  },
550
  {
551
+ "epoch": 0.17,
552
+ "learning_rate": 0.00018290993071593534,
553
+ "loss": 1.4373,
554
  "step": 455
555
  },
556
  {
557
+ "epoch": 0.18,
558
+ "learning_rate": 0.00018271747498075445,
559
+ "loss": 1.424,
560
  "step": 460
561
  },
562
  {
563
+ "epoch": 0.18,
564
+ "learning_rate": 0.00018252501924557354,
565
+ "loss": 1.4445,
566
  "step": 465
567
  },
568
  {
569
+ "epoch": 0.18,
570
+ "learning_rate": 0.00018233256351039262,
571
+ "loss": 1.4569,
572
  "step": 470
573
  },
574
  {
575
+ "epoch": 0.18,
576
+ "learning_rate": 0.0001821401077752117,
577
+ "loss": 1.2821,
578
  "step": 475
579
  },
580
  {
581
+ "epoch": 0.18,
582
+ "learning_rate": 0.00018194765204003079,
583
+ "loss": 1.4447,
584
  "step": 480
585
  },
586
  {
587
+ "epoch": 0.19,
588
+ "learning_rate": 0.0001817551963048499,
589
+ "loss": 1.5219,
590
  "step": 485
591
  },
592
  {
593
+ "epoch": 0.19,
594
+ "learning_rate": 0.00018156274056966898,
595
+ "loss": 1.3272,
596
  "step": 490
597
  },
598
  {
599
+ "epoch": 0.19,
600
+ "learning_rate": 0.0001813702848344881,
601
+ "loss": 1.3045,
602
  "step": 495
603
  },
604
  {
605
+ "epoch": 0.19,
606
+ "learning_rate": 0.00018117782909930717,
607
+ "loss": 1.4005,
608
  "step": 500
609
  },
610
  {
611
+ "epoch": 0.19,
612
+ "learning_rate": 0.00018098537336412625,
613
+ "loss": 1.277,
614
  "step": 505
615
  },
616
  {
617
+ "epoch": 0.2,
618
+ "learning_rate": 0.00018079291762894534,
619
+ "loss": 1.4171,
620
  "step": 510
621
  },
622
  {
623
+ "epoch": 0.2,
624
+ "learning_rate": 0.00018060046189376445,
625
+ "loss": 1.3951,
626
  "step": 515
627
  },
628
  {
629
+ "epoch": 0.2,
630
+ "learning_rate": 0.00018040800615858353,
631
+ "loss": 1.3163,
632
  "step": 520
633
  },
634
  {
635
+ "epoch": 0.2,
636
+ "learning_rate": 0.00018021555042340264,
637
+ "loss": 1.4673,
638
  "step": 525
639
  },
640
  {
641
+ "epoch": 0.2,
642
+ "learning_rate": 0.00018002309468822172,
643
+ "loss": 1.356,
644
  "step": 530
645
  },
646
  {
647
+ "epoch": 0.21,
648
+ "learning_rate": 0.00017983063895304083,
649
+ "loss": 1.4882,
650
  "step": 535
651
  },
652
  {
653
+ "epoch": 0.21,
654
+ "learning_rate": 0.00017963818321785991,
655
+ "loss": 1.4118,
656
  "step": 540
657
  },
658
  {
659
+ "epoch": 0.21,
660
+ "learning_rate": 0.00017944572748267897,
661
+ "loss": 1.4022,
662
  "step": 545
663
  },
664
  {
665
+ "epoch": 0.21,
666
+ "learning_rate": 0.00017925327174749808,
667
+ "loss": 1.4223,
668
  "step": 550
669
  },
670
  {
671
+ "epoch": 0.21,
672
+ "learning_rate": 0.00017906081601231716,
673
+ "loss": 1.3543,
674
  "step": 555
675
  },
676
  {
677
+ "epoch": 0.22,
678
+ "learning_rate": 0.00017886836027713627,
679
+ "loss": 1.2972,
680
  "step": 560
681
  },
682
  {
683
+ "epoch": 0.22,
684
+ "learning_rate": 0.00017867590454195535,
685
+ "loss": 1.3908,
686
  "step": 565
687
  },
688
  {
689
+ "epoch": 0.22,
690
+ "learning_rate": 0.00017848344880677446,
691
+ "loss": 1.4811,
692
  "step": 570
693
  },
694
  {
695
+ "epoch": 0.22,
696
+ "learning_rate": 0.00017829099307159355,
697
+ "loss": 1.4248,
698
  "step": 575
699
  },
700
  {
701
+ "epoch": 0.22,
702
+ "learning_rate": 0.00017809853733641263,
703
+ "loss": 1.4548,
704
  "step": 580
705
  },
706
  {
707
+ "epoch": 0.22,
708
+ "learning_rate": 0.0001779060816012317,
709
+ "loss": 1.4073,
710
  "step": 585
711
  },
712
  {
713
+ "epoch": 0.23,
714
+ "learning_rate": 0.00017771362586605082,
715
+ "loss": 1.3499,
716
  "step": 590
717
  },
718
  {
719
+ "epoch": 0.23,
720
+ "learning_rate": 0.0001775211701308699,
721
+ "loss": 1.2826,
722
  "step": 595
723
  },
724
  {
725
+ "epoch": 0.23,
726
+ "learning_rate": 0.00017732871439568902,
727
+ "loss": 1.4861,
728
  "step": 600
729
  },
730
  {
731
+ "epoch": 0.23,
732
+ "learning_rate": 0.0001771362586605081,
733
+ "loss": 1.4161,
734
  "step": 605
735
  },
736
  {
737
+ "epoch": 0.23,
738
+ "learning_rate": 0.00017694380292532718,
739
+ "loss": 1.3263,
740
  "step": 610
741
  },
742
  {
743
+ "epoch": 0.24,
744
+ "learning_rate": 0.00017675134719014626,
745
+ "loss": 1.3883,
746
  "step": 615
747
+ },
748
+ {
749
+ "epoch": 0.24,
750
+ "learning_rate": 0.00017655889145496535,
751
+ "loss": 1.3328,
752
+ "step": 620
753
+ },
754
+ {
755
+ "epoch": 0.24,
756
+ "learning_rate": 0.00017636643571978446,
757
+ "loss": 1.3174,
758
+ "step": 625
759
+ },
760
+ {
761
+ "epoch": 0.24,
762
+ "learning_rate": 0.00017617397998460354,
763
+ "loss": 1.3876,
764
+ "step": 630
765
+ },
766
+ {
767
+ "epoch": 0.24,
768
+ "learning_rate": 0.00017598152424942265,
769
+ "loss": 1.3541,
770
+ "step": 635
771
+ },
772
+ {
773
+ "epoch": 0.25,
774
+ "learning_rate": 0.00017578906851424173,
775
+ "loss": 1.3319,
776
+ "step": 640
777
+ },
778
+ {
779
+ "epoch": 0.25,
780
+ "learning_rate": 0.00017559661277906084,
781
+ "loss": 1.2886,
782
+ "step": 645
783
+ },
784
+ {
785
+ "epoch": 0.25,
786
+ "learning_rate": 0.00017540415704387992,
787
+ "loss": 1.3338,
788
+ "step": 650
789
+ },
790
+ {
791
+ "epoch": 0.25,
792
+ "learning_rate": 0.000175211701308699,
793
+ "loss": 1.4036,
794
+ "step": 655
795
+ },
796
+ {
797
+ "epoch": 0.25,
798
+ "learning_rate": 0.0001750192455735181,
799
+ "loss": 1.4543,
800
+ "step": 660
801
+ },
802
+ {
803
+ "epoch": 0.26,
804
+ "learning_rate": 0.0001748267898383372,
805
+ "loss": 1.2902,
806
+ "step": 665
807
+ },
808
+ {
809
+ "epoch": 0.26,
810
+ "learning_rate": 0.00017463433410315628,
811
+ "loss": 1.3255,
812
+ "step": 670
813
+ },
814
+ {
815
+ "epoch": 0.26,
816
+ "learning_rate": 0.00017444187836797537,
817
+ "loss": 1.2875,
818
+ "step": 675
819
+ },
820
+ {
821
+ "epoch": 0.26,
822
+ "learning_rate": 0.00017424942263279448,
823
+ "loss": 1.3105,
824
+ "step": 680
825
+ },
826
+ {
827
+ "epoch": 0.26,
828
+ "learning_rate": 0.00017405696689761356,
829
+ "loss": 1.2533,
830
+ "step": 685
831
+ },
832
+ {
833
+ "epoch": 0.27,
834
+ "learning_rate": 0.00017386451116243264,
835
+ "loss": 1.3827,
836
+ "step": 690
837
+ },
838
+ {
839
+ "epoch": 0.27,
840
+ "learning_rate": 0.00017367205542725172,
841
+ "loss": 1.3549,
842
+ "step": 695
843
+ },
844
+ {
845
+ "epoch": 0.27,
846
+ "learning_rate": 0.00017347959969207083,
847
+ "loss": 1.4396,
848
+ "step": 700
849
+ },
850
+ {
851
+ "epoch": 0.27,
852
+ "learning_rate": 0.00017328714395688992,
853
+ "loss": 1.3179,
854
+ "step": 705
855
+ },
856
+ {
857
+ "epoch": 0.27,
858
+ "learning_rate": 0.00017309468822170903,
859
+ "loss": 1.356,
860
+ "step": 710
861
+ },
862
+ {
863
+ "epoch": 0.27,
864
+ "learning_rate": 0.0001729022324865281,
865
+ "loss": 1.2499,
866
+ "step": 715
867
+ },
868
+ {
869
+ "epoch": 0.28,
870
+ "learning_rate": 0.00017270977675134722,
871
+ "loss": 1.4139,
872
+ "step": 720
873
+ },
874
+ {
875
+ "epoch": 0.28,
876
+ "learning_rate": 0.0001725173210161663,
877
+ "loss": 1.249,
878
+ "step": 725
879
+ },
880
+ {
881
+ "epoch": 0.28,
882
+ "learning_rate": 0.00017232486528098538,
883
+ "loss": 1.3054,
884
+ "step": 730
885
+ },
886
+ {
887
+ "epoch": 0.28,
888
+ "learning_rate": 0.00017213240954580447,
889
+ "loss": 1.3415,
890
+ "step": 735
891
+ },
892
+ {
893
+ "epoch": 0.28,
894
+ "learning_rate": 0.00017193995381062355,
895
+ "loss": 1.3323,
896
+ "step": 740
897
+ },
898
+ {
899
+ "epoch": 0.29,
900
+ "learning_rate": 0.00017174749807544266,
901
+ "loss": 1.3559,
902
+ "step": 745
903
+ },
904
+ {
905
+ "epoch": 0.29,
906
+ "learning_rate": 0.00017155504234026174,
907
+ "loss": 1.3771,
908
+ "step": 750
909
+ },
910
+ {
911
+ "epoch": 0.29,
912
+ "learning_rate": 0.00017136258660508085,
913
+ "loss": 1.3811,
914
+ "step": 755
915
+ },
916
+ {
917
+ "epoch": 0.29,
918
+ "learning_rate": 0.00017117013086989993,
919
+ "loss": 1.3644,
920
+ "step": 760
921
+ },
922
+ {
923
+ "epoch": 0.29,
924
+ "learning_rate": 0.00017097767513471902,
925
+ "loss": 1.2619,
926
+ "step": 765
927
+ },
928
+ {
929
+ "epoch": 0.3,
930
+ "learning_rate": 0.0001707852193995381,
931
+ "loss": 1.3795,
932
+ "step": 770
933
+ },
934
+ {
935
+ "epoch": 0.3,
936
+ "learning_rate": 0.0001705927636643572,
937
+ "loss": 1.4482,
938
+ "step": 775
939
+ },
940
+ {
941
+ "epoch": 0.3,
942
+ "learning_rate": 0.0001704003079291763,
943
+ "loss": 1.3213,
944
+ "step": 780
945
+ },
946
+ {
947
+ "epoch": 0.3,
948
+ "learning_rate": 0.0001702078521939954,
949
+ "loss": 1.3406,
950
+ "step": 785
951
+ },
952
+ {
953
+ "epoch": 0.3,
954
+ "learning_rate": 0.00017001539645881449,
955
+ "loss": 1.254,
956
+ "step": 790
957
+ },
958
+ {
959
+ "epoch": 0.31,
960
+ "learning_rate": 0.0001698229407236336,
961
+ "loss": 1.4489,
962
+ "step": 795
963
+ },
964
+ {
965
+ "epoch": 0.31,
966
+ "learning_rate": 0.00016963048498845268,
967
+ "loss": 1.3327,
968
+ "step": 800
969
+ },
970
+ {
971
+ "epoch": 0.31,
972
+ "learning_rate": 0.00016943802925327173,
973
+ "loss": 1.2697,
974
+ "step": 805
975
+ },
976
+ {
977
+ "epoch": 0.31,
978
+ "learning_rate": 0.00016924557351809084,
979
+ "loss": 1.3808,
980
+ "step": 810
981
+ },
982
+ {
983
+ "epoch": 0.31,
984
+ "learning_rate": 0.00016905311778290993,
985
+ "loss": 1.3519,
986
+ "step": 815
987
+ },
988
+ {
989
+ "epoch": 0.32,
990
+ "learning_rate": 0.00016886066204772904,
991
+ "loss": 1.3051,
992
+ "step": 820
993
+ },
994
+ {
995
+ "epoch": 0.32,
996
+ "learning_rate": 0.00016866820631254812,
997
+ "loss": 1.3251,
998
+ "step": 825
999
+ },
1000
+ {
1001
+ "epoch": 0.32,
1002
+ "learning_rate": 0.00016847575057736723,
1003
+ "loss": 1.3595,
1004
+ "step": 830
1005
+ },
1006
+ {
1007
+ "epoch": 0.32,
1008
+ "learning_rate": 0.0001682832948421863,
1009
+ "loss": 1.366,
1010
+ "step": 835
1011
+ },
1012
+ {
1013
+ "epoch": 0.32,
1014
+ "learning_rate": 0.0001680908391070054,
1015
+ "loss": 1.3157,
1016
+ "step": 840
1017
+ },
1018
+ {
1019
+ "epoch": 0.32,
1020
+ "learning_rate": 0.00016789838337182448,
1021
+ "loss": 1.3809,
1022
+ "step": 845
1023
+ },
1024
+ {
1025
+ "epoch": 0.33,
1026
+ "learning_rate": 0.0001677059276366436,
1027
+ "loss": 1.3305,
1028
+ "step": 850
1029
+ },
1030
+ {
1031
+ "epoch": 0.33,
1032
+ "learning_rate": 0.00016751347190146267,
1033
+ "loss": 1.3454,
1034
+ "step": 855
1035
+ },
1036
+ {
1037
+ "epoch": 0.33,
1038
+ "learning_rate": 0.00016732101616628175,
1039
+ "loss": 1.3914,
1040
+ "step": 860
1041
+ },
1042
+ {
1043
+ "epoch": 0.33,
1044
+ "learning_rate": 0.00016712856043110086,
1045
+ "loss": 1.2801,
1046
+ "step": 865
1047
+ },
1048
+ {
1049
+ "epoch": 0.33,
1050
+ "learning_rate": 0.00016693610469591995,
1051
+ "loss": 1.2451,
1052
+ "step": 870
1053
+ },
1054
+ {
1055
+ "epoch": 0.34,
1056
+ "learning_rate": 0.00016674364896073903,
1057
+ "loss": 1.3802,
1058
+ "step": 875
1059
+ },
1060
+ {
1061
+ "epoch": 0.34,
1062
+ "learning_rate": 0.0001665511932255581,
1063
+ "loss": 1.383,
1064
+ "step": 880
1065
+ },
1066
+ {
1067
+ "epoch": 0.34,
1068
+ "learning_rate": 0.00016635873749037722,
1069
+ "loss": 1.3572,
1070
+ "step": 885
1071
+ },
1072
+ {
1073
+ "epoch": 0.34,
1074
+ "learning_rate": 0.0001661662817551963,
1075
+ "loss": 1.381,
1076
+ "step": 890
1077
+ },
1078
+ {
1079
+ "epoch": 0.34,
1080
+ "learning_rate": 0.0001659738260200154,
1081
+ "loss": 1.253,
1082
+ "step": 895
1083
+ },
1084
+ {
1085
+ "epoch": 0.35,
1086
+ "learning_rate": 0.0001657813702848345,
1087
+ "loss": 1.3824,
1088
+ "step": 900
1089
+ },
1090
+ {
1091
+ "epoch": 0.35,
1092
+ "learning_rate": 0.0001655889145496536,
1093
+ "loss": 1.3366,
1094
+ "step": 905
1095
+ },
1096
+ {
1097
+ "epoch": 0.35,
1098
+ "learning_rate": 0.0001653964588144727,
1099
+ "loss": 1.2716,
1100
+ "step": 910
1101
+ },
1102
+ {
1103
+ "epoch": 0.35,
1104
+ "learning_rate": 0.00016520400307929177,
1105
+ "loss": 1.317,
1106
+ "step": 915
1107
+ },
1108
+ {
1109
+ "epoch": 0.35,
1110
+ "learning_rate": 0.00016501154734411085,
1111
+ "loss": 1.27,
1112
+ "step": 920
1113
+ },
1114
+ {
1115
+ "epoch": 0.36,
1116
+ "learning_rate": 0.00016481909160892994,
1117
+ "loss": 1.463,
1118
+ "step": 925
1119
+ },
1120
+ {
1121
+ "epoch": 0.36,
1122
+ "learning_rate": 0.00016462663587374905,
1123
+ "loss": 1.3101,
1124
+ "step": 930
1125
+ },
1126
+ {
1127
+ "epoch": 0.36,
1128
+ "learning_rate": 0.00016443418013856813,
1129
+ "loss": 1.3305,
1130
+ "step": 935
1131
+ },
1132
+ {
1133
+ "epoch": 0.36,
1134
+ "learning_rate": 0.00016424172440338724,
1135
+ "loss": 1.2637,
1136
+ "step": 940
1137
+ },
1138
+ {
1139
+ "epoch": 0.36,
1140
+ "learning_rate": 0.00016404926866820632,
1141
+ "loss": 1.3564,
1142
+ "step": 945
1143
+ },
1144
+ {
1145
+ "epoch": 0.36,
1146
+ "learning_rate": 0.0001638568129330254,
1147
+ "loss": 1.1473,
1148
+ "step": 950
1149
+ },
1150
+ {
1151
+ "epoch": 0.37,
1152
+ "learning_rate": 0.0001636643571978445,
1153
+ "loss": 1.2369,
1154
+ "step": 955
1155
+ },
1156
+ {
1157
+ "epoch": 0.37,
1158
+ "learning_rate": 0.0001634719014626636,
1159
+ "loss": 1.2508,
1160
+ "step": 960
1161
+ },
1162
+ {
1163
+ "epoch": 0.37,
1164
+ "learning_rate": 0.00016327944572748268,
1165
+ "loss": 1.3019,
1166
+ "step": 965
1167
+ },
1168
+ {
1169
+ "epoch": 0.37,
1170
+ "learning_rate": 0.0001630869899923018,
1171
+ "loss": 1.2893,
1172
+ "step": 970
1173
+ },
1174
+ {
1175
+ "epoch": 0.37,
1176
+ "learning_rate": 0.00016289453425712087,
1177
+ "loss": 1.3052,
1178
+ "step": 975
1179
+ },
1180
+ {
1181
+ "epoch": 0.38,
1182
+ "learning_rate": 0.00016270207852193998,
1183
+ "loss": 1.3312,
1184
+ "step": 980
1185
+ },
1186
+ {
1187
+ "epoch": 0.38,
1188
+ "learning_rate": 0.00016250962278675907,
1189
+ "loss": 1.3215,
1190
+ "step": 985
1191
+ },
1192
+ {
1193
+ "epoch": 0.38,
1194
+ "learning_rate": 0.00016231716705157812,
1195
+ "loss": 1.3382,
1196
+ "step": 990
1197
+ },
1198
+ {
1199
+ "epoch": 0.38,
1200
+ "learning_rate": 0.00016212471131639723,
1201
+ "loss": 1.3263,
1202
+ "step": 995
1203
+ },
1204
+ {
1205
+ "epoch": 0.38,
1206
+ "learning_rate": 0.00016193225558121631,
1207
+ "loss": 1.2434,
1208
+ "step": 1000
1209
+ },
1210
+ {
1211
+ "epoch": 0.39,
1212
+ "learning_rate": 0.00016173979984603542,
1213
+ "loss": 1.3063,
1214
+ "step": 1005
1215
+ },
1216
+ {
1217
+ "epoch": 0.39,
1218
+ "learning_rate": 0.0001615473441108545,
1219
+ "loss": 1.3254,
1220
+ "step": 1010
1221
+ },
1222
+ {
1223
+ "epoch": 0.39,
1224
+ "learning_rate": 0.00016135488837567362,
1225
+ "loss": 1.2654,
1226
+ "step": 1015
1227
+ },
1228
+ {
1229
+ "epoch": 0.39,
1230
+ "learning_rate": 0.0001611624326404927,
1231
+ "loss": 1.348,
1232
+ "step": 1020
1233
+ },
1234
+ {
1235
+ "epoch": 0.39,
1236
+ "learning_rate": 0.00016096997690531178,
1237
+ "loss": 1.3155,
1238
+ "step": 1025
1239
+ },
1240
+ {
1241
+ "epoch": 0.4,
1242
+ "learning_rate": 0.00016077752117013086,
1243
+ "loss": 1.2864,
1244
+ "step": 1030
1245
+ },
1246
+ {
1247
+ "epoch": 0.4,
1248
+ "learning_rate": 0.00016058506543494997,
1249
+ "loss": 1.364,
1250
+ "step": 1035
1251
+ },
1252
+ {
1253
+ "epoch": 0.4,
1254
+ "learning_rate": 0.00016039260969976906,
1255
+ "loss": 1.2698,
1256
+ "step": 1040
1257
+ },
1258
+ {
1259
+ "epoch": 0.4,
1260
+ "learning_rate": 0.00016020015396458817,
1261
+ "loss": 1.3632,
1262
+ "step": 1045
1263
+ },
1264
+ {
1265
+ "epoch": 0.4,
1266
+ "learning_rate": 0.00016000769822940725,
1267
+ "loss": 1.2741,
1268
+ "step": 1050
1269
+ },
1270
+ {
1271
+ "epoch": 0.41,
1272
+ "learning_rate": 0.00015981524249422633,
1273
+ "loss": 1.3533,
1274
+ "step": 1055
1275
+ },
1276
+ {
1277
+ "epoch": 0.41,
1278
+ "learning_rate": 0.00015962278675904542,
1279
+ "loss": 1.3571,
1280
+ "step": 1060
1281
+ },
1282
+ {
1283
+ "epoch": 0.41,
1284
+ "learning_rate": 0.0001594303310238645,
1285
+ "loss": 1.2827,
1286
+ "step": 1065
1287
+ },
1288
+ {
1289
+ "epoch": 0.41,
1290
+ "learning_rate": 0.0001592378752886836,
1291
+ "loss": 1.2547,
1292
+ "step": 1070
1293
+ },
1294
+ {
1295
+ "epoch": 0.41,
1296
+ "learning_rate": 0.0001590454195535027,
1297
+ "loss": 1.4101,
1298
+ "step": 1075
1299
+ },
1300
+ {
1301
+ "epoch": 0.41,
1302
+ "learning_rate": 0.0001588529638183218,
1303
+ "loss": 1.3149,
1304
+ "step": 1080
1305
+ },
1306
+ {
1307
+ "epoch": 0.42,
1308
+ "learning_rate": 0.00015866050808314088,
1309
+ "loss": 1.3008,
1310
+ "step": 1085
1311
+ },
1312
+ {
1313
+ "epoch": 0.42,
1314
+ "learning_rate": 0.00015846805234796,
1315
+ "loss": 1.2859,
1316
+ "step": 1090
1317
+ },
1318
+ {
1319
+ "epoch": 0.42,
1320
+ "learning_rate": 0.00015827559661277908,
1321
+ "loss": 1.1892,
1322
+ "step": 1095
1323
+ },
1324
+ {
1325
+ "epoch": 0.42,
1326
+ "learning_rate": 0.00015808314087759816,
1327
+ "loss": 1.364,
1328
+ "step": 1100
1329
+ },
1330
+ {
1331
+ "epoch": 0.42,
1332
+ "learning_rate": 0.00015789068514241724,
1333
+ "loss": 1.2708,
1334
+ "step": 1105
1335
+ },
1336
+ {
1337
+ "epoch": 0.43,
1338
+ "learning_rate": 0.00015769822940723635,
1339
+ "loss": 1.3591,
1340
+ "step": 1110
1341
+ },
1342
+ {
1343
+ "epoch": 0.43,
1344
+ "learning_rate": 0.00015750577367205543,
1345
+ "loss": 1.2828,
1346
+ "step": 1115
1347
+ },
1348
+ {
1349
+ "epoch": 0.43,
1350
+ "learning_rate": 0.00015731331793687452,
1351
+ "loss": 1.3861,
1352
+ "step": 1120
1353
+ },
1354
+ {
1355
+ "epoch": 0.43,
1356
+ "learning_rate": 0.00015712086220169363,
1357
+ "loss": 1.3752,
1358
+ "step": 1125
1359
+ },
1360
+ {
1361
+ "epoch": 0.43,
1362
+ "learning_rate": 0.0001569284064665127,
1363
+ "loss": 1.299,
1364
+ "step": 1130
1365
+ },
1366
+ {
1367
+ "epoch": 0.44,
1368
+ "learning_rate": 0.0001567359507313318,
1369
+ "loss": 1.2744,
1370
+ "step": 1135
1371
+ },
1372
+ {
1373
+ "epoch": 0.44,
1374
+ "learning_rate": 0.00015654349499615087,
1375
+ "loss": 1.3226,
1376
+ "step": 1140
1377
+ },
1378
+ {
1379
+ "epoch": 0.44,
1380
+ "learning_rate": 0.00015635103926096998,
1381
+ "loss": 1.2342,
1382
+ "step": 1145
1383
+ },
1384
+ {
1385
+ "epoch": 0.44,
1386
+ "learning_rate": 0.00015615858352578907,
1387
+ "loss": 1.2236,
1388
+ "step": 1150
1389
+ },
1390
+ {
1391
+ "epoch": 0.44,
1392
+ "learning_rate": 0.00015596612779060818,
1393
+ "loss": 1.3178,
1394
+ "step": 1155
1395
+ },
1396
+ {
1397
+ "epoch": 0.45,
1398
+ "learning_rate": 0.00015577367205542726,
1399
+ "loss": 1.315,
1400
+ "step": 1160
1401
+ },
1402
+ {
1403
+ "epoch": 0.45,
1404
+ "learning_rate": 0.00015558121632024637,
1405
+ "loss": 1.2071,
1406
+ "step": 1165
1407
+ },
1408
+ {
1409
+ "epoch": 0.45,
1410
+ "learning_rate": 0.00015538876058506545,
1411
+ "loss": 1.2485,
1412
+ "step": 1170
1413
+ },
1414
+ {
1415
+ "epoch": 0.45,
1416
+ "learning_rate": 0.00015519630484988454,
1417
+ "loss": 1.3034,
1418
+ "step": 1175
1419
+ },
1420
+ {
1421
+ "epoch": 0.45,
1422
+ "learning_rate": 0.00015500384911470362,
1423
+ "loss": 1.3113,
1424
+ "step": 1180
1425
+ },
1426
+ {
1427
+ "epoch": 0.46,
1428
+ "learning_rate": 0.0001548113933795227,
1429
+ "loss": 1.4071,
1430
+ "step": 1185
1431
+ },
1432
+ {
1433
+ "epoch": 0.46,
1434
+ "learning_rate": 0.0001546189376443418,
1435
+ "loss": 1.2806,
1436
+ "step": 1190
1437
+ },
1438
+ {
1439
+ "epoch": 0.46,
1440
+ "learning_rate": 0.0001544264819091609,
1441
+ "loss": 1.2668,
1442
+ "step": 1195
1443
+ },
1444
+ {
1445
+ "epoch": 0.46,
1446
+ "learning_rate": 0.00015423402617398,
1447
+ "loss": 1.2156,
1448
+ "step": 1200
1449
+ },
1450
+ {
1451
+ "epoch": 0.46,
1452
+ "learning_rate": 0.00015404157043879909,
1453
+ "loss": 1.2713,
1454
+ "step": 1205
1455
+ },
1456
+ {
1457
+ "epoch": 0.46,
1458
+ "learning_rate": 0.00015384911470361817,
1459
+ "loss": 1.3547,
1460
+ "step": 1210
1461
+ },
1462
+ {
1463
+ "epoch": 0.47,
1464
+ "learning_rate": 0.00015365665896843725,
1465
+ "loss": 1.352,
1466
+ "step": 1215
1467
+ },
1468
+ {
1469
+ "epoch": 0.47,
1470
+ "learning_rate": 0.00015346420323325636,
1471
+ "loss": 1.2166,
1472
+ "step": 1220
1473
+ },
1474
+ {
1475
+ "epoch": 0.47,
1476
+ "learning_rate": 0.00015327174749807544,
1477
+ "loss": 1.3411,
1478
+ "step": 1225
1479
+ },
1480
+ {
1481
+ "epoch": 0.47,
1482
+ "learning_rate": 0.00015307929176289455,
1483
+ "loss": 1.3127,
1484
+ "step": 1230
1485
+ },
1486
+ {
1487
+ "epoch": 0.47,
1488
+ "learning_rate": 0.00015288683602771364,
1489
+ "loss": 1.2179,
1490
+ "step": 1235
1491
+ },
1492
+ {
1493
+ "epoch": 0.48,
1494
+ "learning_rate": 0.00015269438029253275,
1495
+ "loss": 1.3099,
1496
+ "step": 1240
1497
+ },
1498
+ {
1499
+ "epoch": 0.48,
1500
+ "learning_rate": 0.00015250192455735183,
1501
+ "loss": 1.3071,
1502
+ "step": 1245
1503
+ },
1504
+ {
1505
+ "epoch": 0.48,
1506
+ "learning_rate": 0.00015230946882217089,
1507
+ "loss": 1.2789,
1508
+ "step": 1250
1509
+ },
1510
+ {
1511
+ "epoch": 0.48,
1512
+ "learning_rate": 0.00015211701308699,
1513
+ "loss": 1.265,
1514
+ "step": 1255
1515
+ },
1516
+ {
1517
+ "epoch": 0.48,
1518
+ "learning_rate": 0.00015192455735180908,
1519
+ "loss": 1.3423,
1520
+ "step": 1260
1521
+ },
1522
+ {
1523
+ "epoch": 0.49,
1524
+ "learning_rate": 0.0001517321016166282,
1525
+ "loss": 1.3561,
1526
+ "step": 1265
1527
+ },
1528
+ {
1529
+ "epoch": 0.49,
1530
+ "learning_rate": 0.00015153964588144727,
1531
+ "loss": 1.3766,
1532
+ "step": 1270
1533
+ },
1534
+ {
1535
+ "epoch": 0.49,
1536
+ "learning_rate": 0.00015134719014626638,
1537
+ "loss": 1.1815,
1538
+ "step": 1275
1539
+ },
1540
+ {
1541
+ "epoch": 0.49,
1542
+ "learning_rate": 0.00015115473441108546,
1543
+ "loss": 1.2731,
1544
+ "step": 1280
1545
+ },
1546
+ {
1547
+ "epoch": 0.49,
1548
+ "learning_rate": 0.00015096227867590455,
1549
+ "loss": 1.4211,
1550
+ "step": 1285
1551
+ },
1552
+ {
1553
+ "epoch": 0.5,
1554
+ "learning_rate": 0.00015076982294072363,
1555
+ "loss": 1.3037,
1556
+ "step": 1290
1557
+ },
1558
+ {
1559
+ "epoch": 0.5,
1560
+ "learning_rate": 0.00015057736720554274,
1561
+ "loss": 1.2689,
1562
+ "step": 1295
1563
+ },
1564
+ {
1565
+ "epoch": 0.5,
1566
+ "learning_rate": 0.00015038491147036182,
1567
+ "loss": 1.3821,
1568
+ "step": 1300
1569
+ },
1570
+ {
1571
+ "epoch": 0.5,
1572
+ "learning_rate": 0.00015019245573518093,
1573
+ "loss": 1.3421,
1574
+ "step": 1305
1575
+ },
1576
+ {
1577
+ "epoch": 0.5,
1578
+ "learning_rate": 0.00015000000000000001,
1579
+ "loss": 1.2462,
1580
+ "step": 1310
1581
+ },
1582
+ {
1583
+ "epoch": 0.51,
1584
+ "learning_rate": 0.0001498075442648191,
1585
+ "loss": 1.3296,
1586
+ "step": 1315
1587
+ },
1588
+ {
1589
+ "epoch": 0.51,
1590
+ "learning_rate": 0.00014961508852963818,
1591
+ "loss": 1.2991,
1592
+ "step": 1320
1593
+ },
1594
+ {
1595
+ "epoch": 0.51,
1596
+ "learning_rate": 0.00014942263279445726,
1597
+ "loss": 1.2695,
1598
+ "step": 1325
1599
+ },
1600
+ {
1601
+ "epoch": 0.51,
1602
+ "learning_rate": 0.00014923017705927637,
1603
+ "loss": 1.3681,
1604
+ "step": 1330
1605
+ },
1606
+ {
1607
+ "epoch": 0.51,
1608
+ "learning_rate": 0.00014903772132409545,
1609
+ "loss": 1.3166,
1610
+ "step": 1335
1611
+ },
1612
+ {
1613
+ "epoch": 0.51,
1614
+ "learning_rate": 0.00014884526558891456,
1615
+ "loss": 1.2906,
1616
+ "step": 1340
1617
+ },
1618
+ {
1619
+ "epoch": 0.52,
1620
+ "learning_rate": 0.00014865280985373365,
1621
+ "loss": 1.2122,
1622
+ "step": 1345
1623
+ },
1624
+ {
1625
+ "epoch": 0.52,
1626
+ "learning_rate": 0.00014846035411855276,
1627
+ "loss": 1.2767,
1628
+ "step": 1350
1629
+ },
1630
+ {
1631
+ "epoch": 0.52,
1632
+ "learning_rate": 0.00014826789838337184,
1633
+ "loss": 1.3143,
1634
+ "step": 1355
1635
+ },
1636
+ {
1637
+ "epoch": 0.52,
1638
+ "learning_rate": 0.00014807544264819092,
1639
+ "loss": 1.2427,
1640
+ "step": 1360
1641
+ },
1642
+ {
1643
+ "epoch": 0.52,
1644
+ "learning_rate": 0.00014788298691301,
1645
+ "loss": 1.2413,
1646
+ "step": 1365
1647
+ },
1648
+ {
1649
+ "epoch": 0.53,
1650
+ "learning_rate": 0.00014769053117782912,
1651
+ "loss": 1.2943,
1652
+ "step": 1370
1653
+ },
1654
+ {
1655
+ "epoch": 0.53,
1656
+ "learning_rate": 0.0001474980754426482,
1657
+ "loss": 1.2527,
1658
+ "step": 1375
1659
+ },
1660
+ {
1661
+ "epoch": 0.53,
1662
+ "learning_rate": 0.00014730561970746728,
1663
+ "loss": 1.3178,
1664
+ "step": 1380
1665
+ },
1666
+ {
1667
+ "epoch": 0.53,
1668
+ "learning_rate": 0.0001471131639722864,
1669
+ "loss": 1.2924,
1670
+ "step": 1385
1671
+ },
1672
+ {
1673
+ "epoch": 0.53,
1674
+ "learning_rate": 0.00014692070823710547,
1675
+ "loss": 1.1681,
1676
+ "step": 1390
1677
+ },
1678
+ {
1679
+ "epoch": 0.54,
1680
+ "learning_rate": 0.00014672825250192456,
1681
+ "loss": 1.2805,
1682
+ "step": 1395
1683
+ },
1684
+ {
1685
+ "epoch": 0.54,
1686
+ "learning_rate": 0.00014653579676674364,
1687
+ "loss": 1.2903,
1688
+ "step": 1400
1689
+ },
1690
+ {
1691
+ "epoch": 0.54,
1692
+ "learning_rate": 0.00014634334103156275,
1693
+ "loss": 1.3223,
1694
+ "step": 1405
1695
+ },
1696
+ {
1697
+ "epoch": 0.54,
1698
+ "learning_rate": 0.00014615088529638183,
1699
+ "loss": 1.208,
1700
+ "step": 1410
1701
+ },
1702
+ {
1703
+ "epoch": 0.54,
1704
+ "learning_rate": 0.00014595842956120094,
1705
+ "loss": 1.2135,
1706
+ "step": 1415
1707
+ },
1708
+ {
1709
+ "epoch": 0.55,
1710
+ "learning_rate": 0.00014576597382602002,
1711
+ "loss": 1.291,
1712
+ "step": 1420
1713
+ },
1714
+ {
1715
+ "epoch": 0.55,
1716
+ "learning_rate": 0.00014557351809083913,
1717
+ "loss": 1.1809,
1718
+ "step": 1425
1719
+ },
1720
+ {
1721
+ "epoch": 0.55,
1722
+ "learning_rate": 0.00014538106235565822,
1723
+ "loss": 1.3176,
1724
+ "step": 1430
1725
+ },
1726
+ {
1727
+ "epoch": 0.55,
1728
+ "learning_rate": 0.0001451886066204773,
1729
+ "loss": 1.2837,
1730
+ "step": 1435
1731
+ },
1732
+ {
1733
+ "epoch": 0.55,
1734
+ "learning_rate": 0.00014499615088529638,
1735
+ "loss": 1.2517,
1736
+ "step": 1440
1737
+ },
1738
+ {
1739
+ "epoch": 0.56,
1740
+ "learning_rate": 0.00014480369515011547,
1741
+ "loss": 1.3029,
1742
+ "step": 1445
1743
+ },
1744
+ {
1745
+ "epoch": 0.56,
1746
+ "learning_rate": 0.00014461123941493458,
1747
+ "loss": 1.2432,
1748
+ "step": 1450
1749
+ },
1750
+ {
1751
+ "epoch": 0.56,
1752
+ "learning_rate": 0.00014441878367975366,
1753
+ "loss": 1.2917,
1754
+ "step": 1455
1755
+ },
1756
+ {
1757
+ "epoch": 0.56,
1758
+ "learning_rate": 0.00014422632794457277,
1759
+ "loss": 1.3542,
1760
+ "step": 1460
1761
+ },
1762
+ {
1763
+ "epoch": 0.56,
1764
+ "learning_rate": 0.00014403387220939185,
1765
+ "loss": 1.2598,
1766
+ "step": 1465
1767
+ },
1768
+ {
1769
+ "epoch": 0.56,
1770
+ "learning_rate": 0.00014384141647421093,
1771
+ "loss": 1.2475,
1772
+ "step": 1470
1773
+ },
1774
+ {
1775
+ "epoch": 0.57,
1776
+ "learning_rate": 0.00014364896073903002,
1777
+ "loss": 1.2112,
1778
+ "step": 1475
1779
+ },
1780
+ {
1781
+ "epoch": 0.57,
1782
+ "learning_rate": 0.00014345650500384913,
1783
+ "loss": 1.2839,
1784
+ "step": 1480
1785
+ },
1786
+ {
1787
+ "epoch": 0.57,
1788
+ "learning_rate": 0.0001432640492686682,
1789
+ "loss": 1.1871,
1790
+ "step": 1485
1791
+ },
1792
+ {
1793
+ "epoch": 0.57,
1794
+ "learning_rate": 0.00014307159353348732,
1795
+ "loss": 1.1936,
1796
+ "step": 1490
1797
+ },
1798
+ {
1799
+ "epoch": 0.57,
1800
+ "learning_rate": 0.0001428791377983064,
1801
+ "loss": 1.2619,
1802
+ "step": 1495
1803
+ },
1804
+ {
1805
+ "epoch": 0.58,
1806
+ "learning_rate": 0.0001426866820631255,
1807
+ "loss": 1.243,
1808
+ "step": 1500
1809
+ },
1810
+ {
1811
+ "epoch": 0.58,
1812
+ "learning_rate": 0.00014249422632794457,
1813
+ "loss": 1.3483,
1814
+ "step": 1505
1815
+ },
1816
+ {
1817
+ "epoch": 0.58,
1818
+ "learning_rate": 0.00014230177059276365,
1819
+ "loss": 1.241,
1820
+ "step": 1510
1821
+ },
1822
+ {
1823
+ "epoch": 0.58,
1824
+ "learning_rate": 0.00014210931485758276,
1825
+ "loss": 1.2585,
1826
+ "step": 1515
1827
+ },
1828
+ {
1829
+ "epoch": 0.58,
1830
+ "learning_rate": 0.00014191685912240184,
1831
+ "loss": 1.2525,
1832
+ "step": 1520
1833
+ },
1834
+ {
1835
+ "epoch": 0.59,
1836
+ "learning_rate": 0.00014172440338722095,
1837
+ "loss": 1.2934,
1838
+ "step": 1525
1839
+ },
1840
+ {
1841
+ "epoch": 0.59,
1842
+ "learning_rate": 0.00014153194765204003,
1843
+ "loss": 1.2277,
1844
+ "step": 1530
1845
+ },
1846
+ {
1847
+ "epoch": 0.59,
1848
+ "learning_rate": 0.00014133949191685914,
1849
+ "loss": 1.2294,
1850
+ "step": 1535
1851
+ },
1852
+ {
1853
+ "epoch": 0.59,
1854
+ "learning_rate": 0.00014114703618167823,
1855
+ "loss": 1.245,
1856
+ "step": 1540
1857
+ },
1858
+ {
1859
+ "epoch": 0.59,
1860
+ "learning_rate": 0.0001409545804464973,
1861
+ "loss": 1.1826,
1862
+ "step": 1545
1863
+ },
1864
+ {
1865
+ "epoch": 0.6,
1866
+ "learning_rate": 0.0001407621247113164,
1867
+ "loss": 1.2436,
1868
+ "step": 1550
1869
+ },
1870
+ {
1871
+ "epoch": 0.6,
1872
+ "learning_rate": 0.0001405696689761355,
1873
+ "loss": 1.1588,
1874
+ "step": 1555
1875
+ },
1876
+ {
1877
+ "epoch": 0.6,
1878
+ "learning_rate": 0.00014037721324095459,
1879
+ "loss": 1.3642,
1880
+ "step": 1560
1881
+ },
1882
+ {
1883
+ "epoch": 0.6,
1884
+ "learning_rate": 0.0001401847575057737,
1885
+ "loss": 1.2621,
1886
+ "step": 1565
1887
+ },
1888
+ {
1889
+ "epoch": 0.6,
1890
+ "learning_rate": 0.00013999230177059278,
1891
+ "loss": 1.2909,
1892
+ "step": 1570
1893
+ },
1894
+ {
1895
+ "epoch": 0.61,
1896
+ "learning_rate": 0.00013979984603541186,
1897
+ "loss": 1.2259,
1898
+ "step": 1575
1899
+ },
1900
+ {
1901
+ "epoch": 0.61,
1902
+ "learning_rate": 0.00013960739030023094,
1903
+ "loss": 1.2078,
1904
+ "step": 1580
1905
+ },
1906
+ {
1907
+ "epoch": 0.61,
1908
+ "learning_rate": 0.00013941493456505003,
1909
+ "loss": 1.2307,
1910
+ "step": 1585
1911
+ },
1912
+ {
1913
+ "epoch": 0.61,
1914
+ "learning_rate": 0.00013922247882986914,
1915
+ "loss": 1.2816,
1916
+ "step": 1590
1917
+ },
1918
+ {
1919
+ "epoch": 0.61,
1920
+ "learning_rate": 0.00013903002309468822,
1921
+ "loss": 1.2587,
1922
+ "step": 1595
1923
+ },
1924
+ {
1925
+ "epoch": 0.61,
1926
+ "learning_rate": 0.00013883756735950733,
1927
+ "loss": 1.3018,
1928
+ "step": 1600
1929
+ },
1930
+ {
1931
+ "epoch": 0.62,
1932
+ "learning_rate": 0.0001386451116243264,
1933
+ "loss": 1.2167,
1934
+ "step": 1605
1935
+ },
1936
+ {
1937
+ "epoch": 0.62,
1938
+ "learning_rate": 0.00013845265588914552,
1939
+ "loss": 1.2399,
1940
+ "step": 1610
1941
+ },
1942
+ {
1943
+ "epoch": 0.62,
1944
+ "learning_rate": 0.0001382602001539646,
1945
+ "loss": 1.2355,
1946
+ "step": 1615
1947
+ },
1948
+ {
1949
+ "epoch": 0.62,
1950
+ "learning_rate": 0.0001380677444187837,
1951
+ "loss": 1.2099,
1952
+ "step": 1620
1953
+ },
1954
+ {
1955
+ "epoch": 0.62,
1956
+ "learning_rate": 0.00013787528868360277,
1957
+ "loss": 1.2794,
1958
+ "step": 1625
1959
+ },
1960
+ {
1961
+ "epoch": 0.63,
1962
+ "learning_rate": 0.00013768283294842188,
1963
+ "loss": 1.1232,
1964
+ "step": 1630
1965
+ },
1966
+ {
1967
+ "epoch": 0.63,
1968
+ "learning_rate": 0.00013749037721324096,
1969
+ "loss": 1.2563,
1970
+ "step": 1635
1971
+ },
1972
+ {
1973
+ "epoch": 0.63,
1974
+ "learning_rate": 0.00013729792147806005,
1975
+ "loss": 1.2364,
1976
+ "step": 1640
1977
+ },
1978
+ {
1979
+ "epoch": 0.63,
1980
+ "learning_rate": 0.00013710546574287916,
1981
+ "loss": 1.2036,
1982
+ "step": 1645
1983
+ },
1984
+ {
1985
+ "epoch": 0.63,
1986
+ "learning_rate": 0.00013691301000769824,
1987
+ "loss": 1.209,
1988
+ "step": 1650
1989
+ },
1990
+ {
1991
+ "epoch": 0.64,
1992
+ "learning_rate": 0.00013672055427251732,
1993
+ "loss": 1.2653,
1994
+ "step": 1655
1995
+ },
1996
+ {
1997
+ "epoch": 0.64,
1998
+ "learning_rate": 0.0001365280985373364,
1999
+ "loss": 1.3329,
2000
+ "step": 1660
2001
+ },
2002
+ {
2003
+ "epoch": 0.64,
2004
+ "learning_rate": 0.0001363356428021555,
2005
+ "loss": 1.2629,
2006
+ "step": 1665
2007
+ },
2008
+ {
2009
+ "epoch": 0.64,
2010
+ "learning_rate": 0.0001361431870669746,
2011
+ "loss": 1.2393,
2012
+ "step": 1670
2013
+ },
2014
+ {
2015
+ "epoch": 0.64,
2016
+ "learning_rate": 0.0001359507313317937,
2017
+ "loss": 1.1669,
2018
+ "step": 1675
2019
+ },
2020
+ {
2021
+ "epoch": 0.65,
2022
+ "learning_rate": 0.0001357582755966128,
2023
+ "loss": 1.2177,
2024
+ "step": 1680
2025
+ },
2026
+ {
2027
+ "epoch": 0.65,
2028
+ "learning_rate": 0.0001355658198614319,
2029
+ "loss": 1.1363,
2030
+ "step": 1685
2031
+ },
2032
+ {
2033
+ "epoch": 0.65,
2034
+ "learning_rate": 0.00013537336412625098,
2035
+ "loss": 1.1773,
2036
+ "step": 1690
2037
+ },
2038
+ {
2039
+ "epoch": 0.65,
2040
+ "learning_rate": 0.00013518090839107006,
2041
+ "loss": 1.2482,
2042
+ "step": 1695
2043
+ },
2044
+ {
2045
+ "epoch": 0.65,
2046
+ "learning_rate": 0.00013498845265588915,
2047
+ "loss": 1.266,
2048
+ "step": 1700
2049
+ },
2050
+ {
2051
+ "epoch": 0.66,
2052
+ "learning_rate": 0.00013479599692070823,
2053
+ "loss": 1.2936,
2054
+ "step": 1705
2055
+ },
2056
+ {
2057
+ "epoch": 0.66,
2058
+ "learning_rate": 0.00013460354118552734,
2059
+ "loss": 1.2855,
2060
+ "step": 1710
2061
+ },
2062
+ {
2063
+ "epoch": 0.66,
2064
+ "learning_rate": 0.00013441108545034642,
2065
+ "loss": 1.2413,
2066
+ "step": 1715
2067
+ },
2068
+ {
2069
+ "epoch": 0.66,
2070
+ "learning_rate": 0.00013421862971516553,
2071
+ "loss": 1.2874,
2072
+ "step": 1720
2073
+ },
2074
+ {
2075
+ "epoch": 0.66,
2076
+ "learning_rate": 0.00013402617397998461,
2077
+ "loss": 1.2527,
2078
+ "step": 1725
2079
+ },
2080
+ {
2081
+ "epoch": 0.66,
2082
+ "learning_rate": 0.0001338337182448037,
2083
+ "loss": 1.2527,
2084
+ "step": 1730
2085
+ },
2086
+ {
2087
+ "epoch": 0.67,
2088
+ "learning_rate": 0.00013364126250962278,
2089
+ "loss": 1.2423,
2090
+ "step": 1735
2091
+ },
2092
+ {
2093
+ "epoch": 0.67,
2094
+ "learning_rate": 0.0001334488067744419,
2095
+ "loss": 1.2544,
2096
+ "step": 1740
2097
+ },
2098
+ {
2099
+ "epoch": 0.67,
2100
+ "learning_rate": 0.00013325635103926097,
2101
+ "loss": 1.1613,
2102
+ "step": 1745
2103
+ },
2104
+ {
2105
+ "epoch": 0.67,
2106
+ "learning_rate": 0.00013306389530408008,
2107
+ "loss": 1.2273,
2108
+ "step": 1750
2109
+ },
2110
+ {
2111
+ "epoch": 0.67,
2112
+ "learning_rate": 0.00013287143956889917,
2113
+ "loss": 1.2789,
2114
+ "step": 1755
2115
+ },
2116
+ {
2117
+ "epoch": 0.68,
2118
+ "learning_rate": 0.00013267898383371828,
2119
+ "loss": 1.2755,
2120
+ "step": 1760
2121
+ },
2122
+ {
2123
+ "epoch": 0.68,
2124
+ "learning_rate": 0.00013248652809853733,
2125
+ "loss": 1.2144,
2126
+ "step": 1765
2127
+ },
2128
+ {
2129
+ "epoch": 0.68,
2130
+ "learning_rate": 0.00013229407236335641,
2131
+ "loss": 1.1615,
2132
+ "step": 1770
2133
+ },
2134
+ {
2135
+ "epoch": 0.68,
2136
+ "learning_rate": 0.00013210161662817552,
2137
+ "loss": 1.2173,
2138
+ "step": 1775
2139
+ },
2140
+ {
2141
+ "epoch": 0.68,
2142
+ "learning_rate": 0.0001319091608929946,
2143
+ "loss": 1.2408,
2144
+ "step": 1780
2145
+ },
2146
+ {
2147
+ "epoch": 0.69,
2148
+ "learning_rate": 0.00013171670515781372,
2149
+ "loss": 1.1671,
2150
+ "step": 1785
2151
+ },
2152
+ {
2153
+ "epoch": 0.69,
2154
+ "learning_rate": 0.0001315242494226328,
2155
+ "loss": 1.1994,
2156
+ "step": 1790
2157
+ },
2158
+ {
2159
+ "epoch": 0.69,
2160
+ "learning_rate": 0.0001313317936874519,
2161
+ "loss": 1.3369,
2162
+ "step": 1795
2163
+ },
2164
+ {
2165
+ "epoch": 0.69,
2166
+ "learning_rate": 0.000131139337952271,
2167
+ "loss": 1.1598,
2168
+ "step": 1800
2169
+ },
2170
+ {
2171
+ "epoch": 0.69,
2172
+ "learning_rate": 0.00013094688221709007,
2173
+ "loss": 1.1784,
2174
+ "step": 1805
2175
+ },
2176
+ {
2177
+ "epoch": 0.7,
2178
+ "learning_rate": 0.00013075442648190916,
2179
+ "loss": 1.2743,
2180
+ "step": 1810
2181
+ },
2182
+ {
2183
+ "epoch": 0.7,
2184
+ "learning_rate": 0.00013056197074672827,
2185
+ "loss": 1.2127,
2186
+ "step": 1815
2187
+ },
2188
+ {
2189
+ "epoch": 0.7,
2190
+ "learning_rate": 0.00013036951501154735,
2191
+ "loss": 1.2738,
2192
+ "step": 1820
2193
+ },
2194
+ {
2195
+ "epoch": 0.7,
2196
+ "learning_rate": 0.00013017705927636643,
2197
+ "loss": 1.2232,
2198
+ "step": 1825
2199
+ },
2200
+ {
2201
+ "epoch": 0.7,
2202
+ "learning_rate": 0.00012998460354118554,
2203
+ "loss": 1.1547,
2204
+ "step": 1830
2205
+ },
2206
+ {
2207
+ "epoch": 0.7,
2208
+ "learning_rate": 0.00012979214780600463,
2209
+ "loss": 1.2235,
2210
+ "step": 1835
2211
+ },
2212
+ {
2213
+ "epoch": 0.71,
2214
+ "learning_rate": 0.0001295996920708237,
2215
+ "loss": 1.332,
2216
+ "step": 1840
2217
+ },
2218
+ {
2219
+ "epoch": 0.71,
2220
+ "learning_rate": 0.0001294072363356428,
2221
+ "loss": 1.1242,
2222
+ "step": 1845
2223
+ },
2224
+ {
2225
+ "epoch": 0.71,
2226
+ "learning_rate": 0.0001292147806004619,
2227
+ "loss": 1.1874,
2228
+ "step": 1850
2229
+ },
2230
+ {
2231
+ "epoch": 0.71,
2232
+ "learning_rate": 0.00012902232486528098,
2233
+ "loss": 1.1394,
2234
+ "step": 1855
2235
+ },
2236
+ {
2237
+ "epoch": 0.71,
2238
+ "learning_rate": 0.0001288298691301001,
2239
+ "loss": 1.1734,
2240
+ "step": 1860
2241
+ },
2242
+ {
2243
+ "epoch": 0.72,
2244
+ "learning_rate": 0.00012863741339491918,
2245
+ "loss": 1.2245,
2246
+ "step": 1865
2247
+ },
2248
+ {
2249
+ "epoch": 0.72,
2250
+ "learning_rate": 0.00012844495765973829,
2251
+ "loss": 1.1869,
2252
+ "step": 1870
2253
+ },
2254
+ {
2255
+ "epoch": 0.72,
2256
+ "learning_rate": 0.00012825250192455737,
2257
+ "loss": 1.2558,
2258
+ "step": 1875
2259
  }
2260
  ],
2261
+ "max_steps": 5206,
2262
  "num_train_epochs": 2,
2263
+ "total_flos": 19105201520640.0,
2264
  "trial_name": null,
2265
  "trial_params": null
2266
  }
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f43edfee59f374cac51ff219426a9f3a4f9ade245c05ad2c22d14f5e8e9c1a4
3
  size 6139
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:762f87ade0fa6168b36ab9775461c0f8a3950e947fee8d4e9ec9e059b0407a8a
3
  size 6139
last-checkpoint/zero_to_fp32.py CHANGED
@@ -5,7 +5,7 @@
5
 
6
  # DeepSpeed Team
7
 
8
- # This script extracts fp32 consolidated weights from a zero 2 and 3 DeepSpeed checkpoints. It gets
9
  # copied into the top level checkpoint dir, so the user can easily do the conversion at any point in
10
  # the future. Once extracted, the weights don't require DeepSpeed and can be used in any
11
  # application.
@@ -63,7 +63,7 @@ def get_model_state_file(checkpoint_dir, zero_stage):
63
  raise FileNotFoundError(f"Directory '{checkpoint_dir}' doesn't exist")
64
 
65
  # there should be only one file
66
- if zero_stage == 2:
67
  file = os.path.join(checkpoint_dir, "mp_rank_00_model_states.pt")
68
  elif zero_stage == 3:
69
  file = os.path.join(checkpoint_dir, "zero_pp_rank_0_mp_rank_00_model_states.pt")
@@ -143,7 +143,11 @@ def parse_optim_states(files, ds_checkpoint_dir):
143
  total_files = len(files)
144
  state_dicts = []
145
  for f in files:
146
- state_dicts.append(torch.load(f, map_location=device))
 
 
 
 
147
 
148
  if not ZERO_STAGE in state_dicts[0][OPTIMIZER_STATE_DICT]:
149
  raise ValueError(f"{files[0]} is not a zero checkpoint")
@@ -164,14 +168,14 @@ def parse_optim_states(files, ds_checkpoint_dir):
164
  )
165
 
166
  # the groups are named differently in each stage
167
- if zero_stage == 2:
168
  fp32_groups_key = SINGLE_PARTITION_OF_FP32_GROUPS
169
  elif zero_stage == 3:
170
  fp32_groups_key = FP32_FLAT_GROUPS
171
  else:
172
  raise ValueError(f"unknown zero stage {zero_stage}")
173
 
174
- if zero_stage == 2:
175
  fp32_flat_groups = [state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key] for i in range(len(state_dicts))]
176
  elif zero_stage == 3:
177
  # if there is more than one param group, there will be multiple flattened tensors - one
@@ -206,7 +210,7 @@ def _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir):
206
  zero_model_states = parse_model_states(model_files)
207
  print(f'Parsing checkpoint created by deepspeed=={zero_model_states[0].ds_version}')
208
 
209
- if zero_stage == 2:
210
  return _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states)
211
  elif zero_stage == 3:
212
  return _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states)
 
5
 
6
  # DeepSpeed Team
7
 
8
+ # This script extracts fp32 consolidated weights from a zero 1, 2 and 3 DeepSpeed checkpoints. It gets
9
  # copied into the top level checkpoint dir, so the user can easily do the conversion at any point in
10
  # the future. Once extracted, the weights don't require DeepSpeed and can be used in any
11
  # application.
 
63
  raise FileNotFoundError(f"Directory '{checkpoint_dir}' doesn't exist")
64
 
65
  # there should be only one file
66
+ if zero_stage <= 2:
67
  file = os.path.join(checkpoint_dir, "mp_rank_00_model_states.pt")
68
  elif zero_stage == 3:
69
  file = os.path.join(checkpoint_dir, "zero_pp_rank_0_mp_rank_00_model_states.pt")
 
143
  total_files = len(files)
144
  state_dicts = []
145
  for f in files:
146
+ state_dict = torch.load(f, map_location=device)
147
+ # immediately discard the potentially huge 2 optimizer states as we only care for fp32 master weights
148
+ # and also handle the case where it was already removed by another helper script
149
+ state_dict["optimizer_state_dict"].pop("optimizer_state_dict", None)
150
+ state_dicts.append(state_dict)
151
 
152
  if not ZERO_STAGE in state_dicts[0][OPTIMIZER_STATE_DICT]:
153
  raise ValueError(f"{files[0]} is not a zero checkpoint")
 
168
  )
169
 
170
  # the groups are named differently in each stage
171
+ if zero_stage <= 2:
172
  fp32_groups_key = SINGLE_PARTITION_OF_FP32_GROUPS
173
  elif zero_stage == 3:
174
  fp32_groups_key = FP32_FLAT_GROUPS
175
  else:
176
  raise ValueError(f"unknown zero stage {zero_stage}")
177
 
178
+ if zero_stage <= 2:
179
  fp32_flat_groups = [state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key] for i in range(len(state_dicts))]
180
  elif zero_stage == 3:
181
  # if there is more than one param group, there will be multiple flattened tensors - one
 
210
  zero_model_states = parse_model_states(model_files)
211
  print(f'Parsing checkpoint created by deepspeed=={zero_model_states[0].ds_version}')
212
 
213
+ if zero_stage <= 2:
214
  return _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states)
215
  elif zero_stage == 3:
216
  return _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states)
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f43edfee59f374cac51ff219426a9f3a4f9ade245c05ad2c22d14f5e8e9c1a4
3
  size 6139
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:762f87ade0fa6168b36ab9775461c0f8a3950e947fee8d4e9ec9e059b0407a8a
3
  size 6139