KevinKibe commited on
Commit
0908dce
1 Parent(s): 1e20729

Training in progress, step 10, checkpoint

Browse files
last-checkpoint/adapter_config.json CHANGED
@@ -23,8 +23,8 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "q_proj",
27
- "v_proj"
28
  ],
29
  "task_type": null,
30
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "v_proj",
27
+ "q_proj"
28
  ],
29
  "task_type": null,
30
  "use_dora": false,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48a2738892776e844e26a0ee6703b1fb99b3618598d0797ad41bb5969dc70b07
3
  size 14176064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eab84c97625eac5e0880e90af56f69c7ac1c1037c854b22bc5157c6bc30f7fb5
3
  size 14176064
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8b5427323a83099922056b799577815d11adbca7043c84179c99c26e81d30a52
3
  size 4832762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c58d4f7b9f3ff3ae48c6a6d543cb52233eb7cbd1b55850f8d16d9e523ade094
3
  size 4832762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7b33ed5fc6487acbf6e26dedc64c25e7f851e363e054fc82b98e44babc584391
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1c7e68895b7eac717d7e4bf5ab2889881ac66b5cbef829d15b86f30725fd17f
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a7409a3db23ca30d748433633adb25a16b5943f83075d54979b20be428dcaf7c
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1eca0ddb3ec890c90e8e6e7325bec40cdb8b59e4bebf5c0da8473b02809de875
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,98 +1,34 @@
1
  {
2
- "best_metric": 142.5,
3
- "best_model_checkpoint": "../openai/whisper-small-finetuned/checkpoint-50",
4
- "epoch": 12.5,
5
  "eval_steps": 10,
6
- "global_step": 50,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 2.5,
13
- "grad_norm": 10.211004257202148,
14
  "learning_rate": 9e-05,
15
- "loss": 5.6632,
16
  "step": 10
17
  },
18
  {
19
- "epoch": 2.5,
20
- "eval_loss": 4.246429443359375,
21
- "eval_runtime": 6.907,
22
- "eval_samples_per_second": 0.869,
23
- "eval_steps_per_second": 0.145,
24
  "eval_wer": 157.5,
25
  "step": 10
26
- },
27
- {
28
- "epoch": 5.0,
29
- "grad_norm": 2.3187050819396973,
30
- "learning_rate": 0.0001,
31
- "loss": 4.3134,
32
- "step": 20
33
- },
34
- {
35
- "epoch": 5.0,
36
- "eval_loss": 3.409444808959961,
37
- "eval_runtime": 6.0126,
38
- "eval_samples_per_second": 0.998,
39
- "eval_steps_per_second": 0.166,
40
- "eval_wer": 145.0,
41
- "step": 20
42
- },
43
- {
44
- "epoch": 7.5,
45
- "grad_norm": 1.6774072647094727,
46
- "learning_rate": 0.0001,
47
- "loss": 3.5339,
48
- "step": 30
49
- },
50
- {
51
- "epoch": 7.5,
52
- "eval_loss": 2.968590021133423,
53
- "eval_runtime": 6.2596,
54
- "eval_samples_per_second": 0.959,
55
- "eval_steps_per_second": 0.16,
56
- "eval_wer": 145.0,
57
- "step": 30
58
- },
59
- {
60
- "epoch": 10.0,
61
- "grad_norm": 2.3679263591766357,
62
- "learning_rate": 0.0001,
63
- "loss": 2.9729,
64
- "step": 40
65
- },
66
- {
67
- "epoch": 10.0,
68
- "eval_loss": 2.7336347103118896,
69
- "eval_runtime": 7.5505,
70
- "eval_samples_per_second": 0.795,
71
- "eval_steps_per_second": 0.132,
72
- "eval_wer": 160.0,
73
- "step": 40
74
- },
75
- {
76
- "epoch": 12.5,
77
- "grad_norm": 1.7330552339553833,
78
- "learning_rate": 0.0001,
79
- "loss": 2.6911,
80
- "step": 50
81
- },
82
- {
83
- "epoch": 12.5,
84
- "eval_loss": 2.5959129333496094,
85
- "eval_runtime": 7.3414,
86
- "eval_samples_per_second": 0.817,
87
- "eval_steps_per_second": 0.136,
88
- "eval_wer": 142.5,
89
- "step": 50
90
  }
91
  ],
92
  "logging_steps": 10,
93
- "max_steps": 50,
94
  "num_input_tokens_seen": 0,
95
- "num_train_epochs": 13,
96
  "save_steps": 10,
97
  "stateful_callbacks": {
98
  "TrainerControl": {
@@ -101,12 +37,12 @@
101
  "should_evaluate": false,
102
  "should_log": false,
103
  "should_save": true,
104
- "should_training_stop": true
105
  },
106
  "attributes": {}
107
  }
108
  },
109
- "total_flos": 2.0322758393856e+17,
110
  "train_batch_size": 16,
111
  "trial_name": null,
112
  "trial_params": null
 
1
  {
2
+ "best_metric": 157.5,
3
+ "best_model_checkpoint": "../openai/whisper-small-finetuned/checkpoint-10",
4
+ "epoch": 2.066666666666667,
5
  "eval_steps": 10,
6
+ "global_step": 10,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 2.066666666666667,
13
+ "grad_norm": 12.599740982055664,
14
  "learning_rate": 9e-05,
15
+ "loss": 5.9124,
16
  "step": 10
17
  },
18
  {
19
+ "epoch": 2.066666666666667,
20
+ "eval_loss": 4.316458225250244,
21
+ "eval_runtime": 6.6642,
22
+ "eval_samples_per_second": 0.9,
23
+ "eval_steps_per_second": 0.15,
24
  "eval_wer": 157.5,
25
  "step": 10
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  }
27
  ],
28
  "logging_steps": 10,
29
+ "max_steps": 30,
30
  "num_input_tokens_seen": 0,
31
+ "num_train_epochs": 9223372036854775807,
32
  "save_steps": 10,
33
  "stateful_callbacks": {
34
  "TrainerControl": {
 
37
  "should_evaluate": false,
38
  "should_log": false,
39
  "should_save": true,
40
+ "should_training_stop": false
41
  },
42
  "attributes": {}
43
  }
44
  },
45
+ "total_flos": 4.170277011456e+16,
46
  "train_batch_size": 16,
47
  "trial_name": null,
48
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4e3463515450077cbaac6ef5248a6cbcde1055cef5783be11b4bd1c11daf2248
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94ef2d56d7f2173783036ac4eb53c992be8cb36de1230bc10392218f27610460
3
  size 5368