KevinKibe commited on
Commit
79ff963
1 Parent(s): 01bc302

Training in progress, step 500, checkpoint

Browse files
last-checkpoint/README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
- base_model: openai/whisper-small
3
  library_name: peft
 
4
  ---
5
 
6
  # Model Card for Model ID
 
1
  ---
 
2
  library_name: peft
3
+ base_model: openai/whisper-small
4
  ---
5
 
6
  # Model Card for Model ID
last-checkpoint/adapter_config.json CHANGED
@@ -23,8 +23,8 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "v_proj",
27
- "q_proj"
28
  ],
29
  "task_type": null,
30
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "q_proj",
27
+ "v_proj"
28
  ],
29
  "task_type": null,
30
  "use_dora": false,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5bf48bc81f22ed4a3d63abaf440f475fd90680f94f87ba2e5965161c5aac2e45
3
  size 14176064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a81e82bda9dcd00b0e1c33cfe6f1fdaf3e89ae5aec7b7e43b6645ca7b2c6dc9
3
  size 14176064
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a927dd4e6fd6ed7a110258e3d8e0cdbe0ebce7f7847a0294bc7c468330b8b979
3
  size 4832890
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc954ae19af7f7676fdaf86e1bbd134ba266d9a7226e99b400d544d8eeba6358
3
  size 4832890
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:47586dfc0afc6cfec360e3d39632c4aee9a22a56ca08f36766bb611ed8e4e252
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb8290f7060f840c0e94500ccf61924054dd37481a06525cc06c6a3905e66f47
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:647072c6e45e0dfca80c8b6481afc718d32502c5b24237edb51c87996b591c22
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df6b5589c1af5c61bb3bbb4c1c654a8cc71368279b069213acf87ffc7febc71c
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,115 +1,35 @@
1
  {
2
- "best_metric": 92.05103042198233,
3
- "best_model_checkpoint": "../openai/whisper-small-finetuned/checkpoint-200",
4
- "epoch": 18.024,
5
- "eval_steps": 100,
6
- "global_step": 600,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 3.004,
13
- "grad_norm": 16.915180206298828,
14
- "learning_rate": 9.214285714285714e-05,
15
- "loss": 3.4928,
16
- "step": 100
17
- },
18
- {
19
- "epoch": 3.004,
20
- "eval_loss": 3.5212297439575195,
21
- "eval_runtime": 254.3007,
22
- "eval_samples_per_second": 0.983,
23
- "eval_steps_per_second": 0.031,
24
- "eval_wer": 96.4180569185476,
25
- "step": 100
26
- },
27
- {
28
- "epoch": 6.008,
29
- "grad_norm": 1.997205376625061,
30
- "learning_rate": 8.19387755102041e-05,
31
- "loss": 2.4572,
32
- "step": 200
33
- },
34
- {
35
- "epoch": 6.008,
36
- "eval_loss": 3.190507411956787,
37
- "eval_runtime": 192.1727,
38
- "eval_samples_per_second": 1.301,
39
- "eval_steps_per_second": 0.042,
40
- "eval_wer": 92.05103042198233,
41
- "step": 200
42
- },
43
- {
44
- "epoch": 9.012,
45
- "grad_norm": 1.6556462049484253,
46
- "learning_rate": 7.173469387755103e-05,
47
- "loss": 1.7435,
48
- "step": 300
49
- },
50
- {
51
- "epoch": 9.012,
52
- "eval_loss": 2.6634905338287354,
53
- "eval_runtime": 241.7692,
54
- "eval_samples_per_second": 1.034,
55
- "eval_steps_per_second": 0.033,
56
- "eval_wer": 99.06771344455349,
57
- "step": 300
58
- },
59
- {
60
- "epoch": 12.016,
61
- "grad_norm": 2.3752479553222656,
62
- "learning_rate": 6.153061224489797e-05,
63
- "loss": 1.4418,
64
- "step": 400
65
- },
66
- {
67
- "epoch": 12.016,
68
- "eval_loss": 2.67608642578125,
69
- "eval_runtime": 188.8586,
70
- "eval_samples_per_second": 1.324,
71
- "eval_steps_per_second": 0.042,
72
- "eval_wer": 101.66830225711483,
73
- "step": 400
74
- },
75
- {
76
- "epoch": 15.02,
77
- "grad_norm": 3.0957250595092773,
78
- "learning_rate": 5.132653061224491e-05,
79
- "loss": 1.2986,
80
  "step": 500
81
  },
82
  {
83
- "epoch": 15.02,
84
- "eval_loss": 2.6544599533081055,
85
- "eval_runtime": 243.0171,
86
- "eval_samples_per_second": 1.029,
87
- "eval_steps_per_second": 0.033,
88
- "eval_wer": 105.59371933267909,
89
  "step": 500
90
- },
91
- {
92
- "epoch": 18.024,
93
- "grad_norm": 2.199568510055542,
94
- "learning_rate": 4.112244897959184e-05,
95
- "loss": 1.1683,
96
- "step": 600
97
- },
98
- {
99
- "epoch": 18.024,
100
- "eval_loss": 2.682807683944702,
101
- "eval_runtime": 190.2924,
102
- "eval_samples_per_second": 1.314,
103
- "eval_steps_per_second": 0.042,
104
- "eval_wer": 92.98331697742886,
105
- "step": 600
106
  }
107
  ],
108
- "logging_steps": 100,
109
- "max_steps": 1000,
110
  "num_input_tokens_seen": 0,
111
  "num_train_epochs": 9223372036854775807,
112
- "save_steps": 100,
113
  "stateful_callbacks": {
114
  "TrainerControl": {
115
  "args": {
@@ -122,7 +42,7 @@
122
  "attributes": {}
123
  }
124
  },
125
- "total_flos": 5.51181401063424e+18,
126
  "train_batch_size": 32,
127
  "trial_name": null,
128
  "trial_params": null
 
1
  {
2
+ "best_metric": 94.3081452404318,
3
+ "best_model_checkpoint": "../openai/whisper-small-finetuned/checkpoint-500",
4
+ "epoch": 0.25,
5
+ "eval_steps": 500,
6
+ "global_step": 500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.25,
13
+ "grad_norm": 1.1737864017486572,
14
+ "learning_rate": 0.000759090909090909,
15
+ "loss": 1.6334,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  "step": 500
17
  },
18
  {
19
+ "epoch": 0.25,
20
+ "eval_loss": 2.2990105152130127,
21
+ "eval_runtime": 602.9832,
22
+ "eval_samples_per_second": 0.415,
23
+ "eval_steps_per_second": 0.053,
24
+ "eval_wer": 94.3081452404318,
25
  "step": 500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  }
27
  ],
28
+ "logging_steps": 500,
29
+ "max_steps": 2000,
30
  "num_input_tokens_seen": 0,
31
  "num_train_epochs": 9223372036854775807,
32
+ "save_steps": 500,
33
  "stateful_callbacks": {
34
  "TrainerControl": {
35
  "args": {
 
42
  "attributes": {}
43
  }
44
  },
45
+ "total_flos": 4.69890367488e+18,
46
  "train_batch_size": 32,
47
  "trial_name": null,
48
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3acbf58c4f322256ef2a28a4429e1ca55f9f5c1504873d96c73ffd64cb4852d0
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3d8d8c7c3bad7ca67b09ea445b40e35b02fcb6febd1b6e09fc11cc8930ec4da
3
  size 5368