KevinKibe commited on
Commit
ba86afb
1 Parent(s): 956c27f

Training in progress, step 10, checkpoint

Browse files
last-checkpoint/adapter_config.json CHANGED
@@ -23,8 +23,8 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "v_proj",
27
- "q_proj"
28
  ],
29
  "task_type": null,
30
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "q_proj",
27
+ "v_proj"
28
  ],
29
  "task_type": null,
30
  "use_dora": false,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f604dc87057072c6ea75e2256a7f52b189a3910b55463bc71f46b0594487ac55
3
  size 14176064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7b748f132f1b007b4b46c29b79576b9032a654e0c44680137c558710f38b8ea
3
  size 14176064
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e864ba831a16c7c9cb6a577948f8294a6b9956b437893e1b91868d8acdd201fb
3
  size 4832762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7da4a13f61129710c3e65e3794c64c276377cf47bba8ecc7b5ff65663120f31f
3
  size 4832762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bfb3811b63313489c6bb7fbbd93a86fcbb2760116919b75cb163af7587926db5
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fd34c05f32915cfddb244845f306181a1764de738fcce7d4ddb675210c64c1c
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:df73be84f0c00f13e358d00950761edb4aec7bfdee9e97f169f67cba14d6b4b5
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4dcd0406e202df2a3c65a5d416d9b75e9f9f16a0af78526a7608fc8a266c503
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,83 +1,35 @@
1
  {
2
- "best_metric": 81.60919540229885,
3
- "best_model_checkpoint": "../openai/whisper-small-finetuned/checkpoint-100",
4
- "epoch": 99.01,
5
- "eval_steps": 25,
6
- "global_step": 100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 24.01,
13
- "grad_norm": 0.658126711845398,
14
  "learning_rate": 0.0001,
15
- "loss": 1.5717,
16
- "step": 25
17
  },
18
  {
19
- "epoch": 24.01,
20
- "eval_loss": 3.452148675918579,
21
- "eval_runtime": 18.355,
22
- "eval_samples_per_second": 0.545,
23
- "eval_steps_per_second": 0.054,
24
- "eval_wer": 89.65517241379311,
25
- "step": 25
26
- },
27
- {
28
- "epoch": 49.01,
29
- "grad_norm": 0.2500029504299164,
30
- "learning_rate": 0.0001,
31
- "loss": 0.8875,
32
- "step": 50
33
- },
34
- {
35
- "epoch": 49.01,
36
- "eval_loss": 3.3323276042938232,
37
- "eval_runtime": 18.862,
38
- "eval_samples_per_second": 0.53,
39
- "eval_steps_per_second": 0.053,
40
- "eval_wer": 83.9080459770115,
41
- "step": 50
42
- },
43
- {
44
- "epoch": 74.01,
45
- "grad_norm": 25.93928337097168,
46
- "learning_rate": 0.0001,
47
- "loss": 0.6714,
48
- "step": 75
49
- },
50
- {
51
- "epoch": 74.01,
52
- "eval_loss": 3.3124895095825195,
53
- "eval_runtime": 19.825,
54
- "eval_samples_per_second": 0.504,
55
- "eval_steps_per_second": 0.05,
56
- "eval_wer": 82.75862068965517,
57
- "step": 75
58
- },
59
- {
60
- "epoch": 99.01,
61
- "grad_norm": 0.16914552450180054,
62
- "learning_rate": 0.0001,
63
- "loss": 0.5819,
64
- "step": 100
65
- },
66
- {
67
- "epoch": 99.01,
68
- "eval_loss": 3.2785234451293945,
69
- "eval_runtime": 20.2621,
70
- "eval_samples_per_second": 0.494,
71
- "eval_steps_per_second": 0.049,
72
- "eval_wer": 81.60919540229885,
73
- "step": 100
74
  }
75
  ],
76
- "logging_steps": 25,
77
- "max_steps": 100,
78
  "num_input_tokens_seen": 0,
79
  "num_train_epochs": 9223372036854775807,
80
- "save_steps": 25,
81
  "stateful_callbacks": {
82
  "TrainerControl": {
83
  "args": {
@@ -85,12 +37,12 @@
85
  "should_evaluate": false,
86
  "should_log": false,
87
  "should_save": true,
88
- "should_training_stop": true
89
  },
90
  "attributes": {}
91
  }
92
  },
93
- "total_flos": 5.8736295936e+17,
94
  "train_batch_size": 64,
95
  "trial_name": null,
96
  "trial_params": null
 
1
  {
2
+ "best_metric": 85.0574712643678,
3
+ "best_model_checkpoint": "../openai/whisper-small-finetuned/checkpoint-10",
4
+ "epoch": 9.05,
5
+ "eval_steps": 10,
6
+ "global_step": 10,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 9.05,
13
+ "grad_norm": 3.5340542793273926,
14
  "learning_rate": 0.0001,
15
+ "loss": 3.955,
16
+ "step": 10
17
  },
18
  {
19
+ "epoch": 9.05,
20
+ "eval_loss": 4.58884334564209,
21
+ "eval_runtime": 20.8147,
22
+ "eval_samples_per_second": 0.48,
23
+ "eval_steps_per_second": 0.048,
24
+ "eval_wer": 85.0574712643678,
25
+ "step": 10
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  }
27
  ],
28
+ "logging_steps": 10,
29
+ "max_steps": 20,
30
  "num_input_tokens_seen": 0,
31
  "num_train_epochs": 9223372036854775807,
32
+ "save_steps": 10,
33
  "stateful_callbacks": {
34
  "TrainerControl": {
35
  "args": {
 
37
  "should_evaluate": false,
38
  "should_log": false,
39
  "should_save": true,
40
+ "should_training_stop": false
41
  },
42
  "attributes": {}
43
  }
44
  },
45
+ "total_flos": 5.8736295936e+16,
46
  "train_batch_size": 64,
47
  "trial_name": null,
48
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d09de538d839ccb79e7d7b0f42cfd620c304edae9ecd58357e0e446b3b151f24
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85cb21ac25d27f454ad5e322f2dfdf92f2e016f1413497c1f02b2aa3d8ef540a
3
  size 5368