yashcode00 commited on
Commit
73c1c23
·
1 Parent(s): e78e760

yashcode00/wav2vec2-large-xlsr-indian-language-classification-featureExtractor

Browse files
README.md CHANGED
@@ -17,8 +17,8 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [yashcode00/wav2vec2-large-xlsr-indian-language-classification-featureExtractor](https://huggingface.co/yashcode00/wav2vec2-large-xlsr-indian-language-classification-featureExtractor) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.4481
21
- - Accuracy: 0.8710
22
 
23
  ## Model description
24
 
@@ -45,10 +45,13 @@ The following hyperparameters were used during training:
45
  - total_train_batch_size: 64
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
48
- - num_epochs: 60
49
 
50
  ### Training results
51
 
 
 
 
52
 
53
 
54
  ### Framework versions
 
17
 
18
  This model is a fine-tuned version of [yashcode00/wav2vec2-large-xlsr-indian-language-classification-featureExtractor](https://huggingface.co/yashcode00/wav2vec2-large-xlsr-indian-language-classification-featureExtractor) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.1164
21
+ - Accuracy: 0.9677
22
 
23
  ## Model description
24
 
 
45
  - total_train_batch_size: 64
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
48
+ - num_epochs: 300
49
 
50
  ### Training results
51
 
52
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
53
+ |:-------------:|:------:|:----:|:---------------:|:--------:|
54
+ | 0.0194 | 203.39 | 1500 | 0.1154 | 0.9677 |
55
 
56
 
57
  ### Framework versions
all_results.json CHANGED
@@ -1,15 +1,15 @@
1
  {
2
- "epoch": 56.95,
3
- "eval_accuracy": 0.8709677457809448,
4
- "eval_loss": 0.4480999708175659,
5
- "eval_runtime": 1.4774,
6
  "eval_samples": 93,
7
- "eval_samples_per_second": 62.95,
8
- "eval_steps_per_second": 8.123,
9
- "total_flos": 8.064772262536032e+17,
10
- "train_loss": 0.15708597316628412,
11
- "train_runtime": 790.3622,
12
  "train_samples": 466,
13
- "train_samples_per_second": 35.376,
14
- "train_steps_per_second": 0.531
15
  }
 
1
  {
2
+ "epoch": 284.75,
3
+ "eval_accuracy": 0.9677419066429138,
4
+ "eval_loss": 0.11643270403146744,
5
+ "eval_runtime": 1.6786,
6
  "eval_samples": 93,
7
+ "eval_samples_per_second": 55.402,
8
+ "eval_steps_per_second": 7.149,
9
+ "total_flos": 4.0318045822521795e+18,
10
+ "train_loss": 0.022167698939641316,
11
+ "train_runtime": 3842.9627,
12
  "train_samples": 466,
13
+ "train_samples_per_second": 36.378,
14
+ "train_steps_per_second": 0.546
15
  }
eval_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 56.95,
3
- "eval_accuracy": 0.8709677457809448,
4
- "eval_loss": 0.4480999708175659,
5
- "eval_runtime": 1.4774,
6
  "eval_samples": 93,
7
- "eval_samples_per_second": 62.95,
8
- "eval_steps_per_second": 8.123
9
  }
 
1
  {
2
+ "epoch": 284.75,
3
+ "eval_accuracy": 0.9677419066429138,
4
+ "eval_loss": 0.11643270403146744,
5
+ "eval_runtime": 1.6786,
6
  "eval_samples": 93,
7
+ "eval_samples_per_second": 55.402,
8
+ "eval_steps_per_second": 7.149
9
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c5bd079f25010ad6fed494f319bd825fbda1ad3ba0247b15ee3faa0fc2a04cef
3
  size 1266146037
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40adbc61c446a3a2f2e26e29a2072bbdd48d0ae0bf10ed6afa5870eedb1549ec
3
  size 1266146037
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 56.95,
3
- "total_flos": 8.064772262536032e+17,
4
- "train_loss": 0.15708597316628412,
5
- "train_runtime": 790.3622,
6
  "train_samples": 466,
7
- "train_samples_per_second": 35.376,
8
- "train_steps_per_second": 0.531
9
  }
 
1
  {
2
+ "epoch": 284.75,
3
+ "total_flos": 4.0318045822521795e+18,
4
+ "train_loss": 0.022167698939641316,
5
+ "train_runtime": 3842.9627,
6
  "train_samples": 466,
7
+ "train_samples_per_second": 36.378,
8
+ "train_steps_per_second": 0.546
9
  }
trainer_state.json CHANGED
@@ -1,61 +1,172 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 56.94915254237288,
5
- "eval_steps": 1000,
6
- "global_step": 420,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 13.56,
13
- "learning_rate": 3.821428571428572e-05,
14
- "loss": 0.554,
15
  "step": 100
16
  },
17
  {
18
  "epoch": 27.12,
19
- "learning_rate": 2.6309523809523813e-05,
20
- "loss": 0.0396,
21
  "step": 200
22
  },
23
  {
24
  "epoch": 40.68,
25
- "learning_rate": 1.4404761904761905e-05,
26
- "loss": 0.0312,
27
  "step": 300
28
  },
29
  {
30
  "epoch": 54.24,
31
- "learning_rate": 2.5e-06,
32
- "loss": 0.0308,
33
  "step": 400
34
  },
35
  {
36
- "epoch": 56.95,
37
- "step": 420,
38
- "total_flos": 8.064772262536032e+17,
39
- "train_loss": 0.15708597316628412,
40
- "train_runtime": 790.3622,
41
- "train_samples_per_second": 35.376,
42
- "train_steps_per_second": 0.531
43
  },
44
  {
45
- "epoch": 56.95,
46
- "eval_accuracy": 0.8709677457809448,
47
- "eval_loss": 0.4480999708175659,
48
- "eval_runtime": 1.4774,
49
- "eval_samples_per_second": 62.95,
50
- "eval_steps_per_second": 8.123,
51
- "step": 420
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
52
  }
53
  ],
54
  "logging_steps": 100,
55
- "max_steps": 420,
56
- "num_train_epochs": 60,
57
  "save_steps": 2000,
58
- "total_flos": 8.064772262536032e+17,
59
  "trial_name": null,
60
  "trial_params": null
61
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 284.7457627118644,
5
+ "eval_steps": 1500,
6
+ "global_step": 2100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 13.56,
13
+ "learning_rate": 4.764285714285715e-05,
14
+ "loss": 0.0462,
15
  "step": 100
16
  },
17
  {
18
  "epoch": 27.12,
19
+ "learning_rate": 4.5261904761904766e-05,
20
+ "loss": 0.0281,
21
  "step": 200
22
  },
23
  {
24
  "epoch": 40.68,
25
+ "learning_rate": 4.2880952380952384e-05,
26
+ "loss": 0.0283,
27
  "step": 300
28
  },
29
  {
30
  "epoch": 54.24,
31
+ "learning_rate": 4.05e-05,
32
+ "loss": 0.0289,
33
  "step": 400
34
  },
35
  {
36
+ "epoch": 67.8,
37
+ "learning_rate": 3.811904761904762e-05,
38
+ "loss": 0.018,
39
+ "step": 500
 
 
 
40
  },
41
  {
42
+ "epoch": 81.36,
43
+ "learning_rate": 3.573809523809524e-05,
44
+ "loss": 0.0269,
45
+ "step": 600
46
+ },
47
+ {
48
+ "epoch": 94.92,
49
+ "learning_rate": 3.3357142857142856e-05,
50
+ "loss": 0.0191,
51
+ "step": 700
52
+ },
53
+ {
54
+ "epoch": 108.47,
55
+ "learning_rate": 3.0976190476190474e-05,
56
+ "loss": 0.0211,
57
+ "step": 800
58
+ },
59
+ {
60
+ "epoch": 122.03,
61
+ "learning_rate": 2.85952380952381e-05,
62
+ "loss": 0.0214,
63
+ "step": 900
64
+ },
65
+ {
66
+ "epoch": 135.59,
67
+ "learning_rate": 2.6214285714285713e-05,
68
+ "loss": 0.0272,
69
+ "step": 1000
70
+ },
71
+ {
72
+ "epoch": 149.15,
73
+ "learning_rate": 2.3833333333333334e-05,
74
+ "loss": 0.0217,
75
+ "step": 1100
76
+ },
77
+ {
78
+ "epoch": 162.71,
79
+ "learning_rate": 2.1452380952380956e-05,
80
+ "loss": 0.0207,
81
+ "step": 1200
82
+ },
83
+ {
84
+ "epoch": 176.27,
85
+ "learning_rate": 1.9071428571428574e-05,
86
+ "loss": 0.0192,
87
+ "step": 1300
88
+ },
89
+ {
90
+ "epoch": 189.83,
91
+ "learning_rate": 1.669047619047619e-05,
92
+ "loss": 0.0164,
93
+ "step": 1400
94
+ },
95
+ {
96
+ "epoch": 203.39,
97
+ "learning_rate": 1.4309523809523811e-05,
98
+ "loss": 0.0194,
99
+ "step": 1500
100
+ },
101
+ {
102
+ "epoch": 203.39,
103
+ "eval_accuracy": 0.9677419066429138,
104
+ "eval_loss": 0.115411177277565,
105
+ "eval_runtime": 1.4468,
106
+ "eval_samples_per_second": 64.281,
107
+ "eval_steps_per_second": 8.294,
108
+ "step": 1500
109
+ },
110
+ {
111
+ "epoch": 216.95,
112
+ "learning_rate": 1.192857142857143e-05,
113
+ "loss": 0.0177,
114
+ "step": 1600
115
+ },
116
+ {
117
+ "epoch": 230.51,
118
+ "learning_rate": 9.547619047619049e-06,
119
+ "loss": 0.0168,
120
+ "step": 1700
121
+ },
122
+ {
123
+ "epoch": 244.07,
124
+ "learning_rate": 7.166666666666667e-06,
125
+ "loss": 0.0171,
126
+ "step": 1800
127
+ },
128
+ {
129
+ "epoch": 257.63,
130
+ "learning_rate": 4.785714285714286e-06,
131
+ "loss": 0.0179,
132
+ "step": 1900
133
+ },
134
+ {
135
+ "epoch": 271.19,
136
+ "learning_rate": 2.404761904761905e-06,
137
+ "loss": 0.0171,
138
+ "step": 2000
139
+ },
140
+ {
141
+ "epoch": 284.75,
142
+ "learning_rate": 2.380952380952381e-08,
143
+ "loss": 0.0163,
144
+ "step": 2100
145
+ },
146
+ {
147
+ "epoch": 284.75,
148
+ "step": 2100,
149
+ "total_flos": 4.0318045822521795e+18,
150
+ "train_loss": 0.022167698939641316,
151
+ "train_runtime": 3842.9627,
152
+ "train_samples_per_second": 36.378,
153
+ "train_steps_per_second": 0.546
154
+ },
155
+ {
156
+ "epoch": 284.75,
157
+ "eval_accuracy": 0.9677419066429138,
158
+ "eval_loss": 0.11643270403146744,
159
+ "eval_runtime": 1.6786,
160
+ "eval_samples_per_second": 55.402,
161
+ "eval_steps_per_second": 7.149,
162
+ "step": 2100
163
  }
164
  ],
165
  "logging_steps": 100,
166
+ "max_steps": 2100,
167
+ "num_train_epochs": 300,
168
  "save_steps": 2000,
169
+ "total_flos": 4.0318045822521795e+18,
170
  "trial_name": null,
171
  "trial_params": null
172
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a6fff3406fc6d17e7151844526156d27d071a854fa3b738d788067583d864923
3
  size 4155
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3dc42704ac631ac8dd8732665b42c45bc13ac858f389067d086ba7cc86a62c22
3
  size 4155