Update README.md
Browse files
README.md
CHANGED
@@ -19,7 +19,7 @@ should probably proofread and complete it, then remove this comment. -->
|
|
19 |
|
20 |
# libri-alpha-0.85-Temp-1-processor-change
|
21 |
|
22 |
-
This model is a
|
23 |
It achieves the following results on the evaluation set:
|
24 |
- Loss: 78.4467
|
25 |
- Wer: 0.1153
|
@@ -37,6 +37,7 @@ More information needed
|
|
37 |
More information needed
|
38 |
|
39 |
## Training procedure
|
|
|
40 |
|
41 |
### Training hyperparameters
|
42 |
|
@@ -52,6 +53,9 @@ The following hyperparameters were used during training:
|
|
52 |
- lr_scheduler_warmup_steps: 500
|
53 |
- num_epochs: 30
|
54 |
- mixed_precision_training: Native AMP
|
|
|
|
|
|
|
55 |
|
56 |
### Training results
|
57 |
|
|
|
19 |
|
20 |
# libri-alpha-0.85-Temp-1-processor-change
|
21 |
|
22 |
+
This model is a distilled version of [Wav2vec2](https://huggingface.co/) on the 30% of the Librispeech-clean.100 dataset.
|
23 |
It achieves the following results on the evaluation set:
|
24 |
- Loss: 78.4467
|
25 |
- Wer: 0.1153
|
|
|
37 |
More information needed
|
38 |
|
39 |
## Training procedure
|
40 |
+
Knowledge distillation of Wav2vec2-base-960h teacher model with 6 attention layers for student model.
|
41 |
|
42 |
### Training hyperparameters
|
43 |
|
|
|
53 |
- lr_scheduler_warmup_steps: 500
|
54 |
- num_epochs: 30
|
55 |
- mixed_precision_training: Native AMP
|
56 |
+
- alpha: 0.75(ignore name of repo)
|
57 |
+
- temperature: 1
|
58 |
+
|
59 |
|
60 |
### Training results
|
61 |
|