Commit
·
6d19ce7
1
Parent(s):
620ae71
Update README.md
Browse files
README.md
CHANGED
@@ -2,7 +2,7 @@ This model is fine tuned on the IEMOCAP_speaker_indpt_Ses05F_Ses05M.pickle datas
|
|
2 |
|
3 |
The initial pre-trained model is facebook/wav2vec2-base. The fine tune dataset only contains 4 common emotions of IEMOCAP (happy, angry, sad, neutral), **without frustration**. No audio augmentation is applied. Fine-tune dataset audios are also not padded or trimed to fixed length. The length setting is done when fine tuning the transoformer using max_length = 8 sec in the feature extractor.
|
4 |
|
5 |
-
After **10** epoches of training, the validation accuracy is around **67%**.
|
6 |
|
7 |
In order to impliment this model: run the following code in a python script:
|
8 |
|
|
|
2 |
|
3 |
The initial pre-trained model is facebook/wav2vec2-base. The fine tune dataset only contains 4 common emotions of IEMOCAP (happy, angry, sad, neutral), **without frustration**. No audio augmentation is applied. Fine-tune dataset audios are also not padded or trimed to fixed length. The length setting is done when fine tuning the transoformer using max_length = 8 sec in the feature extractor.
|
4 |
|
5 |
+
After **10** epoches of training, the validation accuracy is around **67%**.
|
6 |
|
7 |
In order to impliment this model: run the following code in a python script:
|
8 |
|