Commit
·
4e169cc
1
Parent(s):
3dfc212
Update README.md
Browse files
README.md
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
This model is fine tuned on the IEMOCAP dataset. We applied volume normalization and data augmentation (noise injection, pitch shift and audio stretching). Also, this is a speaker independent model: We use Ses05F in the IEMOCAP dataset as validation speaker and Ses05M as test speaker.
|
2 |
|
3 |
-
The initial pre-trained model is facebook/wav2vec2-base
|
4 |
|
5 |
After **10** epoches of training, the validation accuracy is around **67%**.
|
6 |
|
|
|
1 |
This model is fine tuned on the IEMOCAP dataset. We applied volume normalization and data augmentation (noise injection, pitch shift and audio stretching). Also, this is a speaker independent model: We use Ses05F in the IEMOCAP dataset as validation speaker and Ses05M as test speaker.
|
2 |
|
3 |
+
The initial pre-trained model is **facebook/wav2vec2-base**. The fine tune dataset only contains 4 common emotions of IEMOCAP (happy, angry, sad, neutral), *without frustration*. The audios are either padded or trimed to 8-sec-long before fine tuning.
|
4 |
|
5 |
After **10** epoches of training, the validation accuracy is around **67%**.
|
6 |
|