canlinzhang commited on
Commit
6d19ce7
1 Parent(s): 620ae71

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -2,7 +2,7 @@ This model is fine tuned on the IEMOCAP_speaker_indpt_Ses05F_Ses05M.pickle datas
2
 
3
  The initial pre-trained model is facebook/wav2vec2-base. The fine tune dataset only contains 4 common emotions of IEMOCAP (happy, angry, sad, neutral), **without frustration**. No audio augmentation is applied. Fine-tune dataset audios are also not padded or trimed to fixed length. The length setting is done when fine tuning the transoformer using max_length = 8 sec in the feature extractor.
4
 
5
- After **10** epoches of training, the validation accuracy is around **67%**.
6
 
7
  In order to impliment this model: run the following code in a python script:
8
 
 
2
 
3
  The initial pre-trained model is facebook/wav2vec2-base. The fine tune dataset only contains 4 common emotions of IEMOCAP (happy, angry, sad, neutral), **without frustration**. No audio augmentation is applied. Fine-tune dataset audios are also not padded or trimed to fixed length. The length setting is done when fine tuning the transoformer using max_length = 8 sec in the feature extractor.
4
 
5
+ After **10** epoches of training, the validation accuracy is around **67%**.
6
 
7
  In order to impliment this model: run the following code in a python script:
8