This is to deplicate the work of [wav2vec2-base-Speech_Emotion_Recognition](https://huggingface.co/DunnBC22/wav2vec2-base-Speech_Emotion_Recognition) *Only little changes are made for success run on google colab.* ### My Version of metrics: |Epoch |Training Loss |Validation Loss |Accuracy |Weighted f1 |Micro f1 |Macro f1 |Weighted recall |Micro recall |Macro recall |Weighted precision |Micro precision |Macro precision | ----|----|----|----|----|----|----|----|----|----|----|----|----| |0 | 1.789200 | 1.548816 | 0.382590 | 0.287415 | 0.382590 | 0.289045 | 0.382590 | 0.382590 | 0.379768 | 0.473585 | 0.382590 | 0.467116 | |1 | 1.789200 | 1.302810 | 0.529823 | 0.511868 | 0.529823 | 0.511619 | 0.529823 | 0.529823 | 0.523766 | 0.552868 | 0.529823 | 0.560496 | |2 | 1.789200 | 1.029921 | 0.672757 | 0.668108 | 0.672757 | 0.669246 | 0.672757 | 0.672757 | 0.676383 | 0.674857 | 0.672757 | 0.673698 | |3 | 1.789200 | 0.968154 | 0.677055 | 0.671986 | 0.677055 | 0.674074 | 0.677055 | 0.677055 | 0.676891 | 0.701300 | 0.677055 | 0.705734 | |4 | 1.789200 | 0.850912 | 0.717894 | 0.714321 | 0.717894 | 0.716527 | 0.717894 | 0.717894 | 0.722476 | 0.716772 | 0.717894 | 0.716698 | |5 | 1.789200 | 0.870916 | 0.710371 | 0.706013 | 0.710371 | 0.708563 | 0.710371 | 0.710371 | 0.713853 | 0.710966 | 0.710371 | 0.712245 | |6 | 1.789200 | 0.827148 | 0.729178 | 0.725336 | 0.729178 | 0.726744 | 0.729178 | 0.729178 | 0.732127 | 0.735935 | 0.729178 | 0.736041 | |7 | 1.789200 | 0.798354 | 0.729715 | 0.727086 | 0.729715 | 0.728847 | 0.729715 | 0.729715 | 0.732476 | 0.729932 | 0.729715 | 0.730688 | |8 | 1.789200 | 0.799373 | 0.735626 | 0.732981 | 0.735626 | 0.735058 | 0.735626 | 0.735626 | 0.738147 | 0.741482 | 0.735626 | 0.742782 | |9 | 1.789200 | 0.810692 | 0.728103 | 0.724754 | 0.728103 | 0.726852 | 0.728103 | 0.728103 | 0.731083 | 0.731919 | 0.728103 | 0.732869 | ```***** Running Evaluation ***** Num examples = 1861 Batch size = 32 [59/59 08:38] {'eval_loss': 0.8106924891471863, 'eval_accuracy': 0.7281031703385277, 'eval_Weighted F1': 0.7247543780750472, 'eval_Micro F1': 0.7281031703385277, 'eval_Macro F1': 0.7268519957485492, 'eval_Weighted Recall': 0.7281031703385277, 'eval_Micro Recall': 0.7281031703385277, 'eval_Macro Recall': 0.7310833557439055, 'eval_Weighted Precision': 0.7319188411210771, 'eval_Micro Precision': 0.7281031703385277, 'eval_Macro Precision': 0.732869407033253, 'eval_runtime': 83.3066, 'eval_samples_per_second': 22.339, 'eval_steps_per_second': 0.708, 'epoch': 9.98} ``` ### Model description This model predicts the emotion of the person speaking in the audio sample. For more information on how it was created, check out the following link: https://github.com/DunnBC22/Vision_Audio_and_Multimodal_Projects/tree/main/Audio-Projects/Emotion%20Detection/Speech%20Emotion%20Detection ### Training and evaluation data Dataset Source: https://www.kaggle.com/datasets/dmitrybabko/speech-emotion-recognition-en