From 874fe6191e247198e8ca9ba44da20f5338ce0995 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Enrique=20Hern=C3=A1ndez=20Calabr=C3=A9s?= Date: Wed, 14 Jul 2021 12:47:57 +0000 Subject: [PATCH] Update README.md --- README.md | 12 ++++++++++-- 1 file changed, 10 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 441cc62..abc5688 100644 --- a/README.md +++ b/README.md @@ -11,9 +11,13 @@ model_index: -# wav2vec2-lg-xlsr-en-speech-emotion-recognition +# Speech Emotion Recognition By Fine-Tuning Wav2Vec 2.0 -This model is a fine-tuned version of [jonatasgrosman/wav2vec2-large-xlsr-53-english](https://huggingface.co/jonatasgrosman/wav2vec2-large-xlsr-53-english) for a Speech Emotion Recognition (SER) task. + +__Important:__ This model it's not yet implementable due to missing built-in functions in HuggingFace for speech classification tasks. I'm working on the instructions of how to use it and a repository where the code will be available soon. Thank you anyway! + + +The model is a fine-tuned version of [jonatasgrosman/wav2vec2-large-xlsr-53-english](https://huggingface.co/jonatasgrosman/wav2vec2-large-xlsr-53-english) for a Speech Emotion Recognition (SER) task. The dataset used to fine-tune the original pre-trained model is the [RAVDESS dataset](https://zenodo.org/record/1188976#.YO6yI-gzaUk). This dataset provides 1440 samples of recordings from actors performing on 8 different emotions in English, which are: @@ -72,6 +76,10 @@ The following hyperparameters were used during training: | 0.4581 | 2.72 | 390 | 0.4719 | 0.8467 | | 0.3967 | 2.93 | 420 | 0.5023 | 0.8223 | +## Contact + +Any doubt, contact me on [Twitter](https://twitter.com/ehcalabres) (GitHub repo soon). + ### Framework versions