Implementation for the paper (Interspeech 2021). The paper has been accepted, its full-text will be shared after publication.
Towards the Explainability of Multimodal Speech Emotion Recognition
Puneet Kumar, Vishesh Kaushik, and Balasubramanian Raman
- Install Anaconda or Miniconda distribution and create a conda environment with Python 3.6+.
- Install the requirements using the following command:
pip install -r Requirements.txt
- Download glove.6B.zip, unzip and keep in
glove.6B
folder. - Download the required datasets.
- For IEMOCAP Dataset:
RunData_Preprocess(IEMOCAP).ipynb
in Jupyter Notebook, then
RunTraining(IEMOCAP).ipynb
in Jupyter Notebook, then
RunAnalysis(IEMOCAP).ipynb
in Jupyter Notebook.
OR
Runmain_IEMOCAP.py
in the terminal/command-line using the following command:
python main_IEMOCAP.py --epoch=100
- For MSP-IMPROV Dataset:
RunData_Preprocess(IMPROV).ipynb
in Jupyter Notebook, then
RunTraining(IMPROV).ipynb
in Jupyter Notebook, then
RunAnalysis(IMPROV).ipynb
in Jupyter Notebook.
OR
Runmain_IMPROV.py
in the terminal/command-line using the following command:
python main_IMPROV.py --epoch=100
- For RAVDESS Dataset:
RunData_Preprocess(RAVDESS).ipynb
in Jupyter Notebook, then
RunTraining(RAVDESS).ipynb
in Jupyter Notebook, then
RunAnalysis(RAVDESS).ipynb
in Jupyter Notebook.
OR
Runmain_RAVDESS.py
in the terminal/command-line using the following command:
python main_RAVDESS.py --epoch=100
By default, the code saves the model checkpoints in the log-1
folder.
The tensorboard log files are saved in the log-1
folder. These files can be accessed using the following command:
tensorboard --logdir "/log-1"