The code related to the paper below: Hongru Ji, Xianghua Li, Mingxin Li, Meng Zhao, Chao Gao, Hybrid relational graphs with sentiment-laden semantic alignment for multimodal emotion recognition in conversation, The 34th International Joint Conference on Artificial Intelligence (IJCAI25), 2025, 2973-2981
Pytorch implementation for the paper: [Hybrid Relational Graphs with Sentiment-laden Semantic Alignment for Multimodal Emotion Recognition in Conversation]
pip install -r requirements.txtThe raw data can be found at IEMOCAP and MELD.
In our paper, we use pre-extracted features. The multimodal features are available at IEMOCAP and MELD.
Before starting, make sure to download the parameters of the T5-base pre-trained model into the pretrained_model directory.
The implementation results may vary with training machines and random seeds.
To train on IEMOCAP:
python main.py -backbone ./pretrained_model -run_type train -dataset iemocap -use_gat -window_size 8 -gat 1 -emotion_first -use_video_mode -use_audio_modeTo train on MELD:
python main.py -backbone ./pretrained_model -run_type train -dataset meld -use_gat -emotion_first -use_video_mode -use_audio_modeWe provide the pre-trained checkpoint on IEMOCAP at here, and the checkpoint on MELD at here.
To predict on IEMOCAP:
python main.py -run_type predict -ckpt ./iemocap-best-model/ckpt -output predict_real.json -dataset iemocap -test_batch_size=64To predict on MELD:
python main.py -run_type predict -ckpt ./meld-best-model/ckpt -output predict_real.json -dataset meld -test_batch_size=64
