Benchmarking and Bridging Emotion Conflicts for Multimodal Emotion Reasoning
Paper
β’
2508.01181
β’
Published
This repository contains the MoSEAR.pth model weights for MoSEAR (Modality-Specific Experts with Attention Reallocation), a framework designed to address emotion conflicts in multimodal emotion reasoning tasks.
Key Features:
This model achieves state-of-the-art performance on emotion conflict scenarios:
import torch
# Load checkpoint
checkpoint = torch.load('MoSEAR.pth', map_location='cpu')
# The checkpoint contains:
# - model state dict
# - optimizer state (if included)
# - training metadata
For complete usage with the MoSEAR framework, please refer to the GitHub repository.
# Clone the code repository
git clone https://github.com/ZhiyuanHan-Aaron/MoSEAR.git
cd MoSEAR
# Download this checkpoint
# Place it in the appropriate directory as per the repository instructions
# Run inference
bash scripts/inference.sh
MoSEAR.pth: Main model checkpoint (best performing model)If you use this model in your research, please cite:
@inproceedings{han2025mosear,
title={Benchmarking and Bridging Emotion Conflicts for Multimodal Emotion Reasoning},
author={Han, Zhiyuan and Li, Yifei and Chen, Yanyan and Liang, Xiaohan and Song, Mingming and Peng, Yongsheng and Yin, Guanghao and Ma, Huadong},
booktitle={Proceedings of the 33rd ACM International Conference on Multimedia},
year={2025}
}
Zhiyuan Han
This work builds upon:
This model is released under the BSD 3-Clause License. See the LICENSE for details.
Copyright Β© 2025 Zhiyuan Han