From Discriminative Emotion Recognition to Generative Emotion Understanding — the 4th edition of the MER series, featuring four new tracks.
MER2026 marks the fourth edition of the MER series of challenges.
The MER series provides valuable data resources to the research community and offers tasks centered on recent research trends,
establishing itself as one of the largest challenges in the field.
Throughout its history, the focus of MER has shifted from discriminative emotion recognition to
generative emotion understanding.
Specifically, MER2023 concentrated on discriminative emotion recognition, restricting the scope to fixed basic labels.
In MER2024 and MER2025, we transitioned to generative emotion understanding, leveraging the extensive vocabulary and
multimodal understanding capabilities of MLLMs to facilitate fine-grained and explainable emotion recognition.
Building on this trajectory, MER2026 contains four tracks:
MER-Cross, MER-FG, MER-Prefer, and MER-PS.
Contact: merchallenge.contact@gmail.com
· lianzheng@tongji.edu.cn
A newly introduced track that shifts focus from individual scenarios to dyadic interactions. When speaker s₁ is talking, MER-Cross targets the emotion of the listening s₂ (interlocutor) rather than the speaker, enabling capture of both sides of a conversation.
Human emotion extends far beyond basic labels. In this track, participants can predict any number of emotion labels across diverse categories, expanding recognition scope from basic to more nuanced emotions.
A newly introduced track predicting which of two emotion descriptions is preferred by human annotators for a given video — a critical component for training reward models in emotion understanding.
Shifts emotion recognition from observable behaviors to physiological evidence. Using synchronized EEG (64-ch, 1000 Hz) and fNIRS (51-ch, 47.62 Hz) signals from 30 participants watching 15 emotion-eliciting video clips.
A key feature is real-time dynamic emotion annotation — subjects continuously report valence and arousal via joystick at 1 Hz. The goal is to estimate the valence–arousal trajectory from brain signals.
[1] Zheng Lian et al. MER 2023: Multi-label Learning, Modality Robustness, and Semi-supervised Learning. Proceedings of the 31st ACM International Conference on Multimedia, 2023.
[2] Zheng Lian et al. MER 2024: Semi-supervised Learning, Noise Robustness, and Open-vocabulary Multimodal Emotion Recognition. Proceedings of the 2nd International Workshop on Multimodal and Responsible Affective Computing, 2024.
[3] Zheng Lian et al. MER 2025: When Affective Computing Meets Large Language Models. Proceedings of the 33rd ACM International Conference on Multimedia, 2025.
[4] Zheng Lian et al. Explainable Multimodal Emotion Reasoning. arXiv:2306.15401, 2023.
[5] Zheng Lian et al. OV-MER: Towards Open-Vocabulary Multimodal Emotion Recognition. Forty-second International Conference on Machine Learning (ICML), 2025.
[6] Zheng Lian et al. AffectGPT: A New Dataset, Model, and Benchmark for Emotion Understanding with Multimodal Large Language Models. Forty-second International Conference on Machine Learning (ICML), 2025.
[7] Zheng Lian et al. EmoPrefer: Can Large Language Models Understand Human Emotion Preferences? International Conference on Learning Representations (ICLR), 2026.
[8] Zheng Lian et al. AffectGPT-R1: Leveraging Reinforcement Learning for Open-Vocabulary Multimodal Emotion Recognition. arXiv:2508.01318, 2026.
Format: ACM double-column template (paper template). Double-blind review — anonymize all author information.
Length:
Review: All papers peer-reviewed; accepted papers published in the ACM Digital Library.
Challenge solutions and any work on multimodal affective computing are welcome.
All deadlines at 23:59 Anywhere on Earth (AoE).
We would like to express our sincere gratitude to the Dataset Chairs for their support during the dataset construction process: Xiaojiang Peng (Shenzhen Technology University), Kele Xu (National University of Defense Technology), Fei Ma (Guangdong Lab of Artificial Intelligence and Digital Economy (SZ)), Ziyu Jia (Institute of Automation, CAS), Laizhong Cui (Shenzhen University), and Zebang Cheng (Shenzhen University). We also thank the members of our Annotation Team for their efforts: Zelin Yan (Tongji University), Liyi Liu (Tongji University), Chenxi Zhou (National University of Defense Technology), Yuan Cao (National University of Defense Technology), Kaiyao Li (Shenzhen University), Dawei Huang (Shenzhen University), and Hanwen Du (Shenzhen University). Finally, we thank the ACM Multimedia 2026 Challenge Chairs for giving us the opportunity to organize MER2026 at this premier conference.