Adaptive multimodal transformer based on exchanging for multimodal sentiment analysis

Abstract Multimodal sentiment analysis significantly improves sentiment classification performance by integrating cross-modal emotional cues. However, existing methods still face challenges in key issues such as modal distribution differences, cross-modal interaction efficiency, and contextual corre...

Full description

Saved in:
Bibliographic Details
Main Authors: Gulanbaier Tuerhong, Feifan Fu, Mairidan Wushouer
Format: Article
Language:English
Published: Nature Portfolio 2025-07-01
Series:Scientific Reports
Subjects:
Online Access:https://doi.org/10.1038/s41598-025-11848-4
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Abstract Multimodal sentiment analysis significantly improves sentiment classification performance by integrating cross-modal emotional cues. However, existing methods still face challenges in key issues such as modal distribution differences, cross-modal interaction efficiency, and contextual correlation modeling. To address these issues, this paper proposes an Adaptive Multimodal Transformer based on Exchanging (AMTE) model, which employs an exchange fusion mechanism. When the local emotional features of one modality are insufficient, AMTE enhances them with the global features of another modality, thus bridging cross-modal semantic differences while retaining modality specificity, achieving efficient fusion. AMTE’s multi-scale hierarchical fusion mechanism constructs an adaptive hyper-modal representation, effectively reducing the distribution differences between modalities. In the cross-modal exchange fusion stage, the language modality serves as the dominant modality, deeply fusing with the hyper-modal representation and combining contextual information for sentiment prediction. Experimental results show that AMTE achieves excellent performance, with binary sentiment classification accuracies of 89.18%, 88.28%, and 81.84% on the public datasets CMU-MOSI, CMU-MOSEI, and CH-SIMS, respectively.
ISSN:2045-2322