ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2202.09263
  4. Cited By
Is Cross-Attention Preferable to Self-Attention for Multi-Modal Emotion
  Recognition?

Is Cross-Attention Preferable to Self-Attention for Multi-Modal Emotion Recognition?

IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2022
18 February 2022
Vandana Rajan
Alessio Brutti
Andrea Cavallaro
ArXiv (abs)PDFHTMLGithub (59★)

Papers citing "Is Cross-Attention Preferable to Self-Attention for Multi-Modal Emotion Recognition?"

17 / 17 papers shown
MF-GCN: A Multi-Frequency Graph Convolutional Network for Tri-Modal Depression Detection Using Eye-Tracking, Facial, and Acoustic Features
MF-GCN: A Multi-Frequency Graph Convolutional Network for Tri-Modal Depression Detection Using Eye-Tracking, Facial, and Acoustic Features
Sejuti Rahman
Swakshar Deb
MD. Sameer Iqbal Chowdhury
MD. Jubair Ahmed Sourov
Mohammad Shamsuddin
152
0
0
19 Nov 2025
IsoNet: Causal Analysis of Multimodal Transformers for Neuromuscular Gesture Classification
IsoNet: Causal Analysis of Multimodal Transformers for Neuromuscular Gesture Classification
Eion Tyacke
Kunal Gupta
Jay Patel
Rui Li
168
0
0
20 Jun 2025
CROSS-GAiT: Cross-Attention-Based Multimodal Representation Fusion for Parametric Gait Adaptation in Complex Terrains
CROSS-GAiT: Cross-Attention-Based Multimodal Representation Fusion for Parametric Gait Adaptation in Complex Terrains
Gershom Seneviratne
K. Weerakoon
Mohamed Bashir Elnoor
Vignesh Rajgopal
Harshavarthan Varatharajan
Mohamed Khalid M Jaffar
Jason Pusey
Wanrong Zhu
CVBM
369
8
0
25 Sep 2024
A Depression Detection Method Based on Multi-Modal Feature Fusion Using
  Cross-Attention
A Depression Detection Method Based on Multi-Modal Feature Fusion Using Cross-Attention
Shengjie Li
Yinhao Xiao
305
8
0
02 Jul 2024
Stock Movement Prediction with Multimodal Stable Fusion via Gated
  Cross-Attention Mechanism
Stock Movement Prediction with Multimodal Stable Fusion via Gated Cross-Attention Mechanism
Chang Zong
Jian Shao
Weiming Lu
Yueting Zhuang
300
12
0
06 Jun 2024
Leveraging Speech for Gesture Detection in Multimodal Communication
Leveraging Speech for Gesture Detection in Multimodal Communication
E. Ghaleb
I. Burenko
Marlou Rasenberg
Wim Pouw
Ivan Toni
Peter Uhrig
Anna Wilson
Judith Holler
Asli Ozyurek
Raquel Fernández
SLR
257
7
0
23 Apr 2024
Cross-Attention Fusion of Visual and Geometric Features for Large
  Vocabulary Arabic Lipreading
Cross-Attention Fusion of Visual and Geometric Features for Large Vocabulary Arabic Lipreading
Samar Daou
Ahmed Rekik
A. Ben-Hamadou
Abdelaziz Kallel
240
6
0
18 Feb 2024
Comment-aided Video-Language Alignment via Contrastive Pre-training for
  Short-form Video Humor Detection
Comment-aided Video-Language Alignment via Contrastive Pre-training for Short-form Video Humor Detection
Yang Liu
Tongfei Shen
Dong Zhang
Qingying Sun
Shoushan Li
Guodong Zhou
297
6
0
14 Feb 2024
HiCMAE: Hierarchical Contrastive Masked Autoencoder for Self-Supervised
  Audio-Visual Emotion Recognition
HiCMAE: Hierarchical Contrastive Masked Autoencoder for Self-Supervised Audio-Visual Emotion RecognitionInformation Fusion (Inf. Fusion), 2024
Guoying Zhao
Zheng Lian
Yinan Han
Jianhua Tao
341
78
0
11 Jan 2024
Joyful: Joint Modality Fusion and Graph Contrastive Learning for
  Multimodal Emotion Recognition
Joyful: Joint Modality Fusion and Graph Contrastive Learning for Multimodal Emotion Recognition
Dongyuan Li
Yusong Wang
Kotaro Funakoshi
Manabu Okumura
215
65
0
18 Nov 2023
Interaction is all You Need? A Study of Robots Ability to Understand and
  Execute
Interaction is all You Need? A Study of Robots Ability to Understand and Execute
Kushal Koshti
Nidhir Bhavsar
243
1
0
13 Nov 2023
Exploring Emotion Expression Recognition in Older Adults Interacting
  with a Virtual Coach
Exploring Emotion Expression Recognition in Older Adults Interacting with a Virtual CoachIEEE Transactions on Affective Computing (IEEE Trans. Affective Comput.), 2023
Cristina Palmero
Mikel de Velasco
Mohamed Amine Hmani
Aymen Mtibaa
Leila Ben Letaifa
...
Anna Esposito
M. El-Yacoubi
Dijana Petrovska – Delacretaz
M. Inés Torres
Sergio Escalera
292
11
0
09 Nov 2023
3M-TRANSFORMER: A Multi-Stage Multi-Stream Multimodal Transformer for
  Embodied Turn-Taking Prediction
3M-TRANSFORMER: A Multi-Stage Multi-Stream Multimodal Transformer for Embodied Turn-Taking PredictionIEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2023
Mehdi Fatan
Emanuele Mincato
Dimitra Pintzou
Mariella Dimiccoli
333
2
0
23 Oct 2023
A Survey on Image-text Multimodal Models
A Survey on Image-text Multimodal Models
Ruifeng Guo
Jingxuan Wei
Linzhuang Sun
Khai-Nguyen Nguyen
Guiyong Chang
Dawei Liu
Sibo Zhang
Zhengbing Yao
Mingjun Xu
Liping Bu
VLM
413
25
0
23 Sep 2023
Cross-Attention is Not Enough: Incongruity-Aware Dynamic Hierarchical
  Fusion for Multimodal Affect Recognition
Cross-Attention is Not Enough: Incongruity-Aware Dynamic Hierarchical Fusion for Multimodal Affect Recognition
Yaoting Wang
Yuanchao Li
Paul Pu Liang
Louis-Philippe Morency
P. Bell
Catherine Lai
CVBM
363
9
0
23 May 2023
Efficient Multimodal Transformer with Dual-Level Feature Restoration for
  Robust Multimodal Sentiment Analysis
Efficient Multimodal Transformer with Dual-Level Feature Restoration for Robust Multimodal Sentiment AnalysisIEEE Transactions on Affective Computing (IEEE TAC), 2022
Guoying Zhao
Zheng Lian
B. Liu
Jianhua Tao
313
122
0
16 Aug 2022
Continuous-Time Audiovisual Fusion with Recurrence vs. Attention for
  In-The-Wild Affect Recognition
Continuous-Time Audiovisual Fusion with Recurrence vs. Attention for In-The-Wild Affect Recognition
Vincent Karas
M. Tellamekala
Adria Mallol-Ragolta
Michel Valstar
Björn W. Schuller
309
17
0
24 Mar 2022
1
Page 1 of 1