Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2312.09818
Cited By
SMILE: Multimodal Dataset for Understanding Laughter in Video with Language Models
15 December 2023
Lee Hyun
Kim Sung-Bin
Seungju Han
Youngjae Yu
Tae-Hyun Oh
Re-assign community
ArXiv
PDF
HTML
Papers citing
"SMILE: Multimodal Dataset for Understanding Laughter in Video with Language Models"
15 / 15 papers shown
Title
Why We Feel: Breaking Boundaries in Emotional Reasoning with Multimodal Large Language Models
Yuxiang Lin
Jingdong Sun
Zhi-Qi Cheng
Jue Wang
Haomin Liang
Zebang Cheng
Yifei Dong
Jun-Yan He
Xiaojiang Peng
Xian-Sheng Hua
39
0
0
10 Apr 2025
Towards Online Multi-Modal Social Interaction Understanding
X. Li
Shijian Deng
Bolin Lai
Weiguo Pian
James M. Rehg
Yapeng Tian
41
0
0
25 Mar 2025
ViSpeak: Visual Instruction Feedback in Streaming Videos
Shenghao Fu
Q. Yang
Yuan-Ming Li
Yi-Xing Peng
Kun-Yu Lin
Xihan Wei
Jian-Fang Hu
Xiaohua Xie
Wei-Shi Zheng
VLM
58
1
0
17 Mar 2025
BottleHumor: Self-Informed Humor Explanation using the Information Bottleneck Principle
EunJeong Hwang
Peter West
Vered Shwartz
36
1
0
22 Feb 2025
video-SALMONN-o1: Reasoning-enhanced Audio-visual Large Language Model
Guangzhi Sun
Yudong Yang
Jimin Zhuang
Changli Tang
Y. Li
W. Li
Z. Ma
Chao Zhang
LRM
MLLM
VLM
64
2
0
17 Feb 2025
Generative Emotion Cause Explanation in Multimodal Conversations
Lin Wang
Xiaocui Yang
Shi Feng
Daling Wang
Yifei Zhang
23
0
0
01 Nov 2024
AVHBench: A Cross-Modal Hallucination Benchmark for Audio-Visual Large Language Models
Kim Sung-Bin
Oh Hyun-Bin
JungMok Lee
Arda Senocak
Joon Son Chung
Tae-Hyun Oh
MLLM
VLM
29
2
0
23 Oct 2024
BQA: Body Language Question Answering Dataset for Video Large Language Models
Shintaro Ozaki
Kazuki Hayashi
Miyu Oba
Yusuke Sakai
Hidetaka Kamigaito
Taro Watanabe
30
1
0
17 Oct 2024
Towards Social AI: A Survey on Understanding Social Interactions
Sangmin Lee
Minzhi Li
Bolin Lai
Wenqi Jia
Fiona Ryan
...
Ozgur Kara
Bikram Boote
Weiyan Shi
Diyi Yang
James M. Rehg
18
4
0
05 Sep 2024
End-to-end Semantic-centric Video-based Multimodal Affective Computing
Ronghao Lin
Ying Zeng
Sijie Mai
Haifeng Hu
VGen
33
0
0
14 Aug 2024
BEAF: Observing BEfore-AFter Changes to Evaluate Hallucination in Vision-language Models
Moon Ye-Bin
Nam Hyeon-Woo
Wonseok Choi
Tae-Hyun Oh
MLLM
32
6
0
18 Jul 2024
Can Language Models Laugh at YouTube Short-form Videos?
Dayoon Ko
Sangho Lee
Gunhee Kim
19
6
0
22 Oct 2023
SYNAuG: Exploiting Synthetic Data for Data Imbalance Problems
Moon Ye-Bin
Nam Hyeon-Woo
Wonseok Choi
Nayeong Kim
Suha Kwak
Tae-Hyun Oh
DiffM
16
3
0
02 Aug 2023
Language Models with Image Descriptors are Strong Few-Shot Video-Language Learners
Zhenhailong Wang
Manling Li
Ruochen Xu
Luowei Zhou
Jie Lei
...
Chenguang Zhu
Derek Hoiem
Shih-Fu Chang
Mohit Bansal
Heng Ji
MLLM
VLM
164
134
0
22 May 2022
Training language models to follow instructions with human feedback
Long Ouyang
Jeff Wu
Xu Jiang
Diogo Almeida
Carroll L. Wainwright
...
Amanda Askell
Peter Welinder
Paul Christiano
Jan Leike
Ryan J. Lowe
OSLM
ALM
301
11,730
0
04 Mar 2022
1