ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1804.06786
  4. Cited By
Quantifying the visual concreteness of words and topics in multimodal
  datasets
v1v2 (latest)

Quantifying the visual concreteness of words and topics in multimodal datasets

18 April 2018
Jack Hessel
David M. Mimno
Lillian Lee
ArXiv (abs)PDFHTML

Papers citing "Quantifying the visual concreteness of words and topics in multimodal datasets"

20 / 20 papers shown
The impact of abstract and object tags on image privacy classification
The impact of abstract and object tags on image privacy classification
Darya Baranouskaya
Andrea Cavallaro
VLM
115
0
0
09 Oct 2025
Uncovering Visual-Semantic Psycholinguistic Properties from the Distributional Structure of Text Embedding Space
Uncovering Visual-Semantic Psycholinguistic Properties from the Distributional Structure of Text Embedding SpaceAnnual Meeting of the Association for Computational Linguistics (ACL), 2025
Si Wu
Sebastian Bruch
294
0
0
29 May 2025
Multi-Modal Framing Analysis of News
Multi-Modal Framing Analysis of News
Arnav Arora
Srishti Yadav
Maria Antoniak
Serge J. Belongie
Isabelle Augenstein
255
7
0
26 Mar 2025
Unveiling the Mystery of Visual Attributes of Concrete and Abstract
  Concepts: Variability, Nearest Neighbors, and Challenging Categories
Unveiling the Mystery of Visual Attributes of Concrete and Abstract Concepts: Variability, Nearest Neighbors, and Challenging CategoriesConference on Empirical Methods in Natural Language Processing (EMNLP), 2024
Tarun Tater
Sabine Schulte im Walde
Diego Frassinelli
207
2
0
15 Oct 2024
Understanding the Cognitive Complexity in Language Elicited by Product
  Images
Understanding the Cognitive Complexity in Language Elicited by Product Images
Yan-Ying Chen
Shabnam Hakimi
Monica P Van
Francine Chen
Matthew K. Hong
M. Klenk
Charlene C. Wu
256
2
0
25 Sep 2024
Mitigating Open-Vocabulary Caption Hallucinations
Mitigating Open-Vocabulary Caption Hallucinations
Assaf Ben-Kish
Moran Yanuka
Morris Alper
Raja Giryes
Hadar Averbuch-Elor
MLLMVLM
399
14
0
06 Dec 2023
Composition and Deformance: Measuring Imageability with a Text-to-Image
  Model
Composition and Deformance: Measuring Imageability with a Text-to-Image Model
Si Wu
David A. Smith
EGVMCoGe
127
5
0
05 Jun 2023
Is BERT Blind? Exploring the Effect of Vision-and-Language Pretraining
  on Visual Language Understanding
Is BERT Blind? Exploring the Effect of Vision-and-Language Pretraining on Visual Language UnderstandingComputer Vision and Pattern Recognition (CVPR), 2023
Morris Alper
Michael Fiman
Hadar Averbuch-Elor
VLMLRM
243
17
0
21 Mar 2023
PACS: A Dataset for Physical Audiovisual CommonSense Reasoning
PACS: A Dataset for Physical Audiovisual CommonSense ReasoningEuropean Conference on Computer Vision (ECCV), 2022
Samuel Yu
Peter Wu
Paul Pu Liang
Ruslan Salakhutdinov
Louis-Philippe Morency
LRM
281
19
0
21 Mar 2022
Billion-Scale Pretraining with Vision Transformers for Multi-Task Visual
  Representations
Billion-Scale Pretraining with Vision Transformers for Multi-Task Visual RepresentationsIEEE Workshop/Winter Conference on Applications of Computer Vision (WACV), 2021
Josh Beal
Hao Wu
Dong Huk Park
Andrew Zhai
Dmitry Kislyuk
ViT
202
31
0
12 Aug 2021
Separating Skills and Concepts for Novel Visual Question Answering
Separating Skills and Concepts for Novel Visual Question AnsweringComputer Vision and Pattern Recognition (CVPR), 2021
Spencer Whitehead
Hui Wu
Heng Ji
Rogerio Feris
Kate Saenko
CoGe
185
38
0
19 Jul 2021
Revisiting Document Representations for Large-Scale Zero-Shot Learning
Revisiting Document Representations for Large-Scale Zero-Shot LearningNorth American Chapter of the Association for Computational Linguistics (NAACL), 2021
Jihyung Kil
Wei-Lun Chao
VLM
183
11
0
21 Apr 2021
Domain-Specific Lexical Grounding in Noisy Visual-Textual Documents
Domain-Specific Lexical Grounding in Noisy Visual-Textual DocumentsConference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Gregory Yauney
Jack Hessel
David M. Mimno
ObjD
120
0
0
30 Oct 2020
Visual Pivoting for (Unsupervised) Entity Alignment
Visual Pivoting for (Unsupervised) Entity Alignment
Fangyu Liu
Muhao Chen
Dan Roth
Nigel Collier
OCL
308
150
0
28 Sep 2020
What is Learned in Visually Grounded Neural Syntax Acquisition
What is Learned in Visually Grounded Neural Syntax AcquisitionAnnual Meeting of the Association for Computational Linguistics (ACL), 2020
Noriyuki Kojima
Hadar Averbuch-Elor
Alexander M. Rush
Yoav Artzi
218
22
0
04 May 2020
Beyond Instructional Videos: Probing for More Diverse Visual-Textual
  Grounding on YouTube
Beyond Instructional Videos: Probing for More Diverse Visual-Textual Grounding on YouTubeConference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Jack Hessel
Zhenhai Zhu
Bo Pang
Radu Soricut
225
4
0
29 Apr 2020
PIQA: Reasoning about Physical Commonsense in Natural Language
PIQA: Reasoning about Physical Commonsense in Natural LanguageAAAI Conference on Artificial Intelligence (AAAI), 2019
Yonatan Bisk
Rowan Zellers
Ronan Le Bras
Jianfeng Gao
Yejin Choi
OODLRM
1.6K
2,547
0
26 Nov 2019
Visually Grounded Neural Syntax Acquisition
Visually Grounded Neural Syntax AcquisitionAnnual Meeting of the Association for Computational Linguistics (ACL), 2019
Freda Shi
Jiayuan Mao
Kevin Gimpel
Karen Livescu
NAI
213
86
0
07 Jun 2019
Unsupervised Discovery of Multimodal Links in Multi-image,
  Multi-sentence Documents
Unsupervised Discovery of Multimodal Links in Multi-image, Multi-sentence Documents
Jack Hessel
Lillian Lee
David M. Mimno
162
31
0
16 Apr 2019
Multimodal Grounding for Language Processing
Multimodal Grounding for Language Processing
Lisa Beinborn
Teresa Botschen
Iryna Gurevych
160
37
0
17 Jun 2018
1
Page 1 of 1