Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2406.13923
Cited By
PIN: A Knowledge-Intensive Dataset for Paired and Interleaved Multimodal Documents
20 June 2024
Junjie Wang
Yin Zhang
Yatai Ji
Yuxiang Zhang
Chunyang Jiang
Yubo Wang
Kang Zhu
Zekun Wang
Tiezhen Wang
Wenhao Huang
Jie Fu
Bei Chen
Qunshu Lin
Minghao Liu
Ge Zhang
Wenhu Chen
Re-assign community
ArXiv
PDF
HTML
Papers citing
"PIN: A Knowledge-Intensive Dataset for Paired and Interleaved Multimodal Documents"
4 / 4 papers shown
Title
Baichuan-Omni-1.5 Technical Report
Yadong Li
J. Liu
Tao Zhang
Tao Zhang
S. Chen
...
Jianhua Xu
Haoze Sun
Mingan Lin
Zenan Zhou
Weipeng Chen
AuLLM
67
10
0
28 Jan 2025
2.5 Years in Class: A Multimodal Textbook for Vision-Language Pretraining
Wenqi Zhang
Hang Zhang
Xin Li
Jiashuo Sun
Yongliang Shen
Weiming Lu
Deli Zhao
Yueting Zhuang
Lidong Bing
VLM
37
2
0
01 Jan 2025
Conceptual 12M: Pushing Web-Scale Image-Text Pre-Training To Recognize Long-Tail Visual Concepts
Soravit Changpinyo
P. Sharma
Nan Ding
Radu Soricut
VLM
273
1,077
0
17 Feb 2021
Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision
Chao Jia
Yinfei Yang
Ye Xia
Yi-Ting Chen
Zarana Parekh
Hieu H. Pham
Quoc V. Le
Yun-hsuan Sung
Zhen Li
Tom Duerig
VLM
CLIP
293
3,683
0
11 Feb 2021
1