Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2310.04550
Cited By
Module-wise Adaptive Distillation for Multimodality Foundation Models
6 October 2023
Chen Liang
Jiahui Yu
Ming-Hsuan Yang
Matthew A. Brown
Yin Cui
Tuo Zhao
Boqing Gong
Tianyi Zhou
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Module-wise Adaptive Distillation for Multimodality Foundation Models"
4 / 4 papers shown
Title
CLIP-PING: Boosting Lightweight Vision-Language Models with Proximus Intrinsic Neighbors Guidance
Chu Myaet Thwal
Ye Lin Tun
Minh N. H. Nguyen
Eui-nam Huh
Choong Seon Hong
VLM
74
0
0
05 Dec 2024
Dinomaly: The Less Is More Philosophy in Multi-Class Unsupervised Anomaly Detection
Jia Guo
Shuai Lu
Weihang Zhang
Huiqi Li
Huiqi Li
Hongen Liao
ViT
56
7
0
23 May 2024
How Much Can CLIP Benefit Vision-and-Language Tasks?
Sheng Shen
Liunian Harold Li
Hao Tan
Mohit Bansal
Anna Rohrbach
Kai-Wei Chang
Z. Yao
Kurt Keutzer
CLIP
VLM
MLLM
185
403
0
13 Jul 2021
Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision
Chao Jia
Yinfei Yang
Ye Xia
Yi-Ting Chen
Zarana Parekh
Hieu H. Pham
Quoc V. Le
Yun-hsuan Sung
Zhen Li
Tom Duerig
VLM
CLIP
293
3,683
0
11 Feb 2021
1