ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2305.13167
  4. Cited By
VLAB: Enhancing Video Language Pre-training by Feature Adapting and
  Blending

VLAB: Enhancing Video Language Pre-training by Feature Adapting and Blending

IEEE transactions on multimedia (IEEE TMM), 2023
22 May 2023
Xingjian He
Sihan Chen
Fan Ma
Zhicheng Huang
Xiaojie Jin
Zikang Liu
Dongmei Fu
Yi Yang
Qingbin Liu
Jiashi Feng
    VLMCLIP
ArXiv (abs)PDFHTML

Papers citing "VLAB: Enhancing Video Language Pre-training by Feature Adapting and Blending"

7 / 7 papers shown
Title
SSL-SSAW: Self-Supervised Learning with Sigmoid Self-Attention Weighting for Question-Based Sign Language Translation
SSL-SSAW: Self-Supervised Learning with Sigmoid Self-Attention Weighting for Question-Based Sign Language Translation
Zekang Liu
Wei Feng
Fanhua Shang
Lianyu Hu
Jichao Feng
Liqing Gao
SLR
164
0
0
17 Sep 2025
Thinking With Videos: Multimodal Tool-Augmented Reinforcement Learning for Long Video Reasoning
Thinking With Videos: Multimodal Tool-Augmented Reinforcement Learning for Long Video Reasoning
H. Zhang
Xin Gu
Jiawen Li
Chixiang Ma
Sule Bai
Chubin Zhang
Bowen Zhang
Zhichao Zhou
Dongliang He
Yansong Tang
OffRLLRM
149
20
0
06 Aug 2025
Pretrained Image-Text Models are Secretly Video Captioners
Pretrained Image-Text Models are Secretly Video CaptionersNorth American Chapter of the Association for Computational Linguistics (NAACL), 2025
Chunhui Zhang
Yiren Jian
Z. Ouyang
Soroush Vosoughi
VLM
405
12
0
20 Feb 2025
Tarsier: Recipes for Training and Evaluating Large Video Description
  Models
Tarsier: Recipes for Training and Evaluating Large Video Description Models
Jiawei Wang
Liping Yuan
Yuchen Zhang
271
112
0
30 Jun 2024
Video-Language Understanding: A Survey from Model Architecture, Model Training, and Data Perspectives
Video-Language Understanding: A Survey from Model Architecture, Model Training, and Data PerspectivesAnnual Meeting of the Association for Computational Linguistics (ACL), 2024
Thong Nguyen
Yi Bin
Junbin Xiao
Leigang Qu
Yicong Li
Jay Zhangjie Wu
Cong-Duy Nguyen
See-Kiong Ng
Luu Anh Tuan
VLM
466
26
1
09 Jun 2024
VideoPrism: A Foundational Visual Encoder for Video Understanding
VideoPrism: A Foundational Visual Encoder for Video Understanding
Long Zhao
N. B. Gundavarapu
Liangzhe Yuan
Hao Zhou
Shen Yan
...
Huisheng Wang
Hartwig Adam
Mikhail Sirotenko
Ting Liu
Boqing Gong
VGen
337
63
0
20 Feb 2024
Incorporating granularity bias as the margin into contrastive loss for
  video captioning
Incorporating granularity bias as the margin into contrastive loss for video captioning
Jiayang Gu
Fengming Yao
105
0
0
25 Nov 2023
1