ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.22242
  4. Cited By
4D-VLA: Spatiotemporal Vision-Language-Action Pretraining with Cross-Scene Calibration
v1v2 (latest)

4D-VLA: Spatiotemporal Vision-Language-Action Pretraining with Cross-Scene Calibration

27 June 2025
J. Zhang
Yurui Chen
Yueming Xu
Ze Huang
Yanpeng Zhou
Yu-Jie Yuan
Xinyue Cai
Guowei Huang
Xingyue Quan
Hang Xu
Li Zhang
ArXiv (abs)PDFHTMLGithub (41★)

Papers citing "4D-VLA: Spatiotemporal Vision-Language-Action Pretraining with Cross-Scene Calibration"

9 / 9 papers shown
SwiftVLA: Unlocking Spatiotemporal Dynamics for Lightweight VLA Models at Minimal Overhead
SwiftVLA: Unlocking Spatiotemporal Dynamics for Lightweight VLA Models at Minimal Overhead
Chaojun Ni
Cheng Chen
Xiaofeng Wang
Zheng Zhu
Wenzhao Zheng
...
Qiang Zhang
Yun Ye
Yang Wang
Guan Huang
Wenjun Mei
117
0
0
30 Nov 2025
Reinforcing Action Policies by Prophesying
Reinforcing Action Policies by Prophesying
Jiahui Zhang
Ze Huang
Chun Gu
Zipei Ma
Li Zhang
233
1
0
25 Nov 2025
VLA-4D: Embedding 4D Awareness into Vision-Language-Action Models for SpatioTemporally Coherent Robotic Manipulation
VLA-4D: Embedding 4D Awareness into Vision-Language-Action Models for SpatioTemporally Coherent Robotic Manipulation
Hanyu Zhou
Chuanhao Ma
Gim Hee Lee
193
0
0
21 Nov 2025
QDepth-VLA: Quantized Depth Prediction as Auxiliary Supervision for Vision-Language-Action Models
QDepth-VLA: Quantized Depth Prediction as Auxiliary Supervision for Vision-Language-Action Models
Y. Li
Yihao Chen
Mingcai Zhou
Haoran Li
Zhengtao Zhang
Dongbin Zhao
VLM
132
1
0
16 Oct 2025
Spatial Forcing: Implicit Spatial Representation Alignment for Vision-language-action Model
Spatial Forcing: Implicit Spatial Representation Alignment for Vision-language-action Model
Fuhao Li
Wenxuan Song
Han Zhao
Jingbo Wang
Pengxiang Ding
Donglin Wang
Long Zeng
Haoang Li
217
7
0
14 Oct 2025
Pure Vision Language Action (VLA) Models: A Comprehensive Survey
Pure Vision Language Action (VLA) Models: A Comprehensive Survey
Dapeng Zhang
Jin Sun
Chenghui Hu
Xiaoyan Wu
Zhenlong Yuan
R. Zhou
Fei Shen
Qingguo Zhou
LM&Ro
325
15
0
23 Sep 2025
VLA-Adapter: An Effective Paradigm for Tiny-Scale Vision-Language-Action Model
VLA-Adapter: An Effective Paradigm for Tiny-Scale Vision-Language-Action Model
Yihao Wang
Pengxiang Ding
Lingxiao Li
Can Cui
Zirui Ge
...
Yifan Tang
Wenhui Wang
Ru Zhang
Jianyi Liu
Donglin Wang
269
29
0
11 Sep 2025
MemoryVLA: Perceptual-Cognitive Memory in Vision-Language-Action Models for Robotic Manipulation
MemoryVLA: Perceptual-Cognitive Memory in Vision-Language-Action Models for Robotic Manipulation
Hao Shi
Bin Xie
Yingfei Liu
Lin Sun
Fengrong Liu
Tiancai Wang
Erjin Zhou
Haoqiang Fan
Xiangyu Zhang
Gao Huang
LM&Ro
132
26
0
26 Aug 2025
Large VLM-based Vision-Language-Action Models for Robotic Manipulation: A Survey
Large VLM-based Vision-Language-Action Models for Robotic Manipulation: A Survey
Rui Shao
W. Li
Lingsen Zhang
Renshan Zhang
Zhiyang Liu
Ran Chen
Liqiang Nie
LM&Ro
249
29
0
18 Aug 2025
1
Page 1 of 1