Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2406.10165
Cited By
CarLLaVA: Vision language models for camera-only closed-loop driving
14 June 2024
Katrin Renz
Long Chen
Ana-Maria Marcu
Jan Hünermann
Benoît Hanotte
Alice Karnsund
Jamie Shotton
Elahe Arani
Oleg Sinavski
VLM
Re-assign community
ArXiv
PDF
HTML
Papers citing
"CarLLaVA: Vision language models for camera-only closed-loop driving"
8 / 8 papers shown
Title
CoLMDriver: LLM-based Negotiation Benefits Cooperative Autonomous Driving
Changxing Liu
Genjia Liu
Z. Wang
Jinchang Yang
Siheng Chen
62
0
0
11 Mar 2025
VaViM and VaVAM: Autonomous Driving through Video Generative Modeling
Florent Bartoccioni
Elias Ramzi
Victor Besnier
Shashanka Venkataramanan
Tuan-Hung Vu
...
Mickael Chen
Éloi Zablocki
Andrei Bursuc
Eduardo Valle
Matthieu Cord
VGen
68
1
0
24 Feb 2025
OccLLaMA: An Occupancy-Language-Action Generative World Model for Autonomous Driving
Julong Wei
Shanshuai Yuan
Pengfei Li
Qingda Hu
Zhongxue Gan
Wenchao Ding
VLM
16
17
0
05 Sep 2024
CoVLA: Comprehensive Vision-Language-Action Dataset for Autonomous Driving
Hidehisa Arai
Keita Miwa
Kento Sasaki
Yu Yamaguchi
Kohei Watanabe
Shunsuke Aoki
Issei Yamamoto
35
9
0
19 Aug 2024
LanguageMPC: Large Language Models as Decision Makers for Autonomous Driving
Hao Sha
Yao Mu
Yuxuan Jiang
Li Chen
Chenfeng Xu
Ping Luo
Shengbo Eben Li
Masayoshi Tomizuka
Wei Zhan
Mingyu Ding
102
154
0
04 Oct 2023
BLIP-2: Bootstrapping Language-Image Pre-training with Frozen Image Encoders and Large Language Models
Junnan Li
Dongxu Li
Silvio Savarese
Steven C. H. Hoi
VLM
MLLM
244
4,186
0
30 Jan 2023
TransFuser: Imitation with Transformer-Based Sensor Fusion for Autonomous Driving
Kashyap Chitta
Aditya Prakash
Bernhard Jaeger
Zehao Yu
Katrin Renz
Andreas Geiger
ViT
81
291
0
31 May 2022
Masked Autoencoders Are Scalable Vision Learners
Kaiming He
Xinlei Chen
Saining Xie
Yanghao Li
Piotr Dollár
Ross B. Girshick
ViT
TPM
258
7,337
0
11 Nov 2021
1