20
0

METDrive: Multi-modal End-to-end Autonomous Driving with Temporal Guidance

Ziang Guo
Xinhao Lin
Zakhar Yagudin
Artem Lykov
Yong Wang
Yanqiang Li
Dzmitry Tsetserukou
Abstract

Multi-modal end-to-end autonomous driving has shown promising advancements in recent work. By embedding more modalities into end-to-end networks, the system's understanding of both static and dynamic aspects of the driving environment is enhanced, thereby improving the safety of autonomous driving. In this paper, we introduce METDrive, an end-to-end system that leverages temporal guidance from the embedded time series features of ego states, including rotation angles, steering, throttle signals, and waypoint vectors. The geometric features derived from perception sensor data and the time series features of ego state data jointly guide the waypoint prediction with the proposed temporal guidance loss function. We evaluated METDrive on the CARLA leaderboard benchmarks, achieving a driving score of 70%, a route completion score of 94%, and an infraction score of 0.78.

View on arXiv
@article{guo2025_2409.12667,
  title={ METDrive: Multi-modal End-to-end Autonomous Driving with Temporal Guidance },
  author={ Ziang Guo and Xinhao Lin and Zakhar Yagudin and Artem Lykov and Yong Wang and Yanqiang Li and Dzmitry Tsetserukou },
  journal={arXiv preprint arXiv:2409.12667},
  year={ 2025 }
}
Comments on this paper