V2V: Scaling Event-Based Vision through Efficient Video-to-Voxel Simulation

Event-based cameras offer unique advantages such as high temporal resolution, high dynamic range, and low power consumption. However, the massive storage requirements and I/O burdens of existing synthetic data generation pipelines and the scarcity of real data prevent event-based training datasets from scaling up, limiting the development and generalization capabilities of event vision models. To address this challenge, we introduce Video-to-Voxel (V2V), an approach that directly converts conventional video frames into event-based voxel grid representations, bypassing the storage-intensive event stream generation entirely. V2V enables a 150 times reduction in storage requirements while supporting on-the-fly parameter randomization for enhanced model robustness. Leveraging this efficiency, we train several video reconstruction and optical flow estimation model architectures on 10,000 diverse videos totaling 52 hours--an order of magnitude larger than existing event datasets, yielding substantial improvements.
View on arXiv@article{lou2025_2505.16797, title={ V2V: Scaling Event-Based Vision through Efficient Video-to-Voxel Simulation }, author={ Hanyue Lou and Jinxiu Liang and Minggui Teng and Yi Wang and Boxin Shi }, journal={arXiv preprint arXiv:2505.16797}, year={ 2025 } }