SMITE: Segment Me In TimE
International Conference on Learning Representations (ICLR), 2024
- VLMVOS
Main:10 Pages
14 Figures
Bibliography:7 Pages
12 Tables
Appendix:9 Pages
Abstract
Segmenting an object in a video presents significant challenges. Each pixel must be accurately labelled, and these labels must remain consistent across frames. The difficulty increases when the segmentation is with arbitrary granularity, meaning the number of segments can vary arbitrarily, and masks are defined based on only one or a few sample images. In this paper, we address this issue by employing a pre-trained text to image diffusion model supplemented with an additional tracking mechanism. We demonstrate that our approach can effectively manage various segmentation scenarios and outperforms state-of-the-art alternatives.
View on arXivComments on this paper
