7

LIEREx: Language-Image Embeddings for Robotic Exploration

Felix Igelbrink
Lennart Niecksch
Marian Renz
Martin Günther
Martin Atzmueller
Main:6 Pages
4 Figures
Bibliography:1 Pages
Appendix:2 Pages
Abstract

Semantic maps allow a robot to reason about its surroundings to fulfill tasks such as navigating known environments, finding specific objects, and exploring unmapped areas. Traditional mapping approaches provide accurate geometric representations but are often constrained by pre-designed symbolic vocabularies. The reliance on fixed object classes makes it impractical to handle out-of-distribution knowledge not defined at design time. Recent advances in Vision-Language Foundation Models, such as CLIP, enable open-set mapping, where objects are encoded as high-dimensional embeddings rather than fixed labels. In LIEREx, we integrate these VLFMs with established 3D Semantic Scene Graphs to enable target-directed exploration by an autonomous agent in partially unknown environments.

View on arXiv
Comments on this paper