Geometric Reasoning in the Embedding Space

In this contribution, we demonstrate that Graph Neural Networks and Transformers can learn to reason about geometric constraints. We train them to predict spatial position of points in a discrete 2D grid from a set of constraints that uniquely describe hidden figures containing these points. Both models are able to predict the position of points and interestingly, they form the hidden figures described by the input constraints in the embedding space during the reasoning process. Our analysis shows that both models recover the grid structure during training so that the embeddings corresponding to the points within the grid organize themselves in a 2D subspace and reflect the neighborhood structure of the grid. We also show that the Graph Neural Network we design for the task performs significantly better than the Transformer and is also easier to scale.
View on arXiv@article{hůla2025_2504.02018, title={ Geometric Reasoning in the Embedding Space }, author={ Jan Hůla and David Mojžíšek and Jiří Janeček and David Herel and Mikoláš Janota }, journal={arXiv preprint arXiv:2504.02018}, year={ 2025 } }