68
1

REALM-Bench: A Real-World Planning Benchmark for LLMs and Multi-Agent Systems

Abstract

This benchmark suite provides a comprehensive evaluation framework for assessing both individual LLMs and multi-agent systems in real-world planning scenarios. The suite encompasses eleven designed problems that progress from basic to highly complex, incorporating key aspects such as multi-agent coordination, inter-agent dependencies, and dynamic environmental disruptions. Each problem can be scaled along three dimensions: the number of parallel planning threads, the complexity of inter-dependencies, and the frequency of unexpected disruptions requiring real-time adaptation. The benchmark includes detailed specifications, evaluation metrics, and baseline implementations using contemporary frameworks like LangGraph, enabling rigorous testing of both single-agent and multi-agent planning capabilities. Through standardized evaluation criteria and scalable complexity, this benchmark aims to drive progress in developing more robust and adaptable AI planning systems for real-world applications.

View on arXiv
@article{geng2025_2502.18836,
  title={ REALM-Bench: A Real-World Planning Benchmark for LLMs and Multi-Agent Systems },
  author={ Longling Geng and Edward Y. Chang },
  journal={arXiv preprint arXiv:2502.18836},
  year={ 2025 }
}
Comments on this paper