ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2508.09125
54
1

Complex Logical Instruction Generation

12 August 2025
Mian Zhang
Shujian Liu
Sixun Dong
Ming Yin
Yebowen Hu
Xun Wang
Steven Ma
Song Wang
Sathish Indurthi
Haoyun Deng
Zhiyu Zoey Chen
Kaiqiang Song
    LRM
ArXiv (abs)PDFHTMLHuggingFace (38 upvotes)Github (17★)
Main:10 Pages
10 Figures
Bibliography:2 Pages
1 Tables
Appendix:31 Pages
Abstract

Instruction following has catalyzed the recent era of Large Language Models (LLMs) and is the foundational skill underpinning more advanced capabilities such as reasoning and agentic behaviors. As tasks grow more challenging, the logic structures embedded in natural language instructions becomes increasingly intricate. However, how well LLMs perform on such logic-rich instructions remains under-explored. We propose LogicIFGen and LogicIFEval. LogicIFGen is a scalable, automated framework for generating verifiable instructions from code functions, which can naturally express rich logic such as conditionals, nesting, recursion, and function calls. We further curate a collection of complex code functions and use LogicIFGen to construct LogicIFEval, a benchmark comprising 426 verifiable logic-rich instructions. Our experiments demonstrate that current state-of-the-art LLMs still struggle to correctly follow the instructions in LogicIFEval. Most LLMs can only follow fewer than 60% of the instructions, revealing significant deficiencies in the instruction-following ability. Code and Benchmark:this https URL

View on arXiv
Comments on this paper