ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2304.07438
14
38

Tractable Control for Autoregressive Language Generation

15 April 2023
Honghua Zhang
Meihua Dang
Nanyun Peng
Guy Van den Broeck
    BDL
ArXivPDFHTML
Abstract

Despite the success of autoregressive large language models in text generation, it remains a major challenge to generate text that satisfies complex constraints: sampling from the conditional distribution Pr⁡(text∣α){\Pr}(\text{text} | \alpha)Pr(text∣α) is intractable for even the simplest lexical constraints α\alphaα. To overcome this challenge, we propose to use tractable probabilistic models (TPMs) to impose lexical constraints in autoregressive text generation models, which we refer to as GeLaTo (Generating Language with Tractable Constraints). To demonstrate the effectiveness of this framework, we use distilled hidden Markov models, where we can efficiently compute Pr⁡(text∣α){\Pr}(\text{text} | \alpha)Pr(text∣α), to guide autoregressive generation from GPT2. GeLaTo achieves state-of-the-art performance on challenging benchmarks for constrained text generation (e.g., CommonGen), beating various strong baselines by a large margin. Our work not only opens up new avenues for controlling large language models but also motivates the development of more expressive TPMs.

View on arXiv
Comments on this paper