ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2410.13147
16
0

Utilizing Large Language Models in an iterative paradigm with domain feedback for molecule optimization

17 October 2024
Khiem Le
Nitesh V. Chawla
ArXivPDFHTML
Abstract

Molecule optimization is a critical task in drug discovery to optimize desired properties of a given molecule through chemical modification. Despite Large Language Models (LLMs) holding the potential to efficiently simulate this task by using natural language to direct the optimization, straightforwardly utilizing them shows limited performance. In this work, we facilitate utilizing LLMs in an iterative paradigm by proposing a simple yet highly effective domain feedback provider, namely Re3\text{Re}^3Re3DF. In detail, Re3\text{Re}^3Re3DF harnesses an external toolkit, RDKit, to handle the molecule hallucination, if the modified molecule is chemically invalid. Otherwise, its desired properties are computed and compared to the original one, establishing reliable domain feedback with correct direction and distance towards the objective, followed by a retrieved example, to guide the LLM to refine the modified molecule. We conduct experiments across both single- and multi-property objectives with 2 thresholds, where Re3\text{Re}^3Re3DF shows significant improvements. Particularly, for 20 single-property objectives, Re3\text{Re}^3Re3DF enhances Hit ratio by 16.95% and 20.76% under loose (\texttt{l}) and strict (\texttt{s}) thresholds, respectively. For 32 multi-property objectives, Re3\text{Re}^3Re3DF enhances Hit ratio by 6.04% and 5.25%.

View on arXiv
Comments on this paper