ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.19743
474
0
v1v2v3 (latest)

Token-level Accept or Reject: A Micro Alignment Approach for Large Language Models

International Joint Conference on Artificial Intelligence (IJCAI), 2025
26 May 2025
Y. Zhang
Yu Yu
Bo Tang
Yu Zhu
Chuxiong Sun
Wenqiang Wei
Jie Hu
Zipeng Xie
Zhiyu Li
Feiyu Xiong
Edward Chung
ArXiv (abs)PDFHTML
Main:6 Pages
7 Figures
Bibliography:4 Pages
11 Tables
Appendix:6 Pages
Abstract

With the rapid development of Large Language Models (LLMs), aligning these models with human preferences and values is critical to ensuring ethical and safe applications. However, existing alignment techniques such as RLHF or DPO often require direct fine-tuning on LLMs with billions of parameters, resulting in substantial computational costs and inefficiencies. To address this, we propose Micro token-level Accept-Reject Aligning (MARA) approach designed to operate independently of the language models. MARA simplifies the alignment process by decomposing sentence-level preference learning into token-level binary classification, where a compact three-layer fully-connected network determines whether candidate tokens are "Accepted" or "Rejected" as part of the response. Extensive experiments across seven different LLMs and three open-source datasets show that MARA achieves significant improvements in alignment performance while reducing computational costs.

View on arXiv
Comments on this paper