ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2405.10587
  4. Cited By
RDRec: Rationale Distillation for LLM-based Recommendation

RDRec: Rationale Distillation for LLM-based Recommendation

17 May 2024
Xinfeng Wang
Jin Cui
Yoshimi Suzuki
Fumiyo Fukumoto
    LRM
ArXivPDFHTML

Papers citing "RDRec: Rationale Distillation for LLM-based Recommendation"

4 / 4 papers shown
Title
LlamaRec: Two-Stage Recommendation using Large Language Models for
  Ranking
LlamaRec: Two-Stage Recommendation using Large Language Models for Ranking
Zhenrui Yue
Sara Rabhi
G. D. S. P. Moreira
Dong Wang
Even Oldridge
LRM
44
36
0
25 Oct 2023
EXPLAIN, EDIT, GENERATE: Rationale-Sensitive Counterfactual Data
  Augmentation for Multi-hop Fact Verification
EXPLAIN, EDIT, GENERATE: Rationale-Sensitive Counterfactual Data Augmentation for Multi-hop Fact Verification
Yingjie Zhu
Jiasheng Si
Yibo Zhao
Haiyang Zhu
Deyu Zhou
Yulan He
23
6
0
23 Oct 2023
Distilling Step-by-Step! Outperforming Larger Language Models with Less
  Training Data and Smaller Model Sizes
Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes
Lokesh Nagalapatti
Chun-Liang Li
Chih-Kuan Yeh
Hootan Nakhost
Yasuhisa Fujii
Alexander Ratner
Ranjay Krishna
Chen-Yu Lee
Tomas Pfister
ALM
204
498
0
03 May 2023
A Token-level Reference-free Hallucination Detection Benchmark for
  Free-form Text Generation
A Token-level Reference-free Hallucination Detection Benchmark for Free-form Text Generation
Tianyu Liu
Yizhe Zhang
Chris Brockett
Yi Mao
Zhifang Sui
Weizhu Chen
W. Dolan
HILM
217
140
0
18 Apr 2021
1