ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2406.14115
  4. Cited By
Take the essence and discard the dross: A Rethinking on Data Selection for Fine-Tuning Large Language Models

Take the essence and discard the dross: A Rethinking on Data Selection for Fine-Tuning Large Language Models

20 June 2024
Ziche Liu
Rui Ke
Feng Jiang
Feng Jiang
Haizhou Li
ArXivPDFHTML

Papers citing "Take the essence and discard the dross: A Rethinking on Data Selection for Fine-Tuning Large Language Models"

3 / 3 papers shown
Title
Ensembles of Low-Rank Expert Adapters
Ensembles of Low-Rank Expert Adapters
Yinghao Li
Vianne Gao
Chao Zhang
MohamadAli Torkamani
50
0
0
31 Jan 2025
LESS: Selecting Influential Data for Targeted Instruction Tuning
LESS: Selecting Influential Data for Targeted Instruction Tuning
Mengzhou Xia
Sadhika Malladi
Suchin Gururangan
Sanjeev Arora
Danqi Chen
68
180
0
06 Feb 2024
Scaling Laws for Neural Language Models
Scaling Laws for Neural Language Models
Jared Kaplan
Sam McCandlish
T. Henighan
Tom B. Brown
B. Chess
R. Child
Scott Gray
Alec Radford
Jeff Wu
Dario Amodei
220
3,054
0
23 Jan 2020
1