ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2411.08968
  4. Cited By
Sparse Upcycling: Inference Inefficient Finetuning

Sparse Upcycling: Inference Inefficient Finetuning

13 November 2024
Sasha Doubov
Nikhil Sardana
Vitaliy Chiley
    MoE
ArXiv (abs)PDFHTML

Papers citing "Sparse Upcycling: Inference Inefficient Finetuning"

2 / 2 papers shown
Title
Scaling Fine-Grained MoE Beyond 50B Parameters: Empirical Evaluation and Practical Insights
Scaling Fine-Grained MoE Beyond 50B Parameters: Empirical Evaluation and Practical Insights
Jakub Krajewski
Marcin Chochowski
Daniel Korzekwa
MoEALM
178
0
0
03 Jun 2025
Beyond Chinchilla-Optimal: Accounting for Inference in Language Model Scaling Laws
Beyond Chinchilla-Optimal: Accounting for Inference in Language Model Scaling LawsInternational Conference on Machine Learning (ICML), 2023
Nikhil Sardana
Jacob P. Portes
Sasha Doubov
Jonathan Frankle
LRM
867
120
0
31 Dec 2023
1