Baby Llama: knowledge distillation from an ensemble of teachers trained
  on a small dataset with no performance penalty
v1v2 (latest)

Baby Llama: knowledge distillation from an ensemble of teachers trained on a small dataset with no performance penalty

Papers citing "Baby Llama: knowledge distillation from an ensemble of teachers trained on a small dataset with no performance penalty"

43 / 43 papers shown
Title
GREEN-CODE: Learning to Optimize Energy Efficiency in LLM-based Code Generation
GREEN-CODE: Learning to Optimize Energy Efficiency in LLM-based Code GenerationIEEE/ACM International Symposium on Cluster, Cloud and Internet Computing (CCGrid), 2025
144
2
0
19 Jan 2025

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.