65
64

In-Context Learning with Long-Context Models: An In-Depth Exploration

Abstract

As model context lengths continue to increase, the number of demonstrations that can be provided in-context approaches the size of entire training datasets. We study the behavior of in-context learning (ICL) at this extreme scale on multiple datasets and models. We show that, for many datasets with large label spaces, performance continues to increase with thousands of demonstrations. We contrast this with example retrieval and finetuning: example retrieval shows excellent performance at low context lengths but has diminished gains with more demonstrations; finetuning is more data hungry than ICL but can exceed long-context ICL performance with additional data. We use the ICL setting to study several properties of both in-context learning and long-context models. We show that long-context ICL is less sensitive to random input shuffling than short-context ICL, that grouping of same-label examples negatively impacts performance, and that the performance boosts do not arise from cumulative gain from encoding many examples together. We conclude that long-context ICL can be an effective tool, and may not require long-context for encoding the demonstration set at all.

View on arXiv
@article{bertsch2025_2405.00200,
  title={ In-Context Learning with Long-Context Models: An In-Depth Exploration },
  author={ Amanda Bertsch and Maor Ivgi and Emily Xiao and Uri Alon and Jonathan Berant and Matthew R. Gormley and Graham Neubig },
  journal={arXiv preprint arXiv:2405.00200},
  year={ 2025 }
}
Comments on this paper