LLM Stability: A detailed analysis with some surprises
Main:8 Pages
12 Figures
Bibliography:4 Pages
4 Tables
Appendix:3 Pages
Abstract
A concerning property of our nearly magical LLMs involves the variation of results given the exact same input and deterministic hyper-parameters. While AI has always had a certain level of noisiness from inputs outside of training data, we have generally had deterministic results for any particular input; that is no longer true. While most LLM practitioners are "in the know", we are unaware of any work that attempts to quantify current LLM stability. We suspect no one has taken the trouble because it is just too boring a paper to execute and write. But we have done it and there are some surprises.
View on arXivComments on this paper
