134
4
v1v2 (latest)

Foundations of Large Language Models

Main:6 Pages
66 Figures
8 Tables
Appendix:271 Pages
Abstract

This is a book about large language models. As indicated by the title, it primarily focuses on foundational concepts rather than comprehensive coverage of all cutting-edge technologies. The book is structured into five main chapters, each exploring a key area: pre-training, generative models, prompting, alignment, and inference. It is intended for college students, professionals, and practitioners in natural language processing and related fields, and can serve as a reference for anyone interested in large language models.

View on arXiv
@article{xiao2025_2501.09223,
  title={ Foundations of Large Language Models },
  author={ Tong Xiao and Jingbo Zhu },
  journal={arXiv preprint arXiv:2501.09223},
  year={ 2025 }
}
Comments on this paper