v1v2 (latest)
Foundations of Large Language Models
- 3DGSAILawVLM

Main:6 Pages
66 Figures
8 Tables
Appendix:271 Pages
Abstract
This is a book about large language models. As indicated by the title, it primarily focuses on foundational concepts rather than comprehensive coverage of all cutting-edge technologies. The book is structured into five main chapters, each exploring a key area: pre-training, generative models, prompting, alignment, and inference. It is intended for college students, professionals, and practitioners in natural language processing and related fields, and can serve as a reference for anyone interested in large language models.
View on arXiv@article{xiao2025_2501.09223, title={ Foundations of Large Language Models }, author={ Tong Xiao and Jingbo Zhu }, journal={arXiv preprint arXiv:2501.09223}, year={ 2025 } }
Comments on this paper