522
v1v2v3 (latest)

Scaling Embedding Layers in Language Models

Main:8 Pages
20 Figures
Bibliography:7 Pages
8 Tables
Appendix:12 Pages
Abstract

We propose SCONESCONE (SScalable, CContextualized, OOffloaded, NN-gram EEmbedding), a new method for extending input embedding layers to enhance language model performance. To avoid increased decoding costs, SCONESCONE retains the original vocabulary while introducing embeddings for a set of frequent n-grams. These embeddings provide contextualized representation for each input token and are learned with a separate model during training. After training, embeddings are precomputed and stored in off-accelerator memory; during inference, querying them has minimal impact on latency due to the low complexity of embedding lookups. SCONESCONE enables two new scaling strategies: increasing the number of n-gram embeddings and scaling the model used to learn them, both while maintaining fixed accelerator usage during inference (in terms of FLOPS and memory). We show that scaling both aspects enables a model with 1B accelerator-resident parameters to outperform a 1.9B-parameter baseline across diverse corpora, while using only about half the FLOPS and accelerator memory during inference.

View on arXiv
Comments on this paper