24
22

Generalization in Quantum Machine Learning: a Quantum Information Perspective

Abstract

Quantum classification and hypothesis testing are two tightly related subjects, the main difference being that the former is data driven: how to assign to quantum states ρ(x)\rho(x) the corresponding class cc (or hypothesis) is learnt from examples during training, where xx can be either tunable experimental parameters or classical data "embedded" into quantum states. Does the model generalize? This is the main question in any data-driven strategy, namely the ability to predict the correct class even of previously unseen states. Here we establish a link between quantum machine learning classification and quantum hypothesis testing (state and channel discrimination) and then show that the accuracy and generalization capability of quantum classifiers depend on the (R\ényi) mutual informations I(C:Q)I(C{:}Q) and I2(X:Q)I_2(X{:}Q) between the quantum state space QQ and the classical parameter space XX or class space CC. Based on the above characterization, we then show how different properties of QQ affect classification accuracy and generalization, such as the dimension of the Hilbert space, the amount of noise, and the amount of neglected information from XX via, e.g., pooling layers. Moreover, we introduce a quantum version of the Information Bottleneck principle that allows us to explore the various tradeoffs between accuracy and generalization. Finally, in order to check our theoretical predictions, we study the classification of the quantum phases of an Ising spin chain, and we propose the Variational Quantum Information Bottleneck (VQIB) method to optimize quantum embeddings of classical data to favor generalization.

View on arXiv
Comments on this paper