• Open Access

Generalization in Quantum Machine Learning: A Quantum Information Standpoint

Leonardo Banchi, Jason Pereira, and Stefano Pirandola
PRX Quantum 2, 040321 – Published 1 November 2021

Abstract

Quantum classification and hypothesis testing (state and channel discrimination) are two tightly related subjects, the main difference being that the former is data driven: how to assign to quantum states ρ(x) the corresponding class c (or hypothesis) is learnt from examples during training, where x can be either tunable experimental parameters or classical data “embedded” into quantum states. Does the model generalize? This is the main question in any data-driven strategy, namely the ability to predict the correct class even of previously unseen states. Here we establish a link between quantum classification and quantum information theory, by showing that the accuracy and generalization capability of quantum classifiers depend on the (Rényi) mutual information I(C:Q) and I2(X:Q) between the quantum state space Q and the classical parameter space X or class space C. Based on the above characterization, we then show how different properties of Q affect classification accuracy and generalization, such as the dimension of the Hilbert space, the amount of noise, and the amount of neglected information from X via, e.g., pooling layers. Moreover, we introduce a quantum version of the information bottleneck principle that allows us to explore the various trade-offs between accuracy and generalization. Finally, in order to check our theoretical predictions, we study the classification of the quantum phases of an Ising spin chain, and we propose the variational quantum information bottleneck method to optimize quantum embeddings of classical data to favor generalization.

  • Figure
  • Figure
  • Figure
  • Figure
  • Figure
  • Figure
  • Figure
1 More
  • Received 4 March 2021
  • Accepted 30 September 2021

DOI:https://doi.org/10.1103/PRXQuantum.2.040321

Published by the American Physical Society under the terms of the Creative Commons Attribution 4.0 International license. Further distribution of this work must maintain attribution to the author(s) and the published article's title, journal citation, and DOI.

Published by the American Physical Society

Physics Subject Headings (PhySH)

Quantum Information, Science & TechnologyAtomic, Molecular & OpticalStatistical Physics & ThermodynamicsInterdisciplinary Physics

Authors & Affiliations

Leonardo Banchi1,2,*, Jason Pereira3,1, and Stefano Pirandola3

  • 1Department of Physics and Astronomy, University of Florence, Via Giovanni Sansone 1, Sesto Fiorentino (FI) I-50019 , Italy
  • 2INFN Sezione di Firenze, Via Giovanni Sansone 1, Sesto Fiorentino (FI) I-50019, Italy
  • 3Department of Computer Science, University of York, York YO10 5GH, United Kingdom

  • *leonardo.banchi@unifi.it

Popular Summary

Data-driven strategies are responsible for the tremendous advances in many fields such as pattern recognition, where the classification algorithm is directly ”learnt” from data rather than being built via human-designed mathematical models. Quantum machine learning is a rapidly evolving new discipline that applies data-driven strategies to the quantum realm, where faster algorithms and more accurate estimation methods are theoretically available. Nonetheless, much theoretical work is still needed to understand how to fully exploit genuinely quantum features, such as entanglement, the no-cloning theorem, and the destructive role of quantum measurements, for machine learning applications.

Here we use tools from quantum information theory to study the problem of generalization in quantum data classification, where the data are either already given in a quantum form or first embedded onto quantum states. We show that after ”training” a quantum classifier using a few examples, the classification error that we get with new, previously unseen data can be bounded by entropic quantities, more specifically by the mutual information between the quantum state space and other classical spaces. Loosely speaking, our theorem shows that a quantum classifier generalizes well when it is capable of discarding as much as possible of the information that is unnecessary to predict the class.

We apply our general results to study fundamental properties, e.g., introducing a quantum version of the bias-variance trade-off, and to derive new algorithms, such as what we call the variational quantum information bottleneck principle. Our predictions are then tested in numerical experiments.

Key Image

Article Text

Click to Expand

References

Click to Expand
Issue

Vol. 2, Iss. 4 — November - December 2021

Reuse & Permissions
Author publication services for translation and copyediting assistance advertisement

Authorization Required


×
×

Images

×

Sign up to receive regular email alerts from PRX Quantum

Reuse & Permissions

It is not necessary to obtain permission to reuse this article or its components as it is available under the terms of the Creative Commons Attribution 4.0 International license. This license permits unrestricted use, distribution, and reproduction in any medium, provided attribution to the author(s) and the published article's title, journal citation, and DOI are maintained. Please note that some figures may have been included with permission from other third parties. It is your responsibility to obtain the proper permission from the rights holder directly for these figures.

×

Log In

Cancel
×

Search


Article Lookup

Paste a citation or DOI

Enter a citation
×