Abstract
In this Letter, we propose a guiding principle for how to design the architecture of a quantum neural network in order to achieve a high learning efficiency. This principle is inspired by the equivalence between extracting information from the input state to the readout qubit and scrambling information from the readout qubit to input qubits. We characterize the quantum information scrambling by operator size growth. By Haar random averaging over operator sizes, we propose an averaged operator size to describe the information scrambling ability of a given quantum neural network architecture. The key conjecture of this Letter is that this quantity is positively correlated with the learning efficiency of this architecture. To support this conjecture, we consider several different architectures, and we also consider two typical learning tasks. One is a regression task of a quantum problem, and the other is a classification task on classical images. In both cases, we find that, for the architecture with a larger averaged operator size, the loss function decreases faster or the prediction accuracy increases faster as the training epoch increases, which means higher learning efficiency. Our results can be generalized to more complicated quantum versions of machine learning algorithms.
- Received 30 November 2020
- Revised 1 March 2021
- Accepted 23 August 2021
DOI:https://doi.org/10.1103/PhysRevResearch.3.L032057
Published by the American Physical Society under the terms of the Creative Commons Attribution 4.0 International license. Further distribution of this work must maintain attribution to the author(s) and the published article's title, journal citation, and DOI.
Published by the American Physical Society