Abstract
Quantum neural networks (QNNs) have generated excitement around the possibility of efficiently analyzing quantum data. But this excitement has been tempered by the existence of exponentially vanishing gradients, known as barren plateau landscapes, for many QNN architectures. Recently, quantum convolutional neural networks (QCNNs) have been proposed, involving a sequence of convolutional and pooling layers that reduce the number of qubits while preserving information about relevant data features. In this work, we rigorously analyze the gradient scaling for the parameters in the QCNN architecture. We find that the variance of the gradient vanishes no faster than polynomially, implying that QCNNs do not exhibit barren plateaus. This result provides an analytical guarantee for the trainability of randomly initialized QCNNs, which highlights QCNNs as being trainable under random initialization unlike many other QNN architectures. To derive our results, we introduce a novel graph-based method to analyze expectation values over Haar-distributed unitaries, which will likely be useful in other contexts. Finally, we perform numerical simulations to verify our analytical results.
9 More- Received 12 March 2021
- Revised 13 July 2021
- Accepted 2 August 2021
DOI:https://doi.org/10.1103/PhysRevX.11.041011
Published by the American Physical Society under the terms of the Creative Commons Attribution 4.0 International license. Further distribution of this work must maintain attribution to the author(s) and the published article’s title, journal citation, and DOI.
Published by the American Physical Society
Physics Subject Headings (PhySH)
Popular Summary
With the advent of quantum computers, many different architectures have been proposed that could provide an advantage over their classical counterparts. Quantum neural networks (QNNs) are among the most promising architectures, with applications including physical simulations, optimization, and more general machine-learning tasks. Despite their tremendous potential, QNNs have been shown to exhibit a “barren plateau,” where the gradient of a cost function vanishes exponentially with system size, rendering the architecture untrainable for large problem sizes. Here, we show that one particular QNN architecture does not have this plateau.
We analyze an architecture called the quantum convolutional neural network (QCNN), which has been recently proposed to tackle classification problems on quantum data. As an example, a QCNN could be trained to classify quantum states according to the phase of matter that they belong to. We rigorously prove that QCNNs do not suffer from barren plateaus, thus highlighting them as a potential candidate architecture for achieving quantum advantage in the near term.
As the field of QNNs booms, we believe it will be important to carry out similar analyses for other candidate architectures, and the techniques developed in our work can be used as blueprints for such analyses.