Abstract
Many supervised learning tasks have intrinsic symmetries, such as translational and rotational symmetry in image classifications. These symmetries can be exploited to enhance performance. We formulate the symmetry constraints into a concise mathematical form. We design two ways to adopt the constraints into the cost function, thereby shaping the cost landscape in favor of parameter choices, which respect the given symmetry. Unlike methods that alter the neural network circuit Ansatz to impose symmetry, our method only changes the classical postprocessing of gradient descent, which is simpler to implement. We call the method symmetry-guided gradient descent (SGGD). We illustrate SGGD in entanglement classification of Werner states and in two classification tasks in a two-dimensional feature space. In both cases, the results show that SGGD can accelerate the training, improve the generalization ability, and remove vanishing gradients, especially when the training data is biased. © 2024 American Physical Society.
Original language | English |
---|---|
Article number | 022406 |
Journal | Physical Review A |
Volume | 110 |
Issue number | 2 |
Online published | 5 Aug 2024 |
DOIs | |
Publication status | Published - Aug 2024 |
Funding
We acknowledge inspiring discussions with Dong Yang and Jin-Long Huang. We acknowledge support from HiSilicon, the National Natural Science Foundation of China (Grants No. 12050410246, No. 1200509, No. 12050410245), and the City University of Hong Kong (Project No. 9610623).
Publisher's Copyright Statement
- COPYRIGHT TERMS OF DEPOSITED FINAL PUBLISHED VERSION FILE: Bian, K., Zhang, S., Meng, F., Zhang, W., & Dahlsten, O. (2024). Symmetry-guided gradient descent for quantum neural networks. Physical Review A, 110(2), Article 022406. https://doi.org/10.1103/PhysRevA.110.022406 The copyright of this article is owned by American Physical Society.