Training deep quantum neural networks

Research output: Contribution to journalArticleResearchpeer review

Authors

  • Kerstin Beer
  • Dmytro Bondarenko
  • Terry Farrelly
  • Tobias J. Osborne
  • Robert Salzmann
  • Daniel Scheiermann
  • Ramona Wolf

External Research Organisations

  • University of Queensland
  • University of Cambridge
View graph of relations

Details

Original languageEnglish
Article number808
Pages (from-to)808
JournalNature Communications
Volume11
Issue number1
Publication statusPublished - 10 Feb 2020

Abstract

Neural networks enjoy widespread success in both research and industry and, with the advent of quantum technology, it is a crucial challenge to design quantum neural networks for fully quantum learning tasks. Here we propose a truly quantum analogue of classical neurons, which form quantum feedforward neural networks capable of universal quantum computation. We describe the efficient training of these networks using the fidelity as a cost function, providing both classical and efficient quantum implementations. Our method allows for fast optimisation with reduced memory requirements: the number of qudits required scales with only the width, allowing deep-network optimisation. We benchmark our proposal for the quantum task of learning an unknown unitary and find remarkable generalisation behaviour and a striking robustness to noisy training data.

Cite this

Training deep quantum neural networks. / Beer, Kerstin; Bondarenko, Dmytro; Farrelly, Terry et al.
In: Nature Communications, Vol. 11, No. 1, 808, 10.02.2020, p. 808.

Research output: Contribution to journalArticleResearchpeer review

Beer, K, Bondarenko, D, Farrelly, T, Osborne, TJ, Salzmann, R, Scheiermann, D & Wolf, R 2020, 'Training deep quantum neural networks', Nature Communications, vol. 11, no. 1, 808, pp. 808. https://doi.org/10.1038/s41467-020-14454-2, https://doi.org/10.15488/9906
Beer, K., Bondarenko, D., Farrelly, T., Osborne, T. J., Salzmann, R., Scheiermann, D., & Wolf, R. (2020). Training deep quantum neural networks. Nature Communications, 11(1), 808. Article 808. https://doi.org/10.1038/s41467-020-14454-2, https://doi.org/10.15488/9906
Beer K, Bondarenko D, Farrelly T, Osborne TJ, Salzmann R, Scheiermann D et al. Training deep quantum neural networks. Nature Communications. 2020 Feb 10;11(1):808. 808. doi: 10.1038/s41467-020-14454-2, 10.15488/9906
Beer, Kerstin ; Bondarenko, Dmytro ; Farrelly, Terry et al. / Training deep quantum neural networks. In: Nature Communications. 2020 ; Vol. 11, No. 1. pp. 808.
Download
@article{49cfaf8b35044a4dbcdb09c2d4d441a5,
title = "Training deep quantum neural networks",
abstract = "Neural networks enjoy widespread success in both research and industry and, with the advent of quantum technology, it is a crucial challenge to design quantum neural networks for fully quantum learning tasks. Here we propose a truly quantum analogue of classical neurons, which form quantum feedforward neural networks capable of universal quantum computation. We describe the efficient training of these networks using the fidelity as a cost function, providing both classical and efficient quantum implementations. Our method allows for fast optimisation with reduced memory requirements: the number of qudits required scales with only the width, allowing deep-network optimisation. We benchmark our proposal for the quantum task of learning an unknown unitary and find remarkable generalisation behaviour and a striking robustness to noisy training data.",
author = "Kerstin Beer and Dmytro Bondarenko and Terry Farrelly and Osborne, {Tobias J.} and Robert Salzmann and Daniel Scheiermann and Ramona Wolf",
note = "Funding information: Helpful correspondence and discussions with Lorenzo Cardarelli, Polina Feldmann, Andrew Green, Alexander Hahn, Amit Jamadagni, Maria Kalabakov, Sebastian Kinne-wig, Roger Melko, Laura Niermann, Simone Pfau, Marvin Schwiering, Deniz E. Stiege-mann and E. Miles Stoudenmire are gratefully acknowledged. This work was supported by the DFG through SFB 1227 (DQ-mat), the RTG 1991, and Quantum Frontiers. T.F. was supported by the Australian Research Council Centres of Excellence for Engineered Quantum Systems (EQUS, CE170100009). The publication of this article was funded by the Open Access Fund of the Leibniz Universit{\"a}t Hannover.",
year = "2020",
month = feb,
day = "10",
doi = "10.1038/s41467-020-14454-2",
language = "English",
volume = "11",
pages = "808",
journal = "Nature Communications",
issn = "2041-1723",
publisher = "Nature Publishing Group",
number = "1",

}

Download

TY - JOUR

T1 - Training deep quantum neural networks

AU - Beer, Kerstin

AU - Bondarenko, Dmytro

AU - Farrelly, Terry

AU - Osborne, Tobias J.

AU - Salzmann, Robert

AU - Scheiermann, Daniel

AU - Wolf, Ramona

N1 - Funding information: Helpful correspondence and discussions with Lorenzo Cardarelli, Polina Feldmann, Andrew Green, Alexander Hahn, Amit Jamadagni, Maria Kalabakov, Sebastian Kinne-wig, Roger Melko, Laura Niermann, Simone Pfau, Marvin Schwiering, Deniz E. Stiege-mann and E. Miles Stoudenmire are gratefully acknowledged. This work was supported by the DFG through SFB 1227 (DQ-mat), the RTG 1991, and Quantum Frontiers. T.F. was supported by the Australian Research Council Centres of Excellence for Engineered Quantum Systems (EQUS, CE170100009). The publication of this article was funded by the Open Access Fund of the Leibniz Universität Hannover.

PY - 2020/2/10

Y1 - 2020/2/10

N2 - Neural networks enjoy widespread success in both research and industry and, with the advent of quantum technology, it is a crucial challenge to design quantum neural networks for fully quantum learning tasks. Here we propose a truly quantum analogue of classical neurons, which form quantum feedforward neural networks capable of universal quantum computation. We describe the efficient training of these networks using the fidelity as a cost function, providing both classical and efficient quantum implementations. Our method allows for fast optimisation with reduced memory requirements: the number of qudits required scales with only the width, allowing deep-network optimisation. We benchmark our proposal for the quantum task of learning an unknown unitary and find remarkable generalisation behaviour and a striking robustness to noisy training data.

AB - Neural networks enjoy widespread success in both research and industry and, with the advent of quantum technology, it is a crucial challenge to design quantum neural networks for fully quantum learning tasks. Here we propose a truly quantum analogue of classical neurons, which form quantum feedforward neural networks capable of universal quantum computation. We describe the efficient training of these networks using the fidelity as a cost function, providing both classical and efficient quantum implementations. Our method allows for fast optimisation with reduced memory requirements: the number of qudits required scales with only the width, allowing deep-network optimisation. We benchmark our proposal for the quantum task of learning an unknown unitary and find remarkable generalisation behaviour and a striking robustness to noisy training data.

UR - http://www.scopus.com/inward/record.url?scp=85079230673&partnerID=8YFLogxK

U2 - 10.1038/s41467-020-14454-2

DO - 10.1038/s41467-020-14454-2

M3 - Article

C2 - 32041956

AN - SCOPUS:85079230673

VL - 11

SP - 808

JO - Nature Communications

JF - Nature Communications

SN - 2041-1723

IS - 1

M1 - 808

ER -