Physical neural networks trained with a physics-aware backpropagation algorithm

ORAL · Invited

Abstract

Deep neural networks approximate mathematical functions by learning the parameters of a sequence of nonlinear functions, or layers. The algorithm of choice performing this learning is the backpropagation algorithm for gradient descent. Backpropagation-based learning applied to large, deep neural networks - deep learning - has led to truly stunning progress in the last decade or so, and is the basis for countless popular techniques in modern science and industry.

We have taken a similar approach to build computations from networks of controllable physical systems - physical neural networks. The underlying idea is that the evolution of a physical system inherently performs a computation on its initial conditions, and this transformation can be adjusted by tuning degrees of freedom of that system. By combining multiple such physical transformations, and applying a backpropagation algorithm to learn their physical parameters, we can learn physical functions. Using a version of backpropagation suited to experimental physical systems, we have demonstrated physical neural networks based on nonlinear optics, analog electronics, mechanics, and coupled oscillators [Wright, Onodera et al., Nature (2022)].

In this talk, I will review this and our recent work on physical neural networks, primarily with optical systems. These include:

- The limits of optical neural networks due to quantum noise [Ma et al., arXiv:2307.15712, Wang et al., Nature Comm (2022)]

- Physical neural networks as smart sensors [Wang, Sohoni et al., Nature Photonics (2023)]

- The potential of large-scale neural networks to implement Transformer models energy-efficiently [Anderson et al., arXiv:2302.10360]

- Implementations of physical neural networks with multimode optical waves, and with systems of coupled nonlinear oscillators [Onodera, Stein et al., in prep]

* We wish to thank NTT Research for their financial and technical support. Portions of this work were supported by the National Science Foundation (award no. CCF-1918549), a Kavli Institute at Cornell instrumentation grant, and a David and Lucile Packard Foundation Fellowship. One of us (Peter McMahon) acknowledges membership of the CIFAR Quantum Information Science Program as an Azrieli Global Scholar.

Publication: S.-Y. Ma, T. Wang, J. Laydevant, L.G. Wright, and P.L. McMahon "Quantum-noise-limited
optical neural networks operating at a few quanta per activation", arXiv: 2302.10360

M.G. Anderson, S.-Y. Ma, T. Wang, L.G. Wright, and P.L. McMahon "Optical Transformers",
arXiv: 2302.10360

L.G. Wright†, T. Onodera†, M.M. Stein, T. Wang, D.T. Schachter, Z. Hu, and P.L. McMahon
(2022) "Deep physical neural networks trained with backpropagation", Nature 601, 549–555.

T. Wang†, M.M. Sohoni†, L.G. Wright, M.M. Stein, S.-Y. Ma, T. Onodera, M.G. Anderson,
P.L. McMahon (2023) "Image sensing with multilayer, nonlinear optical neural networks",
Nature Photonics 17, 408–415.

T. Wang, S. Ma, L.G. Wright, T. Onodera, B. Richards, and P. L. McMahon (2022) "A
photonic neural network using less than 1 photon per weight multiplication," Nature
Communications 13, 123.

Other manuscripts in progress

Presenters

  • Logan G Wright

    Cornell University

Authors

  • Logan G Wright

    Cornell University