Mode-Assisted Unsupervised Learning of Restricted Boltzmann Machines
ORAL
Abstract
Restricted Boltzmann Machines (RBMs) are a powerful class of unsupervised models well known in machine learning. However, unlike their more popular supervised counterparts, their training requires computing a gradient that is notoriously difficult even to approximate. In this work we show that properly combining standard gradient approximations with an off-gradient direction, constructed from samples of the RBM ground state (mode), improves their training dramatically over the standard methods. This approach, which we call "mode training", promotes faster training and stability, in addition to lowering the converged relative entropy (KL divergence). We report promising preliminary results with small models on synthetic data sets and discuss extensions to more realistic scenarios, where a physics-based approach, memcomputing [1], is used
to sample the mode efficiently. The mode training we suggest is versatile, as it can be applied with any given gradient method, and is easily extended to more general energy-based neural network structures such as deep, convolutional and unrestricted Boltzmann machines. [1] M. Di Ventra and F.L. Traversa, J. Appl. Phys. 123, 180901 (2018). Work supported in part by CMRR and DARPA.
to sample the mode efficiently. The mode training we suggest is versatile, as it can be applied with any given gradient method, and is easily extended to more general energy-based neural network structures such as deep, convolutional and unrestricted Boltzmann machines. [1] M. Di Ventra and F.L. Traversa, J. Appl. Phys. 123, 180901 (2018). Work supported in part by CMRR and DARPA.
–
Presenters
-
Haik Manukian
University of California, San Diego
Authors
-
Haik Manukian
University of California, San Diego
-
Yan Ru Pei
University of California, San Diego
-
Sean Bearden
University of California, San Diego
-
Massimiliano Di Ventra
University of California, San Diego