ID: 2112.09420

A random energy approach to deep learning

December 17, 2021

View on ArXiv
Rongrong Xie, Matteo Marsili
Condensed Matter
Computer Science
Statistics
Disordered Systems and Neura...
Machine Learning
Machine Learning

We study a generic ensemble of deep belief networks which is parametrized by the distribution of energy levels of the hidden states of each layer. We show that, within a random energy approach, statistical dependence can propagate from the visible to deep layers only if each layer is tuned close to the critical point during learning. As a consequence, efficiently trained learning machines are characterised by a broad distribution of energy levels. The analysis of Deep Belief Networks and Restricted Boltzmann Machines on different datasets confirms these conclusions.

Similar papers 1