ID: cond-mat/0104011

Multilayer neural networks with extensively many hidden units

April 1, 2001

View on ArXiv

Similar papers 5

A statistical mechanics framework for Bayesian deep neural networks beyond the infinite-width limit

September 11, 2022

83% Match
R. Pacelli, S. Ariosto, M. Pastore, F. Ginelli, ... , Rotondo P.
Disordered Systems and Neura...

Despite the practical success of deep neural networks, a comprehensive theoretical framework that can predict practically relevant scores, such as the test accuracy, from knowledge of the training data is currently lacking. Huge simplifications arise in the infinite-width limit, where the number of units $N_\ell$ in each hidden layer ($\ell=1,\dots, L$, being $L$ the depth of the network) far exceeds the number $P$ of training examples. This idealisation, however, blatantly d...

Find SimilarView on arXiv

Quantifying information stored in synaptic connections rather than in firing patterns of neural networks

November 26, 2024

83% Match
Xinhao Fan, Shreesh P Mysore
Neurons and Cognition
Information Theory
Information Theory
Biological Physics

A cornerstone of our understanding of both biological and artificial neural networks is that they store information in the strengths of connections among the constituent neurons. However, in contrast to the well-established theory for quantifying information encoded by the firing patterns of neural networks, little is known about quantifying information encoded by its synaptic connections. Here, we develop a theoretical framework using continuous Hopfield networks as an exemp...

Find SimilarView on arXiv

Storage properties of a quantum perceptron

November 16, 2021

83% Match
Katerina Aikaterini, Gratsea, ... , Lewenstein Maciej
Disordered Systems and Neura...

Driven by growing computational power and algorithmic developments, machine learning methods have become valuable tools for analyzing vast amounts of data. Simultaneously, the fast technological progress of quantum information processing suggests employing quantum hardware for machine learning purposes. Recent works discuss different architectures of quantum perceptrons, but the abilities of such quantum devices remain debated. Here, we investigate the storage capacity of a p...

Find SimilarView on arXiv

Activation function dependence of the storage capacity of treelike neural networks

July 22, 2020

83% Match
Jacob A. Zavatone-Veth, Cengiz Pehlevan
Disordered Systems and Neura...
Machine Learning
Machine Learning

The expressive power of artificial neural networks crucially depends on the nonlinearity of their activation functions. Though a wide variety of nonlinear activation functions have been proposed for use in artificial neural networks, a detailed understanding of their role in determining the expressive power of a network has not emerged. Here, we study how activation functions affect the storage capacity of treelike two-layer networks. We relate the boundedness or divergence o...

Find SimilarView on arXiv

Generalization in multitask deep neural classifiers: a statistical physics approach

October 30, 2019

83% Match
Tyler Lee, Anthony Ndirango
Machine Learning
Machine Learning

A proper understanding of the striking generalization abilities of deep neural networks presents an enduring puzzle. Recently, there has been a growing body of numerically-grounded theoretical work that has contributed important insights to the theory of learning in deep neural nets. There has also been a recent interest in extending these analyses to understanding how multitask learning can further improve the generalization capacity of deep neural nets. These studies deal a...

Find SimilarView on arXiv

Properties of the geometry of solutions and capacity of multi-layer neural networks with Rectified Linear Units activations

July 17, 2019

83% Match
Carlo Baldassi, Enrico M. Malatesta, Riccardo Zecchina
Disordered Systems and Neura...
Machine Learning
Machine Learning

Rectified Linear Units (ReLU) have become the main model for the neural units in current deep learning systems. This choice has been originally suggested as a way to compensate for the so called vanishing gradient problem which can undercut stochastic gradient descent (SGD) learning in networks composed of multiple layers. Here we provide analytical results on the effects of ReLUs on the capacity and on the geometrical landscape of the solution space in two-layer neural netwo...

Find SimilarView on arXiv

Beyond Disorder: Unveiling Cooperativeness in Multidirectional Associative Memories

March 6, 2025

83% Match
Andrea Alessandrelli, Adriano Barra, Andrea Ladiana, ... , Ricci-Tersenghi Federico
Disordered Systems and Neura...
Machine Learning

By leveraging tools from the statistical mechanics of complex systems, in these short notes we extend the architecture of a neural network for hetero-associative memory (called three-directional associative memories, TAM) to explore supervised and unsupervised learning protocols. In particular, by providing entropic-heterogeneous datasets to its various layers, we predict and quantify a new emergent phenomenon -- that we term {\em layer's cooperativeness} -- where the interpl...

Find SimilarView on arXiv

Representational capacity of a set of independent neurons

January 31, 2002

83% Match
Ines Samengo, Alessandro Treves
Disordered Systems and Neura...

The capacity with which a system of independent neuron-like units represents a given set of stimuli is studied by calculating the mutual information between the stimuli and the neural responses. Both discrete noiseless and continuous noisy neurons are analyzed. In both cases, the information grows monotonically with the number of neurons considered. Under the assumption that neurons are independent, the mutual information rises linearly from zero, and approaches exponentially...

Find SimilarView on arXiv

Fundamental problems in statistical physics XIV: Lecture on Machine Learning

February 11, 2022

83% Match
Aurélien Decelle
Disordered Systems and Neura...
Statistical Mechanics

The recent progresses in Machine Learning opened the door to actual applications of learning algorithms but also to new research directions both in the field of Machine Learning directly and, at the edges with other disciplines. The case that interests us is the interface with physics, and more specifically Statistical Physics. In this short lecture, I will try to present first a brief introduction to Machine Learning from the angle of neural networks. After explaining quickl...

Find SimilarView on arXiv

A Functional Perspective on Learning Symmetric Functions with Neural Networks

August 16, 2020

83% Match
Aaron Zweig, Joan Bruna
Machine Learning
Machine Learning

Symmetric functions, which take as input an unordered, fixed-size set, are known to be universally representable by neural networks that enforce permutation invariance. These architectures only give guarantees for fixed input sizes, yet in many practical applications, including point clouds and particle physics, a relevant notion of generalization should include varying the input size. In this work we treat symmetric functions (of any size) as functions over probability measu...

Find SimilarView on arXiv