ID: cond-mat/0104011

Multilayer neural networks with extensively many hidden units

April 1, 2001

View on ArXiv

Similar papers 2

On neural network kernels and the storage capacity problem

January 12, 2022

84% Match
Jacob A. Zavatone-Veth, Cengiz Pehlevan
Disordered Systems and Neura...
Machine Learning

In this short note, we reify the connection between work on the storage capacity problem in wide two-layer treelike neural networks and the rapidly-growing body of literature on kernel limits of wide neural networks. Concretely, we observe that the "effective order parameter" studied in the statistical mechanics literature is exactly equivalent to the infinite-width Neural Network Gaussian Process Kernel. This correspondence connects the expressivity and trainability of wide ...

Find SimilarView on arXiv

Rethinking generalization requires revisiting old ideas: statistical mechanics approaches and complex learning behavior

October 26, 2017

84% Match
Charles H. Martin, Michael W. Mahoney
Machine Learning
Machine Learning

We describe an approach to understand the peculiar and counterintuitive generalization properties of deep neural networks. The approach involves going beyond worst-case theoretical capacity control frameworks that have been popular in machine learning in recent years to revisit old ideas in the statistical mechanics of neural networks. Within this approach, we present a prototypical Very Simple Deep Learning (VSDL) model, whose behavior is controlled by two control parameters...

Find SimilarView on arXiv

Correlations between hidden units in multilayer neural networks and replica symmetry breaking

June 14, 1999

84% Match
D. Malzahn, A. Engel
Disordered Systems and Neura...
Statistical Mechanics

We consider feed-forward neural networks with one hidden layer, tree architecture and a fixed hidden-to-output Boolean function. Focusing on the saturation limit of the storage problem the influence of replica symmetry breaking on the distribution of local fields at the hidden units is investigated. These field distributions determine the probability for finding a specific activation pattern of the hidden units as well as the corresponding correlation coefficients and therefo...

Find SimilarView on arXiv

On the loss of learning capability inside an arrangement of neural networks

January 9, 2020

84% Match
Ivan Arraut, Diana Diaz
Other Computer Science

We analyze the loss of information and the loss of learning capability inside an arrangement of neural networks. Our method is new and based on the formulation of non-unitary Bogoliubov transformations in order to connect the information between different points of the arrangement. This can be done after expanding the activation function in a Fourier series and then assuming that its information is stored inside a Quantum scalar field.

Find SimilarView on arXiv

Memory Capacity of a Random Neural Network

November 14, 2012

84% Match
Matt Stowe
Neural and Evolutionary Comp...

This paper considers the problem of information capacity of a random neural network. The network is represented by matrices that are square and symmetrical. The matrices have a weight which determines the highest and lowest possible value found in the matrix. The examined matrices are randomly generated and analyzed by a computer program. We find the surprising result that the capacity of the network is a maximum for the binary random neural network and it does not change as ...

Find SimilarView on arXiv

Computational Capabilities of Random Automata Networks for Reservoir Computing

December 8, 2012

84% Match
David Snyder, Alireza Goudarzi, Christof Teuscher
Adaptation and Self-Organizi...
Disordered Systems and Neura...
Neural and Evolutionary Comp...

This paper underscores the conjecture that intrinsic computation is maximal in systems at the "edge of chaos." We study the relationship between dynamics and computational capability in Random Boolean Networks (RBN) for Reservoir Computing (RC). RC is a computational paradigm in which a trained readout layer interprets the dynamics of an excitable component (called the reservoir) that is perturbed by external input. The reservoir is often implemented as a homogeneous recurren...

Find SimilarView on arXiv

Memory and Information Processing in Recurrent Neural Networks

April 23, 2016

84% Match
Alireza Goudarzi, Sarah Marzen, Peter Banda, Guy Feldman, ... , Stefanovic Darko
Neural and Evolutionary Comp...

Recurrent neural networks (RNN) are simple dynamical systems whose computational power has been attributed to their short-term memory. Short-term memory of RNNs has been previously studied analytically only for the case of orthogonal networks, and only under annealed approximation, and uncorrelated input. Here for the first time, we present an exact solution to the memory capacity and the task-solving performance as a function of the structure of a given network instance, ena...

Find SimilarView on arXiv

Information in Infinite Ensembles of Infinitely-Wide Neural Networks

November 20, 2019

84% Match
Ravid Shwartz-Ziv, Alexander A. Alemi
Machine Learning
Information Theory
Information Theory
Machine Learning

In this preliminary work, we study the generalization properties of infinite ensembles of infinitely-wide neural networks. Amazingly, this model family admits tractable calculations for many information-theoretic quantities. We report analytical and empirical investigations in the search for signals that correlate with generalization.

Find SimilarView on arXiv

Spontaneous Emergence of Computation in Network Cascades

April 25, 2022

84% Match
Galen Wilkerson, Sotiris Moschoyiannis, Henrik Jeldtoft Jensen
Physics and Society
Disordered Systems and Neura...
Artificial Intelligence
Formal Languages and Automat...

Neuronal network computation and computation by avalanche supporting networks are of interest to the fields of physics, computer science (computation theory as well as statistical or machine learning) and neuroscience. Here we show that computation of complex Boolean functions arises spontaneously in threshold networks as a function of connectivity and antagonism (inhibition), computed by logic automata (motifs) in the form of computational cascades. We explain the emergent i...

Find SimilarView on arXiv

DNNs as Layers of Cooperating Classifiers

January 17, 2020

84% Match
Marelie H. Davel, Marthinus W. Theunissen, ... , Barnard Etienne
Machine Learning
Machine Learning

A robust theoretical framework that can describe and predict the generalization ability of deep neural networks (DNNs) in general circumstances remains elusive. Classical attempts have produced complexity metrics that rely heavily on global measures of compactness and capacity with little investigation into the effects of sub-component collaboration. We demonstrate intriguing regularities in the activation patterns of the hidden nodes within fully-connected feedforward networ...

Find SimilarView on arXiv