ID: 1706.02714

Deep-Learning the Landscape

June 8, 2017

View on ArXiv

Similar papers 2

On the Modeling of Error Functions as High Dimensional Landscapes for Weight Initialization in Learning Networks

July 20, 2016

87% Match
Julius, Gopinath Mahale, ... , Adityakrishna C. S.
Machine Learning
Computer Vision and Pattern ...
Data Analysis, Statistics an...
Machine Learning

Next generation deep neural networks for classification hosted on embedded platforms will rely on fast, efficient, and accurate learning algorithms. Initialization of weights in learning networks has a great impact on the classification accuracy. In this paper we focus on deriving good initial weights by modeling the error function of a deep neural network as a high-dimensional landscape. We observe that due to the inherent complexity in its algebraic structure, such an error...

Find SimilarView on arXiv

The semantic landscape paradigm for neural networks

July 18, 2023

87% Match
Shreyas Gokhale
Machine Learning
Disordered Systems and Neura...
Statistical Mechanics

Deep neural networks exhibit a fascinating spectrum of phenomena ranging from predictable scaling laws to the unpredictable emergence of new capabilities as a function of training time, dataset size and network size. Analysis of these phenomena has revealed the existence of concepts and algorithms encoded within the learned representations of these networks. While significant strides have been made in explaining observed phenomena separately, a unified framework for understan...

Find SimilarView on arXiv

Geometry of energy landscapes and the optimizability of deep neural networks

August 1, 2018

87% Match
Simon Becker, Yao Zhang, Alpha A. Lee
Disordered Systems and Neura...
Machine Learning
Machine Learning

Deep neural networks are workhorse models in machine learning with multiple layers of non-linear functions composed in series. Their loss function is highly non-convex, yet empirically even gradient descent minimisation is sufficient to arrive at accurate and predictive models. It is hitherto unknown why are deep neural networks easily optimizable. We analyze the energy landscape of a spin glass model of deep neural networks using random matrix theory and algebraic geometry. ...

Find SimilarView on arXiv

Learning Curves for Deep Neural Networks: A Gaussian Field Theory Perspective

June 12, 2019

87% Match
Omry Cohen, Or Malka, Zohar Ringel
Machine Learning
Statistical Mechanics
Neural and Evolutionary Comp...
Data Analysis, Statistics an...
Machine Learning

In the past decade, deep neural networks (DNNs) came to the fore as the leading machine learning algorithms for a variety of tasks. Their raise was founded on market needs and engineering craftsmanship, the latter based more on trial and error than on theory. While still far behind the application forefront, the theoretical study of DNNs has recently made important advancements in analyzing the highly over-parameterized regime where some exact results have been obtained. Leve...

Find SimilarView on arXiv

(K)not machine learning

January 21, 2022

87% Match
Jessica Craven, Mark Hughes, ... , Kar Arjun
Geometric Topology

We review recent efforts to machine learn relations between knot invariants. Because these knot invariants have meaning in physics, we explore aspects of Chern-Simons theory and higher dimensional gauge theories. The goal of this work is to translate numerical experiments with Big Data to new analytic results.

Find SimilarView on arXiv

Intelligent Explorations of the String Theory Landscape

April 17, 2022

87% Match
Andrei Constantin
Artificial Intelligence

The goal of identifying the Standard Model of particle physics and its extensions within string theory has been one of the principal driving forces in string phenomenology. Recently, the incorporation of artificial intelligence in string theory and certain theoretical advancements have brought to light unexpected solutions to mathematical hurdles that have so far hindered progress in this direction. In this review we focus on model building efforts in the context of the $E_8\...

Find SimilarView on arXiv

Machine learning for complete intersection Calabi-Yau manifolds: a methodological study

July 30, 2020

87% Match
Harold Erbin, Riccardo Finotello
Machine Learning
Algebraic Geometry

We revisit the question of predicting both Hodge numbers $h^{1,1}$ and $h^{2,1}$ of complete intersection Calabi-Yau (CICY) 3-folds using machine learning (ML), considering both the old and new datasets built respectively by Candelas-Dale-Lutken-Schimmrigk / Green-H\"ubsch-Lutken and by Anderson-Gao-Gray-Lee. In real world applications, implementing a ML system rarely reduces to feed the brute data to the algorithm. Instead, the typical workflow starts with an exploratory dat...

Find SimilarView on arXiv

Narrowing the Gap between Combinatorial and Hyperbolic Knot Invariants via Deep Learning

April 27, 2022

87% Match
Daniel Grünbaum
Geometric Topology

We present a statistical approach for the discovery of relationships between mathematical entities that is based on linear regression and deep learning with fully connected artificial neural networks. The strategy is applied to computational knot data and empirical connections between combinatorial and hyperbolic knot invariants are revealed.

Find SimilarView on arXiv

The Modern Mathematics of Deep Learning

May 9, 2021

87% Match
Julius Berner, Philipp Grohs, ... , Petersen Philipp
Machine Learning
Machine Learning

We describe the new field of mathematical analysis of deep learning. This field emerged around a list of research questions that were not answered within the classical framework of learning theory. These questions concern: the outstanding generalization power of overparametrized neural networks, the role of depth in deep architectures, the apparent absence of the curse of dimensionality, the surprisingly successful optimization performance despite the non-convexity of the pro...

Find SimilarView on arXiv

Physicist's Journeys Through the AI World - A Topical Review. There is no royal road to unsupervised learning

May 2, 2019

87% Match
Imad Alhousseini, Wissam Chemissany, ... , Nasrallah Aly
Machine Learning
Disordered Systems and Neura...
Computational Physics
Machine Learning

Artificial Intelligence (AI), defined in its most simple form, is a technological tool that makes machines intelligent. Since learning is at the core of intelligence, machine learning poses itself as a core sub-field of AI. Then there comes a subclass of machine learning, known as deep learning, to address the limitations of their predecessors. AI has generally acquired its prominence over the past few years due to its considerable progress in various fields. AI has vastly in...

Find SimilarView on arXiv