December 31, 2020
Ricci flat metrics for Calabi-Yau threefolds are not known analytically. In this work, we employ techniques from machine learning to deduce numerical flat metrics for the Fermat quintic, for the Dwork quintic, and for the Tian-Yau manifold. This investigation employs a single neural network architecture that is capable of approximating Ricci flat Kaehler metrics for several Calabi-Yau manifolds of dimensions two and three. We show that measures that assess the Ricci flatness of the geometry decrease after training by three orders of magnitude. This is corroborated on the validation set, where the improvement is more modest. Finally, we demonstrate that discrete symmetries of manifolds can be learned in the process of learning the metric.
Similar papers 1
November 17, 2022
Finding Ricci-flat (Calabi-Yau) metrics is a long standing problem in geometry with deep implications for string theory and phenomenology. A new attack on this problem uses neural networks to engineer approximations to the Calabi-Yau metric within a given K\"ahler class. In this paper we investigate numerical Ricci-flat metrics over smooth and singular K3 surfaces and Calabi-Yau threefolds. Using these Ricci-flat metric approximations for the Cefal\'u family of quartic twofol...
December 9, 2020
We propose machine learning inspired methods for computing numerical Calabi-Yau (Ricci flat K\"ahler) metrics, and implement them using Tensorflow/Keras. We compare them with previous work, and find that they are far more accurate for manifolds with little or no symmetry. We also discuss issues such as overparameterization and choice of optimization methods.
October 18, 2019
We apply machine learning to the problem of finding numerical Calabi-Yau metrics. Building on Donaldson's algorithm for calculating balanced metrics on K\"ahler manifolds, we combine conventional curve fitting and machine-learning techniques to numerically approximate Ricci-flat metrics. We show that machine learning is able to predict the Calabi-Yau metric and quantities associated with it, such as its determinant, having seen only a small sample of training data. Using this...
May 26, 2022
We introduce neural networks to compute numerical Ricci-flat CY metrics for complete intersection and Kreuzer-Skarke Calabi-Yau manifolds at any point in K\"ahler and complex structure moduli space, and introduce the package cymetric which provides computation realizations of these techniques. In particular, we develop and computationally realize methods for point-sampling on these manifolds. The training for the neural networks is carried out subject to a custom loss functio...
December 28, 2023
Calabi-Yau (CY) manifolds play a ubiquitous role in string theory. As a supersymmetry-preserving choice for the 6 extra compact dimensions of superstring compactifications, these spaces provide an arena in which to explore the rich interplay between physics and geometry. These lectures will focus on compact CY manifolds and the long standing problem of determining their Ricci flat metrics. Despite powerful existence theorems, no analytic expressions for these metrics are know...
July 9, 2024
We present new invariant machine learning models that approximate the Ricci-flat metric on Calabi-Yau (CY) manifolds with discrete symmetries. We accomplish this by combining the $\phi$-model of the cymetric package with non-trainable, $G$-invariant, canonicalization layers that project the $\phi$-model's input data (i.e. points sampled from the CY geometry) to the fundamental domain of a given symmetry group $G$. These $G$-invariant layers are easy to concatenate, provided o...
October 30, 2023
We develop a theory of flows in the space of Riemannian metrics induced by neural network gradient descent. This is motivated in part by recent advances in approximating Calabi-Yau metrics with neural networks and is enabled by recent advances in understanding flows in the space of neural networks. We derive the corresponding metric flow equations, which are governed by a metric neural tangent kernel, a complicated, non-local object that evolves in time. However, many archite...
November 2, 2021
We present a new machine learning library for computing metrics of string compactification spaces. We benchmark the performance on Monte-Carlo sampled integrals against previous numerical approximations and find that our neural networks are more sample- and computation-efficient. We are the first to provide the possibility to compute these metrics for arbitrary, user-specified shape and size parameters of the compact space and observe a linear relation between optimization of...
December 20, 2021
We apply machine learning to the problem of finding numerical Calabi-Yau metrics. We extend previous work on learning approximate Ricci-flat metrics calculated using Donaldson's algorithm to the much more accurate "optimal" metrics of Headrick and Nassar. We show that machine learning is able to predict the K\"ahler potential of a Calabi-Yau metric having seen only a small sample of training data.
December 8, 2020
We use machine learning to approximate Calabi-Yau and SU(3)-structure metrics, including for the first time complex structure moduli dependence. Our new methods furthermore improve existing numerical approximations in terms of accuracy and speed. Knowing these metrics has numerous applications, ranging from computations of crucial aspects of the effective field theory of string compactifications such as the canonical normalizations for Yukawa couplings, and the massive string...