ID: 2312.08550

Harmonics of Learning: Universal Fourier Features Emerge in Invariant Networks

December 13, 2023

View on ArXiv
Giovanni Luca Marchetti, Christopher Hillar, Danica Kragic, Sophia Sanborn
Computer Science
Electrical Engineering and S...
Machine Learning
Artificial Intelligence
Signal Processing

In this work, we formally prove that, under certain conditions, if a neural network is invariant to a finite group then its weights recover the Fourier transform on that group. This provides a mathematical explanation for the emergence of Fourier features -- a ubiquitous phenomenon in both biological and artificial learning systems. The results hold even for non-commutative groups, in which case the Fourier transform encodes all the irreducible unitary group representations. Our findings have consequences for the problem of symmetry discovery. Specifically, we demonstrate that the algebraic structure of an unknown group can be recovered from the weights of a network that is at least approximately invariant within certain bounds. Overall, this work contributes to a foundation for an algebraic learning theory of invariant neural network representations.

Similar papers 1

Neural Fourier Transform: A General Approach to Equivariant Representation Learning

May 29, 2023

91% Match
Masanori Koyama, Kenji Fukumizu, ... , Miyato Takeru
Machine Learning
Machine Learning

Symmetry learning has proven to be an effective approach for extracting the hidden structure of data, with the concept of equivariance relation playing the central role. However, most of the current studies are built on architectural theory and corresponding assumptions on the form of data. We propose Neural Fourier Transform (NFT), a general framework of learning the latent linear action of the group without assuming explicit knowledge of how the group acts on data. We prese...

Find SimilarView on arXiv

A New Neural Network Architecture Invariant to the Action of Symmetry Subgroups

December 11, 2020

90% Match
Piotr Kicki, Mete Ozay, Piotr Skrzypczyński
Machine Learning
Artificial Intelligence

We propose a computationally efficient $G$-invariant neural network that approximates functions invariant to the action of a given permutation subgroup $G \leq S_n$ of the symmetric group on input data. The key element of the proposed network architecture is a new $G$-invariant transformation module, which produces a $G$-invariant latent representation of the input data. Theoretical considerations are supported by numerical experiments, which demonstrate the effectiveness and...

Find SimilarView on arXiv

Bispectral Neural Networks

September 7, 2022

90% Match
Sophia Sanborn, Christian Shewmake, ... , Hillar Christopher
Machine Learning

We present a neural network architecture, Bispectral Neural Networks (BNNs) for learning representations that are invariant to the actions of compact commutative groups on the space over which a signal is defined. The model incorporates the ansatz of the bispectrum, an analytically defined group invariant that is complete -- that is, it preserves all signal structure while removing only the variation due to group actions. Here, we demonstrate that BNNs are able to simultaneou...

Find SimilarView on arXiv

A Computationally Efficient Neural Network Invariant to the Action of Symmetry Subgroups

February 18, 2020

90% Match
Piotr Kicki, Mete Ozay, Piotr Skrzypczyński
Machine Learning
Neural and Evolutionary Comp...
Machine Learning

We introduce a method to design a computationally efficient $G$-invariant neural network that approximates functions invariant to the action of a given permutation subgroup $G \leq S_n$ of the symmetric group on input data. The key element of the proposed network architecture is a new $G$-invariant transformation module, which produces a $G$-invariant latent representation of the input data. This latent representation is then processed with a multi-layer perceptron in the net...

Find SimilarView on arXiv

Representing and Learning Functions Invariant Under Crystallographic Groups

June 8, 2023

90% Match
Ryan P. Adams, Peter Orbanz
Machine Learning
Materials Science
Machine Learning

Crystallographic groups describe the symmetries of crystals and other repetitive structures encountered in nature and the sciences. These groups include the wallpaper and space groups. We derive linear and nonlinear representations of functions that are (1) smooth and (2) invariant under such a group. The linear representation generalizes the Fourier basis to crystallographically invariant basis functions. We show that such a basis exists for each crystallographic group, that...

Find SimilarView on arXiv

A Toy Model of Universality: Reverse Engineering How Networks Learn Group Operations

February 6, 2023

90% Match
Bilal Chughtai, Lawrence Chan, Neel Nanda
Machine Learning
Artificial Intelligence
Representation Theory

Universality is a key hypothesis in mechanistic interpretability -- that different models learn similar features and circuits when trained on similar tasks. In this work, we study the universality hypothesis by examining how small neural networks learn to implement group composition. We present a novel algorithm by which neural networks may implement composition for any finite group via mathematical representation theory. We then show that networks consistently learn this alg...

Find SimilarView on arXiv

Joint Group Invariant Functions on Data-Parameter Domain Induce Universal Neural Networks

October 5, 2023

90% Match
Sho Sonoda, Hideyuki Ishi, ... , Ikeda Masahiro
Machine Learning
Machine Learning

The symmetry and geometry of input data are considered to be encoded in the internal data representation inside the neural network, but the specific encoding rule has been less investigated. In this study, we present a systematic method to induce a generalized neural network and its right inverse operator, called the ridgelet transform, from a joint group invariant function on the data-parameter domain. Since the ridgelet transform is an inverse, (1) it can describe the arran...

Find SimilarView on arXiv

Grokking Group Multiplication with Cosets

December 11, 2023

90% Match
Dashiell Stander, Qinan Yu, ... , Biderman Stella
Machine Learning
Artificial Intelligence
Representation Theory

We use the group Fourier transform over the symmetric group $S_n$ to reverse engineer a 1-layer feedforward network that has "grokked" the multiplication of $S_5$ and $S_6$. Each model discovers the true subgroup structure of the full group and converges on circuits that decompose the group multiplication into the multiplication of the group's conjugate subgroups. We demonstrate the value of using the symmetries of the data and models to understand their mechanisms and hold u...

Find SimilarView on arXiv

Universal Equivariant Multilayer Perceptrons

February 7, 2020

89% Match
Siamak Ravanbakhsh
Machine Learning
Neural and Evolutionary Comp...
Group Theory
Machine Learning

Group invariant and equivariant Multilayer Perceptrons (MLP), also known as Equivariant Networks, have achieved remarkable success in learning on a variety of data structures, such as sequences, images, sets, and graphs. Using tools from group theory, this paper proves the universality of a broad class of equivariant MLPs with a single hidden layer. In particular, it is shown that having a hidden layer on which the group acts regularly is sufficient for universal equivariance...

Find SimilarView on arXiv

Learning Stable Group Invariant Representations with Convolutional Networks

January 16, 2013

89% Match
Joan Bruna, Arthur Szlam, Yann LeCun
Artificial Intelligence
Numerical Analysis

Transformation groups, such as translations or rotations, effectively express part of the variability observed in many recognition problems. The group structure enables the construction of invariant signal representations with appealing mathematical properties, where convolutions, together with pooling operators, bring stability to additive and geometric perturbations of the input. Whereas physical transformation groups are ubiquitous in image and audio applications, they do ...

Find SimilarView on arXiv