ID: 2312.06581

Grokking Group Multiplication with Cosets

December 11, 2023

View on ArXiv

Similar papers 2

A Unified Framework for Discovering Discrete Symmetries

September 6, 2023

87% Match
Pavan Karjol, Rohan Kashyap, ... , P Prathosh A.
Machine Learning
Computer Vision and Pattern ...

We consider the problem of learning a function respecting a symmetry from among a class of symmetries. We develop a unified framework that enables symmetry discovery across a broad range of subgroups including locally symmetric, dihedral and cyclic subgroups. At the core of the framework is a novel architecture composed of linear, matrix-valued and non-linear functions that expresses functions invariant to these subgroups in a principled manner. The structure of the architect...

Find SimilarView on arXiv

Neural Fourier Transform: A General Approach to Equivariant Representation Learning

May 29, 2023

86% Match
Masanori Koyama, Kenji Fukumizu, ... , Miyato Takeru
Machine Learning
Machine Learning

Symmetry learning has proven to be an effective approach for extracting the hidden structure of data, with the concept of equivariance relation playing the central role. However, most of the current studies are built on architectural theory and corresponding assumptions on the form of data. We propose Neural Fourier Transform (NFT), a general framework of learning the latent linear action of the group without assuming explicit knowledge of how the group acts on data. We prese...

Find SimilarView on arXiv

Neural network interpretation using descrambler groups

December 2, 2019

86% Match
Jake L. Amey, Jake Keeley, ... , Kuprov Ilya
Signal Processing

The lack of interpretability and trust is a much-criticised feature of deep neural networks. In fully connected nets, the signalling between inner layers is scrambled because backpropagation training does not require perceptrons to be arranged in any particular order. The result is a black box; this problem is particularly severe in scientific computing and digital signal processing (DSP), where neutral nets perform abstract mathematical transformations that do not reduce to ...

Find SimilarView on arXiv

Interpreting Grokked Transformers in Complex Modular Arithmetic

February 26, 2024

86% Match
Hiroki Furuta, Gouki Minegishi, ... , Matsuo Yutaka
Machine Learning
Artificial Intelligence

Grokking has been actively explored to reveal the mystery of delayed generalization. Identifying interpretable algorithms inside the grokked models is a suggestive hint to understanding its mechanism. In this work, beyond the simplest and well-studied modular addition, we observe the internal circuits learned through grokking in complex modular arithmetic via interpretable reverse engineering, which highlights the significant difference in their dynamics: subtraction poses a ...

Find SimilarView on arXiv

Why does Deep Learning work? - A perspective from Group Theory

December 20, 2014

86% Match
Arnab Paul, Suresh Venkatasubramanian
Machine Learning
Neural and Evolutionary Comp...
Machine Learning

Why does Deep Learning work? What representations does it capture? How do higher-order representations emerge? We study these questions from the perspective of group theory, thereby opening a new approach towards a theory of Deep learning. One factor behind the recent resurgence of the subject is a key algorithmic step called pre-training: first search for a good generative model for the input samples, and repeat the process one layer at a time. We show deeper implications ...

Find SimilarView on arXiv

Representation Theory for Geometric Quantum Machine Learning

October 14, 2022

86% Match
Michael Ragone, Paolo Braccia, Quynh T. Nguyen, Louis Schatzki, Patrick J. Coles, Frederic Sauvage, ... , Cerezo M.
Machine Learning
Representation Theory
Machine Learning

Recent advances in classical machine learning have shown that creating models with inductive biases encoding the symmetries of a problem can greatly improve performance. Importation of these ideas, combined with an existing rich body of work at the nexus of quantum theory and symmetry, has given rise to the field of Geometric Quantum Machine Learning (GQML). Following the success of its classical counterpart, it is reasonable to expect that GQML will play a crucial role in de...

Find SimilarView on arXiv

Symmetry Group Equivariant Architectures for Physics

March 11, 2022

86% Match
Alexander Bogatskiy, Sanmay Ganguly, Thomas Kipf, Risi Kondor, David W. Miller, Daniel Murnane, Jan T. Offermann, Mariel Pettee, Phiala Shanahan, ... , Thais Savannah
Machine Learning
Instrumentation and Methods ...
Artificial Intelligence

Physical theories grounded in mathematical symmetries are an essential component of our understanding of a wide range of properties of the universe. Similarly, in the domain of machine learning, an awareness of symmetries such as rotation or permutation invariance has driven impressive performance breakthroughs in computer vision, natural language processing, and other important applications. In this report, we argue that both the physics community and the broader machine lea...

Find SimilarView on arXiv

Learning Linear Groups in Neural Networks

May 29, 2023

86% Match
Emmanouil Theodosis, Karim Helwani, Demba Ba
Machine Learning
Neural and Evolutionary Comp...

Employing equivariance in neural networks leads to greater parameter efficiency and improved generalization performance through the encoding of domain knowledge in the architecture; however, the majority of existing approaches require an a priori specification of the desired symmetries. We present a neural network architecture, Linear Group Networks (LGNs), for learning linear groups acting on the weight space of neural networks. Linear groups are desirable due to their inher...

Find SimilarView on arXiv

Learning Lie Group Symmetry Transformations with Neural Networks

July 4, 2023

86% Match
Alex Gabel, Victoria Klein, Riccardo Valperga, Jeroen S. W. Lamb, Kevin Webster, ... , Gavves Efstratios
Machine Learning
Computer Vision and Pattern ...

The problem of detecting and quantifying the presence of symmetries in datasets is useful for model selection, generative modeling, and data analysis, amongst others. While existing methods for hard-coding transformations in neural networks require prior knowledge of the symmetries of the task at hand, this work focuses on discovering and characterizing unknown symmetries present in the dataset, namely, Lie group symmetry transformations beyond the traditional ones usually co...

Find SimilarView on arXiv

Universal approximations of permutation invariant/equivariant functions by deep neural networks

March 5, 2019

86% Match
Akiyoshi Sannai, Yuuki Takai, Matthieu Cordonnier
Machine Learning
Machine Learning

In this paper, we develop a theory about the relationship between $G$-invariant/equivariant functions and deep neural networks for finite group $G$. Especially, for a given $G$-invariant/equivariant function, we construct its universal approximator by deep neural network whose layers equip $G$-actions and each affine transformations are $G$-equivariant/invariant. Due to representation theory, we can show that this approximator has exponentially fewer free parameters than usua...

Find SimilarView on arXiv