ID: 1809.07850

Optimal Bayesian clustering using non-negative matrix factorization

September 20, 2018

View on ArXiv

Similar papers 3

Variational Bayes Approximations for Clustering via Mixtures of Normal Inverse Gaussian Distributions

September 7, 2013

88% Match
Sanjeena Subedi, Paul D. McNicholas
Methodology
Computation
Machine Learning

Parameter estimation for model-based clustering using a finite mixture of normal inverse Gaussian (NIG) distributions is achieved through variational Bayes approximations. Univariate NIG mixtures and multivariate NIG mixtures are considered. The use of variational Bayes approximations here is a substantial departure from the traditional EM approach and alleviates some of the associated computational complexities and uncertainties. Our variational algorithm is applied to simul...

Find SimilarView on arXiv

Model-Based Hierarchical Clustering

January 16, 2013

88% Match
Shivakumar Vaithyanathan, Byron E Dom
Machine Learning
Artificial Intelligence
Machine Learning

We present an approach to model-based hierarchical clustering by formulating an objective function based on a Bayesian analysis. This model organizes the data into a cluster hierarchy while specifying a complex feature-set partitioning that is a key component of our model. Features can have either a unique distribution in every cluster or a common distribution over some (or even all) of the clusters. The cluster subsets over which these features have such a common distributio...

Find SimilarView on arXiv

Inferring Hierarchical Mixture Structures: A Bayesian Nonparametric Approach

May 13, 2019

88% Match
Weipeng Huang, Nishma Laitonjam, ... , Hurley Neil
Machine Learning
Artificial Intelligence
Machine Learning

This paper focuses on the problem of hierarchical non-overlapping clustering of a dataset. In such a clustering, each data item is associated with exactly one leaf node and each internal node is associated with all the data items stored in the sub-tree beneath it, so that each level of the hierarchy corresponds to a partition of the dataset. We develop a novel Bayesian nonparametric method combining the nested Chinese Restaurant Process (nCRP) and the Hierarchical Dirichlet P...

Find SimilarView on arXiv

Bayesian Clustering via Fusing of Localized Densities

March 31, 2023

87% Match
Alexander Dombowsky, David B. Dunson
Methodology

Bayesian clustering typically relies on mixture models, with each component interpreted as a different cluster. After defining a prior for the component parameters and weights, Markov chain Monte Carlo (MCMC) algorithms are commonly used to produce samples from the posterior distribution of the component labels. The data are then clustered by minimizing the expectation of a clustering loss function that favours similarity to the component labels. Unfortunately, although these...

Find SimilarView on arXiv

Model based clustering of multinomial count data

July 28, 2022

87% Match
Panagiotis Papastamoulis
Methodology
Computation

We consider the problem of inferring an unknown number of clusters in replicated multinomial data. Under a model based clustering point of view, this task can be treated by estimating finite mixtures of multinomial distributions with or without covariates. Both Maximum Likelihood (ML) as well as Bayesian estimation are taken into account. Under a Maximum Likelihood approach, we provide an Expectation--Maximization (EM) algorithm which exploits a careful initialization procedu...

Find SimilarView on arXiv

clusterBMA: Bayesian model averaging for clustering

September 9, 2022

87% Match
Owen Forbes, Edgar Santos-Fernandez, Paul Pao-Yen Wu, Hong-Bo Xie, Paul E. Schwenn, Jim Lagopoulos, Lia Mills, Dashiell D. Sacks, ... , Mengersen Kerrie
Methodology
Applications
Machine Learning

Various methods have been developed to combine inference across multiple sets of results for unsupervised clustering, within the ensemble clustering literature. The approach of reporting results from one `best' model out of several candidate clustering models generally ignores the uncertainty that arises from model selection, and results in inferences that are sensitive to the particular model and parameters chosen. Bayesian model averaging (BMA) is a popular approach for com...

Find SimilarView on arXiv

Fast Clustering and Topic Modeling Based on Rank-2 Nonnegative Matrix Factorization

September 3, 2015

87% Match
Da Kuang, Barry Drake, Haesun Park
Machine Learning
Information Retrieval
Numerical Analysis

The importance of unsupervised clustering and topic modeling is well recognized with ever-increasing volumes of text data. In this paper, we propose a fast method for hierarchical clustering and topic modeling called HierNMF2. Our method is based on fast Rank-2 nonnegative matrix factorization (NMF) that performs binary clustering and an efficient node splitting rule. Further utilizing the final leaf nodes generated in HierNMF2 and the idea of nonnegative least squares fittin...

Find SimilarView on arXiv

Nonnegative Matrix Factorization with Local Similarity Learning

July 9, 2019

87% Match
Chong Peng, Zhao Kang, ... , Cheng Qiang
Machine Learning
Computer Vision and Pattern ...
Machine Learning

Existing nonnegative matrix factorization methods focus on learning global structure of the data to construct basis and coefficient matrices, which ignores the local structure that commonly exists among data. In this paper, we propose a new type of nonnegative matrix factorization method, which learns local similarity and clustering in a mutually enhancing way. The learned new representation is more representative in that it better reveals inherent geometric property of the d...

Find SimilarView on arXiv

Non-negative matrix factorization algorithms greatly improve topic model fits

May 27, 2021

87% Match
Peter Carbonetto, Abhishek Sarkar, ... , Stephens Matthew
Machine Learning
Machine Learning
Computation

We report on the potential for using algorithms for non-negative matrix factorization (NMF) to improve parameter estimation in topic models. While several papers have studied connections between NMF and topic models, none have suggested leveraging these connections to develop new algorithms for fitting topic models. NMF avoids the "sum-to-one" constraints on the topic model parameters, resulting in an optimization problem with simpler structure and more efficient computations...

Find SimilarView on arXiv

Model-based clustering based on sparse finite Gaussian mixtures

June 22, 2016

87% Match
Gertraud Malsiner-Walli, Sylvia Frühwirth-Schnatter, Bettina Grün
Methodology

In the framework of Bayesian model-based clustering based on a finite mixture of Gaussian distributions, we present a joint approach to estimate the number of mixture components and identify cluster-relevant variables simultaneously as well as to obtain an identified model. Our approach consists in specifying sparse hierarchical priors on the mixture weights and component means. In a deliberately overfitting mixture model the sparse prior on the weights empties superfluous co...

Find SimilarView on arXiv