ID: 2207.13984

Model based clustering of multinomial count data

July 28, 2022

View on ArXiv
Panagiotis Papastamoulis
Statistics
Methodology
Computation

We consider the problem of inferring an unknown number of clusters in replicated multinomial data. Under a model based clustering point of view, this task can be treated by estimating finite mixtures of multinomial distributions with or without covariates. Both Maximum Likelihood (ML) as well as Bayesian estimation are taken into account. Under a Maximum Likelihood approach, we provide an Expectation--Maximization (EM) algorithm which exploits a careful initialization procedure combined with a ridge--stabilized implementation of the Newton--Raphson method in the M--step. Under a Bayesian setup, a stochastic gradient Markov chain Monte Carlo (MCMC) algorithm embedded within a prior parallel tempering scheme is devised. The number of clusters is selected according to the Integrated Completed Likelihood criterion in the ML approach and estimating the number of non-empty components in overfitting mixture models in the Bayesian case. Our method is illustrated in simulated data and applied to two real datasets. An R package is available at https://github.com/mqbssppe/multinomialLogitMix.

Similar papers 1

Identifying the number of clusters in discrete mixture models

September 25, 2014

90% Match
Cláudia Silvestre, Margarida G. M. S. Cardoso, Mário A. T. Figueiredo
Methodology
Other Statistics

Research on cluster analysis for categorical data continues to develop, with new clustering algorithms being proposed. However, in this context, the determination of the number of clusters is rarely addressed. In this paper, we propose a new approach in which clustering of categorical data and the estimation of the number of clusters is carried out simultaneously. Assuming that the data originate from a finite mixture of multinomial distributions, we develop a method to selec...

Find SimilarView on arXiv

Bayesian Finite Mixture Models

July 7, 2024

90% Match
Bettina Grün, Gertraud Malsiner-Walli
Methodology

Finite mixture models are a useful statistical model class for clustering and density approximation. In the Bayesian framework finite mixture models require the specification of suitable priors in addition to the data model. These priors allow to avoid spurious results and provide a principled way to define cluster shapes and a preference for specific cluster solutions. A generic model estimation scheme for finite mixtures with a fixed number of components is available using ...

Find SimilarView on arXiv

Multinomial Cluster-Weighted Models for High-Dimensional Data

August 23, 2022

89% Match
Kehinde Olobatuyi, Oludare Ariyo
Methodology

Modeling of high-dimensional data is very important to categorize different classes. We develop a new mixture model called Multinomial cluster-weighted model (MCWM). We derive the identifiability of a general class of MCWM. We estimate the proposed model through Expectation-Maximization (EM) algorithm via an iteratively reweighted least squares (EM-IRLS) and Stochastic Gradient Descent (EM-SGD). Model selection is carried out using different information criteria. Various Adju...

Find SimilarView on arXiv

BayesBinMix: an R Package for Model Based Clustering of Multivariate Binary Data

September 22, 2016

89% Match
Panagiotis Papastamoulis, Magnus Rattray
Computation

The BayesBinMix package offers a Bayesian framework for clustering binary data with or without missing values by fitting mixtures of multivariate Bernoulli distributions with an unknown number of components. It allows the joint estimation of the number of clusters and model parameters using Markov chain Monte Carlo sampling. Heated chains are run in parallel and accelerate the convergence to the target posterior distribution. Identifiability issues are addressed by implementi...

Find SimilarView on arXiv

Clustering Multivariate Data using Factor Analytic Bayesian Mixtures with an Unknown Number of Components

June 2, 2019

89% Match
Panagiotis Papastamoulis
Methodology
Computation

Recent work on overfitting Bayesian mixtures of distributions offers a powerful framework for clustering multivariate data using a latent Gaussian model which resembles the factor analysis model. The flexibility provided by overfitting mixture models yields a simple and efficient way in order to estimate the unknown number of clusters and model parameters by Markov chain Monte Carlo (MCMC) sampling. The present study extends this approach by considering a set of eight paramet...

Find SimilarView on arXiv

Overfitting Bayesian Mixture Models with an Unknown Number of Components

February 18, 2015

89% Match
Havre Zoe Universite Paris-Dauphine & QUT, Brisbane van, Nicole QUT, Brisbane White, ... , Mengersen Kerrie QUT, Brisbane
Methodology

This paper proposes solutions to three issues pertaining to the estimation of finite mixture models with an unknown number of components: the non-identifiability induced by overfitting the number of components, the mixing limitations of standard Markov Chain Monte Carlo (MCMC) sampling techniques, and the related label switching problem. An overfitting approach is used to estimate the number of components in a finite mixture model via a Zmix algorithm. Zmix provides a bridge ...

Find SimilarView on arXiv

Non-parametric Multi-Partitions Clustering

January 6, 2023

89% Match
Chaumaray Marie du Roy de, Vincent Vandewalle
Methodology

In the framework of model-based clustering, a model, called multi-partitions clustering, allowing several latent class variables has been proposed. This model assumes that the distribution of the observed data can be factorized into several independent blocks of variables, each block following its own mixture model. In this paper, we assume that each block follows a non parametric latent class model, {\it i.e.} independence of the variables in each component of the mixture wi...

Find SimilarView on arXiv

Inference and Evaluation of the Multinomial Mixture Model for Text Clustering

June 14, 2006

89% Match
Loïs TSI Rigouste, Olivier TSI Cappé, François TSI Yvon
Information Retrieval
Computation and Language

In this article, we investigate the use of a probabilistic model for unsupervised clustering in text collections. Unsupervised clustering has become a basic module for many intelligent text processing applications, such as information retrieval, text classification or information extraction. The model considered in this contribution consists of a mixture of multinomial distributions over the word counts, each component corresponding to a different theme. We present and contra...

Find SimilarView on arXiv

Identifying Mixtures of Mixtures Using Bayesian Estimation

February 23, 2015

89% Match
Gertraud Malsiner-Walli, Sylvia Frühwirth-Schnatter, Bettina Grün
Methodology

The use of a finite mixture of normal distributions in model-based clustering allows to capture non-Gaussian data clusters. However, identifying the clusters from the normal components is challenging and in general either achieved by imposing constraints on the model or by using post-processing procedures. Within the Bayesian framework we propose a different approach based on sparse finite mixtures to achieve identifiability. We specify a hierarchical prior where the hyperpar...

Find SimilarView on arXiv

A survey on Bayesian inference for Gaussian mixture model

August 20, 2021

89% Match
Jun Lu
Machine Learning
Artificial Intelligence
Machine Learning

Clustering has become a core technology in machine learning, largely due to its application in the field of unsupervised learning, clustering, classification, and density estimation. A frequentist approach exists to hand clustering based on mixture model which is known as the EM algorithm where the parameters of the mixture model are usually estimated into a maximum likelihood estimation framework. Bayesian approach for finite and infinite Gaussian mixture model generates poi...

Find SimilarView on arXiv