McNicholas, Paul D.
Keep It Light! Simplifying Image Clustering Via Text-Free Adapters
Li, Yicen, Borde, Haitz Sáez de Ocáriz, Kratsios, Anastasis, McNicholas, Paul D.
Many competitive clustering pipelines have a multi-modal design, leveraging large language models (LLMs) or other text encoders, and text-image pairs, which are often unavailable in real-world downstream applications. Additionally, such frameworks are generally complicated to train and require substantial computational resources, making widespread adoption challenging. In this work, we show that in deep clustering, competitive performance with more complex state-of-the-art methods can be achieved using a text-free and highly simplified training pipeline. In particular, our approach, Simple Clustering via Pre-trained models (SCP), trains only a small cluster head while leveraging pre-trained vision model feature representations and positive data pairs. Experiments on benchmark datasets including CIFAR-10, CIFAR-20, CIFAR-100, STL-10, ImageNet-10, and ImageNet-Dogs, demonstrate that SCP achieves highly competitive performance. Furthermore, we provide a theoretical result explaining why, at least under ideal conditions, additional text-based embeddings may not be necessary to achieve strong clustering performance in vision.
Finite Mixtures of Multivariate Poisson-Log Normal Factor Analyzers for Clustering Count Data
Payne, Andrea, Silva, Anjali, Rothstein, Steven J., McNicholas, Paul D., Subedi, Sanjeena
A mixture of multivariate Poisson-log normal factor analyzers is introduced by imposing constraints on the covariance matrix, which resulted in flexible models for clustering purposes. In particular, a class of eight parsimonious mixture models based on the mixtures of factor analyzers model are introduced. Variational Gaussian approximation is used for parameter estimation, and information criteria are used for model selection. The proposed models are explored in the context of clustering discrete data arising from RNA sequencing studies. Using real and simulated data, the models are shown to give favourable clustering performance. The GitHub R package for this work is available at https://github.com/anjalisilva/mixMPLNFA and is released under the open-source MIT license.
Clustering Three-Way Data with Outliers
Clark, Katharine M., McNicholas, Paul D.
Matrix-variate normal mixture models are a powerful statistical tool used to represent complex data structures that involve matrices, such as multivariate time series, spatial data, and image data. Detecting outliers in matrix-variate normal mixture models is crucial for identifying anomalous observations that deviate significantly from the underlying distribution. Outliers can provide valuable insights into data quality issues, anomalies, or unexpected patterns. Outliers, and their treatment, is a long-studied topic in the field of applied statistics. The problem of handling outliers in multivariate clustering has been studied in several contexts including work by García-Escudero et al. (2008), Punzo and McNicholas (2016), Punzo et al. (2020), and Clark and McNicholas (2023).
Clustering and Semi-Supervised Classification for Clickstream Data via Mixture Models
Gallaugher, Michael P. B., McNicholas, Paul D.
Finite mixture models have been used for unsupervised learning for some time, and their use within the semi-supervised paradigm is becoming more commonplace. Clickstream data is one of the various emerging data types that demands particular attention because there is a notable paucity of statistical learning approaches currently available. A mixture of first-order continuous time Markov models is introduced for unsupervised and semi-supervised learning of clickstream data. This approach assumes continuous time, which distinguishes it from existing mixture model-based approaches; practically, this allows account to be taken of the amount of time each user spends on each webpage. The approach is evaluated, and compared to the discrete time approach, using simulated and real data.
Clustering Higher Order Data: Finite Mixtures of Multidimensional Arrays
Tait, Peter A., McNicholas, Paul D.
There have been many examples of clustering multivariate (i.e., two-way) data using finite mixture models (see, e.g., reviews by Fraley and Raftery, 2002; Bouveyron and Brunet-Saumard, 2014; McNicholas, 2016b). More recently, there have been some notable examples of clustering threeway data using finite mixtures of matrix-variate distributions (e.g., Viroli, 2011; Anderlucci et al., 2015; Gallaugher and McNicholas, 2018a). This work on clustering three-way data is timely in the sense that the variety of data that require clustering continues to increase. Furthermore, there is no reason to believe that this need ends with three-way data. An approach for clustering multi-way data is introduced based on a finite mixture of multidimensional arrays. While some might refer to such structures as'tensors', and so write about clustering tensor-variate data, we prefer the nomenclature multidimensional array to avoid confusion with the term'tensor' as used in engineering and physics, e.g., tensor fields.
Using Subset Log-Likelihoods to Trim Outliers in Gaussian Mixture Models
Clark, Katharine M., McNicholas, Paul D.
Mixtures of Gaussian distributions are a popular choice in model-based clustering. Outliers can affect parameters estimation and, as such, must be accounted for. Algorithms such as TCLUST discern the most likely outliers, but only when the proportion of outlying points is known \textit{a priori}. It is proved that, for a finite Gaussian mixture model, the log-likelihoods of the subset models are beta-distributed. An algorithm is then proposed that predicts the proportion of outliers by measuring the adherence of a set of subset log-likelihoods to a beta reference distribution. This algorithm removes the least likely points, which are deemed outliers, until model assumptions are met.
Flexible Clustering with a Sparse Mixture of Generalized Hyperbolic Distributions
Gallaugher, Michael P. B., Tang, Yang, McNicholas, Paul D.
Robust clustering of high-dimensional data is an important topic because, in many practical situations, real data sets are heavy-tailed and/or asymmetric. Moreover, traditional model-based clustering often fails for high dimensional data due to the number of free covariance parameters. A parametrization of the component scale matrices for the mixture of generalized hyperbolic distributions is proposed by including a penalty term in the likelihood constraining the parameters resulting in a flexible model for high dimensional data and a meaningful interpretation. An analytically feasible EM algorithm is developed by placing a gamma-Lasso penalty constraining the concentration matrix. The proposed methodology is investigated through simulation studies and two real data sets.
Clustering Discrete Valued Time Series
Roick, Tyler, Karlis, Dimitris, McNicholas, Paul D.
There is a need for the development of models that are able to account for discreteness in data, along with its time series properties and correlation. Our focus falls on INteger-valued AutoRegressive (INAR) type models. The INAR type models can be used in conjunction with existing model-based clustering techniques to cluster discrete valued time series data. With the use of a finite mixture model, several existing techniques such as the selection of the number of clusters, estimation using expectation-maximization and model selection are applicable. The proposed model is then demonstrated on real data to illustrate its clustering applications.
Detecting British Columbia Coastal Rainfall Patterns by Clustering Gaussian Processes
Paton, Forrest, McNicholas, Paul D.
Functional data analysis is a statistical framework where data are assumed to follow some functional form. This method of analysis is commonly applied to time series data, where time, measured continuously or in discrete intervals, serves as the location for a function's value. Gaussian processes are a generalization of the multivariate normal distribution to function space and, in this paper, they are used to shed light on coastal rainfall patterns in British Columbia (BC). Specifically, this work addressed the question over how one should carry out an exploratory cluster analysis for the BC, or any similar, coastal rainfall data. An approach is developed for clustering multiple processes observed on a comparable interval, based on how similar their underlying covariance kernel is. This approach provides significant insights into the BC data, and these insights can be described in terms of El Nino and La Nina; however, the result is not simply one cluster representing El Nino years and another for La Nina years. From one perspective, the results show that clustering annual rainfall can potentially be used to identify extreme weather patterns.
An Evolutionary Algorithm with Crossover and Mutation for Model-Based Clustering
McNicholas, Sharon M., McNicholas, Paul D., Ashlock, Daniel A.
The expectation-maximization (EM) algorithm is almost ubiquitous for parameter estimation in model-based clustering problems; however, it can become stuck at local maxima, due to its single path, monotonic nature. Rather than using an EM algorithm, an evolutionary algorithm (EA) is developed. This EA facilitates a different search of the fitness landscape, i.e., the likelihood surface, utilizing both crossover and mutation. Furthermore, this EA represents an efficient approach to "hard" model-based clustering and so it can be viewed as a sort of generalization of the k-means algorithm, which is itself equivalent to a classification EM algorithm for a Gaussian mixture model with spherical component covariances. The EA is illustrated on several data sets, and its performance is compared to k-means clustering as well as model-based clustering with an EM algorithm.