Lecture 22. Inference on PGMs Cont. & Lecture 23. Gaussian Mixture Models#
This note is completed with the assistance of ChatGPT
Lecture Summary: Gaussian Mixture Model (GMM) in Statistical Machine Learning#
1. Introduction to Unsupervised Learning:#
- Definition: Learning the structure of data without labels. It contrasts with supervised learning where data comes with predefined labels. 
- Main Paradigms: Unsupervised learning is a shift from supervised learning, which is about predicting labels, and the contextual bandits setting, which deals with partial supervision. 
2. Kinds of Unsupervised Learning:#
- Tasks: - Clustering: Grouping data points based on their similarities. 
- Dimensionality Reduction: Reducing the number of random variables. 
- Learning Probabilistic Models: Estimating parameters for statistical models. 
 
- Applications: Include market basket analysis, outlier detection, and various tasks within supervised machine learning pipelines. 
3. Refresher on K-means Clustering:#
- Steps: - Initialize cluster centroids. 
- Assign data points to the nearest centroid and recompute centroids. 
- Terminate when no change in assignments or continue iterating. 
 
- Features: Requires specifying the number of clusters in advance, uses Euclidean distance for dissimilarity, and typically finds spherical clusters. 
4. Gaussian Mixture Model (GMM):#
- Concept: A probabilistic approach to clustering, where each data point can belong to multiple clusters with certain probabilities. 
- Benefits: Allows modeling uncertainty about the origin of each data point. Each point originates from a particular cluster, but we aren’t sure which one. 
- Application: Clustering in a GMM becomes model fitting in a probabilistic sense. 
5. Clustering with a Probabilistic Model:#
- Model: Treats data points as i.i.d. samples from a mixture of distributions. 
- GMM: When the components in the mixture are Gaussian distributions, we have a Gaussian Mixture Model. 
- Normal Distribution: A key mathematical concept, with the 1D Gaussian given by the bell curve and the multi-dimensional Gaussian defined by a mean vector and a covariance matrix. 
6. Key Components of GMM:#
- Cluster Assignment Probabilities: Represents the likelihood a data point belongs to a particular cluster. 
- Location of Point: Governed by the Gaussian distribution of the cluster it’s assigned to. 
- Model Parameters: Cluster weights, means, and covariance matrices. 
- Mixture Distribution: Obtained by marginalizing latent variables and represents the likelihood of observed data points. 
Final Takeaway:#
GMM provides a nuanced probabilistic approach to clustering. Instead of hard assignments as in k-means, GMM gives soft assignments, making it more flexible and expressive for capturing uncertainties in real-world data structures.
