Lecture 24. Subject Review and Exam Info#
This note is completed with the assistance of ChatGPT
2023 S2 Exam Scope#
Stats Background (Lectures 1 and 2)#
- Basics of Statistics & Probability: Descriptive stats, inferential stats, probability axioms, conditional probability. 
- Bayes’ Theorem: 
Linear Regression (Lecture 3)#
- Model: \( y = \beta_0 + \beta_1 x \) 
- MLE (Maximum Likelihood Estimation): Estimation method to find the parameters that maximize the likelihood of the observed data. 
Regularising Linear Regression (Lecture 5)#
- Ridge Regression: Adds L2 penalty to linear regression. 
- Lasso Regression: Adds L1 penalty to linear regression. 
- Bayesian MAP (Maximum A Posteriori): Mode of the posterior distribution. 
Non-linear Regression & Bias-Variance (Lecture 5)#
- Bias-Variance Tradeoff: Total Error = Bias^2 + Variance + Irreducible Error. 
PAC Learning (Lecture 7)#
- PAC (Probably Approximately Correct) Learning: Framework for mathematical analysis of machine learning. 
- VC (Vapnik-Chervonenkis) Dimension: Measure of the capacity of a hypothesis class. 
SVMs (Support Vector Machines) (Lecture 9)#
- Kernel Trick: Efficiently compute dot products in high-dimensional spaces. 
- Popular Kernels: Linear, Polynomial, RBF (Radial Basis Function), Sigmoid. 
- Mercer’s Theorem: Condition for a function to be a valid kernel. 
Neural Networks (Lectures 11-14)#
- Basics: Neurons, activation functions (ReLU, Sigmoid, Tanh). 
- Backpropagation: Algorithm to update network weights using gradients. 
- Autoencoders: Neural networks trained to reproduce their input. 
- CNN (Convolutional Neural Networks): Specialized for grid-like data (e.g., images). 
- RNN (Recurrent Neural Networks): Suited for sequence data (e.g., time series). 
PGMs (Probabilistic Graphical Models) (Lectures 20-22)#
- Directed PGMs: Represent conditional dependencies (e.g., Bayesian Networks). 
- Undirected PGMs: Represent symmetric relationships (e.g., Markov Random Fields). 
- Inference in PGMs: Compute conditional marginals from joint distributions. 
- Exact Inference: Algorithms like variable elimination. 
- Approximate Inference: Techniques like sampling. 
- Parameter Estimation: MLE, EM (Expectation Maximization) for latent variables. 
Application (Project 1)#
- Review Key Steps: Data preprocessing, model selection, evaluation metrics, results, and conclusions. 
Given the extensive topics, it’s essential to have a deeper understanding of each, especially the critical concepts. This cheatsheet provides a quick reference, but detailed notes, practice problems, and examples will help in mastering the material.
