Probabilistic Machine Learning
CS772A
2022-23 (odd semester)


Instructor: Piyush Rai (Office RM-502, email: piyush AT cse DOT iitk DOT ac DOT in)
Office Hours: Wed 16:00-17:00 or by appointment (RM-502)
Class Venue: KD-101
Class Timings: Mon/Thur 18:00 - 19:30
Q/A Forum: Piazza
TAs and email: Abhinav Joshi (ajoshi AT cse), Abhishek Jaiswal (abhijais AT cse), Avideep Mukherjee (avideep AT cse), Soumya Banerjee (soumyab AT cse)
TA Office Hours:
Abhinav: 9:00-10:00 Saturday (RM-511)
Abhishek: Friday 11:30 - 12:30 (KD-108)
Avideep: Tuesday 12:00-13:00 (RM-504)
Soumya: Monday 11:00-12:00 (RM-504)

Background and Course Description

Probabilistic models for data are ubiquitous in many areas of science and engineering, and specific domains such as visual and language understanding, finance, healthcare, biology, climate informatics, etc. This course will be an advanced introduction to probabilistic models of data (often through case studies from these domains) and a deep-dive into advanced inference and optimization methods used to learn such probabilistic models. This is an advanced course and ideally suited for student who are doing research in this area or are interested in doing research in this area.

Pre-requisites

Instructor's consent. The course expects students to have a strong prior background in machine learning (ideally through formal coursework, such as CS771), and ideally also some prior exposure/appreciation to basic principles of probabilistic modeling. The students are expected to have strong foundations in probability and statistics, linear algebra, and optimization, and must also be proficient in programming in Python.

Grading

(Tentative) There will be 4 homeworks (40%) which will consist of a mix of written and programming questions, 2 quizzes (5% each), a mid-sem exam (20%), and an end-sem exam (30%).

Reference materials

There will not be any dedicated textbook for this course. We will use lecture slides/notes, monographs, tutorials, and papers for the topics that will be covered in this course. Some recommended, although not required, reference books are listed below:

Schedule

Lec. No. Date Topics Readings/References/Comments Slides/Notes
1 Aug 1 Course Logistics, Intro to Probabilistic Machine Learning A review article on PML and AI, Probability and statistics refresher slides, PML-2 Sec 3 - 3.1.5.2 PPTX slides, PDF slides
2 Aug 4 Basics of Parameter Estimation in Probabilistic Models PML-2 Section 3.2-3.3, Additional slides on parameter estimation (through a simple example): PPTX, PDF, PML-2 Sec 3 - 3.1.5.2 PPTX slides, PDF slides
3 Aug 11 Parameter Estimation in Probabilistic Models: Examples (Contd) PML-2 Section 3.2-3.3, Wikipedia entry on Dirichlet distribution PPTX slides, PDF slides
4 Aug 13 Parameter Estimation for Gaussians, Probabilistic Linear Regression PML-2 Section 3.2-3.3, Bayesian Inference for Gaussians, PRML 3.3, Recommended Readings: Bayesian Inference tutorial (with Bayesian linear regression as case study; as of now, may read up to section 3) PPTX slides, PDF slides
5 Aug 18 Probabilistic Linear Regression (contd), Exponential Family Distributions PRML 3.3, Recommended Readings: Bayesian Inference tutorial (with Bayesian linear regression as case study; as of now, may read up to section 3), PML-2 Section 2.3 and 3.4 PPTX slides, PDF slides
6 Aug 22 Exp. Family (contd), Logistic/Softmax Regression PML-2 Sec 2.3, PML-1 Sec 2.5, PML-2 Chapter 10 (Sec 10.5 for Bayesian Logistic Regression) PPTX slides, PDF slides
7 Aug 29 Laplace Approximation, Generalized Linear Models PML-2 Sec 7.4.3, PML-2 Chapter 12 PPTX slides, PDF slides
8 Sept 1 Generative Models for Supervised Learning PML-1 Chapter 9 PPTX slides, PDF slides
9 Sept 3 Gaussian Processes (GP) PML-1 Sec 17.2 PPTX slides, PDF slides
10 Sept 5 GP wrap-up, Inference in multi-parameter models, conditional posteriors, local conjugacy Readings listed on the slides (especially the paper on Bayesian matrix factorization) PPTX slides, PDF slides
11 Sept 8 Latent variable models and the Expectation Maximization algorithm PRML Sec 9.3 and 9.4 PPTX slides, PDF slides
12 Sept 12 Latent variable models and the Expectation Maximization algorithm (Contd) PRML Sec 9.3 and 9.4 PPTX slides, PDF slides
13 Sept 15 Variational Inference PRML 10.1,10.2,10.3.10.4, Life after EM (shows the connection between EM, variational EM, and variational inference, through several examples), VI: A Review for Statisticians PPTX slides, PDF slides
14 Sept 26 Variational Inference (Contd) Same readings as those for Lecture 13 PPTX slides, PDF slides
15 Sept 29 Variational Inference (Wrap-up) PML-2 Section 10.3.2 - 10.3.6 (optional readings: papers referenced on the slides) PPTX slides, PDF slides
16 Oct 10 Approximate Inference via Sampling PRML 11.1-11.3, Recommended: Intro to MCMC for Machine Learning, Monte Carlo for Absolute Beginners, Gibbs Sampling for the Uninitiated PPTX slides, PDF slides
17 Oct 13 Approx. Inference via Sampling (Contd): Metropolis Hastings and Gibbs Sampling Same readings as those for Lecture 16 PPTX slides, PDF slides
18 Oct 17 Approx. Inference via Sampling (Contd): MCMC with Gradients, Recent Advances Recommended: SGLD paper and other papers referenced in the slides, Survey paper on SGMCMC methods like SGLD and improvements, Patterns of Scalable Bayes (See sec 4.2 for parallel MCMC), No U-Turn Sampler (section 2 describes the basics of HMC), PPTX slides, PDF slides
19 Oct 20 Approx. Inference via Sampling (wrap-up), Bayesian Deep Learning PML-2 (Sec 17.3) and papers references on the slides PPTX slides, PDF slides
20 Oct 27 Bayesian Deep Learning (contd), (Shallow and Deep) Generative Models Classical generative models (Factor analysis and variants, topic models): PML-2 (Sec 28.3,28.4,28.5) PPTX slides, PDF slides
21 Oct 29 (Shallow/Classical and Deep) Generative Models Same readings as those for Lecture 20; Variational Auto-encoders (VAE): PML-2 Chapter 21 PPTX slides, PDF slides
22 Oct 31 Deep Generative Models VAE: PML-2 Chapter 21, GAN: PML-2 Chapter 26, Diffusion Models: PML-2 Chapter 25 PPTX slides, PDF slides
23 Nov 4 Active Learning and Bayesian Optimization Bayesian Active Learning, An old but classic paper on probabilistic/Bayesian approaches to active learning: Information-Based Objective Functions for Active Data Selection, Bayesian Optimization: PML-2 (Section 6.8), An introduction to Bayesian Optimization (with some code) PPTX slides, PDF slides
24 Nov 7 Other assorted topics in Probabilistic ML (1): Frequentist vs Bayesian Learning, Model Calibration Frequentist Learning: PML-2 Section 4.7; Model Calibration: PML-2 Section 14.2 PPTX slides, PDF slides
25 Nov 10 Other assorted topics in Probabilistic ML (2): Conformal Prediction, Nonparametric Bayesian Methods Conformal Prediction: PML-2 Section 14.3; Nonparametric Bayes: PML-2 Chapter 32; For NPBayes modeling general intro - this tutorial survey paper PPTX slides, PDF slides
26 Nov 14 Other assorted topics in Probabilistic ML (3): Nonparametric Bayesian Methods (contd), Probabilistic Models for Sequential Data, Probabilistic Numerics Same readings as those for Lecture 25 PPTX slides, PDF slides

Course Policies

Anti-cheating policy