Machine Learning

Part V: Probabilistic ML

Deterministic models give a single answer; probabilistic models give a distribution over answers, quantifying what they know and what they don't. This part builds three pillars of probabilistic machine learning: Bayesian inference for updating beliefs, Gaussian processes for non-parametric function learning, and variational inference for scalable approximation. Together they form the foundation of modern probabilistic deep learning.

What you will learn

Derive the posterior distribution for Gaussian and Bernoulli likelihoods
Understand why conjugate priors make Bayesian updating exact
Implement Metropolis-Hastings MCMC for arbitrary posteriors
Define a Gaussian process through its mean and covariance function
Derive the GP posterior predictive mean and variance from scratch
Choose and tune kernel hyperparameters via marginal likelihood
Derive the ELBO and show it lower-bounds the log-evidence
Connect variational inference to the VAE objective of Chapter 12

Prerequisites

Part I: Probability
Bayes' theorem, Gaussian distribution, MLE and MAP estimation
Part I: Linear Algebra
Matrix inversion, positive definite matrices, multivariate Gaussians
Part IV: VAEs (Ch 12)
The reparameterisation trick and the VAE ELBO (revisited in Ch 15)