Part V: Probabilistic ML
Deterministic models give a single answer; probabilistic models give a distribution over answers, quantifying what they know and what they don't. This part builds three pillars of probabilistic machine learning: Bayesian inference for updating beliefs, Gaussian processes for non-parametric function learning, and variational inference for scalable approximation. Together they form the foundation of modern probabilistic deep learning.
Chapter 13: Bayesian Inference
Bayes’ theorem as a learning rule: prior beliefs updated by data to form posteriors. Conjugate priors, MCMC, and the predictive distribution.
Chapter 14: Gaussian Processes
A distribution over functions defined by a mean function and kernel. Full GP regression derivation with posterior predictive and marginal likelihood.
Chapter 15: Variational Inference
When posteriors are intractable, optimise a simpler distribution to approximate them. Full ELBO derivation, mean-field VI, and the VAE connection.