Machine Learning

Part I: Mathematical Foundations

Every machine learning algorithm rests on three mathematical pillars: linear algebra provides the geometry of data and transformations; probability theory formalises uncertainty and inference; optimisation theory explains how models improve. This part builds each pillar from first principles, with full derivations and Python simulations.

What you will learn

โœ“Represent data as vectors and matrices and reason geometrically
โœ“Decompose matrices with eigendecomposition and SVD for compression and analysis
โœ“Model uncertainty with probability distributions and derive MLE/MAP estimators
โœ“Apply Bayesโ€™ theorem to update beliefs as data arrives
โœ“Prove gradient descent converges on convex objectives
โœ“Derive the Adam optimiser from first principles
โœ“Formulate constrained optimisation with Lagrange multipliers and KKT conditions
โœ“Understand every Part IIโ€“VII algorithm through these three lenses