Part II
The mathematical backbone of modern machine learning — from vector spaces and calculus through probability, information theory, convex optimization, and the functional analysis underlying deep learning. Treated rigorously, with ML motivation throughout.
Vector spaces, spectral theory, matrix decompositions, and the geometric machinery behind every ML model — developed from first principles.
Gradients, Hessians, manifolds, and Riemannian geometry — the geometric foundation of learning.
Duality, KKT conditions, proximal methods, and the geometry of non-convex loss landscapes.
Measure-theoretic probability, distributions, convergence modes, limit theorems, and concentration inequalities.
MLE, Bayesian inference, PAC learning, VC dimension, and the theory behind why models generalize.
Entropy, mutual information, KL divergence, channel capacity, and the deep connection between compression and intelligence.
Markov chains, martingales, Brownian motion, Itô calculus, and stochastic differential equations.
Floating point, direct and iterative solvers, automatic differentiation, and numerical stability in practice.
Linear programming, integer programming, network flows, combinatorial optimization, and dynamic programming.
Nash equilibria, minimax, mechanism design, and the multi-agent foundations of GANs and RLHF.
Hilbert and Banach spaces, bounded operators, RKHS, and the operator-theoretic view of attention and kernels.
Spectral graph theory, random graphs, the probabilistic method, and the combinatorial foundations of GNNs.