Probabilistic Modelling Tutorial Series

Self-Directed Tutorial Series, GitHub (Open Access), 2022

Developed and published an open-access tutorial series covering probabilistic modelling techniques, aimed at graduate students and researchers in quantitative fields.

Slides and Code on GitHub

Topics Covered

Probabilistic Programming in PyMC3

  • Fitting a Gaussian distribution
  • Bayesian Linear Regression
  • Changepoint modelling
    • Poisson changepoint
    • Bernoulli changepoint
    • Gaussian changepoint for mean and variance
    • Advanced: multi-changepoint model with mixture emissions for repeated timeseries

Gaussian Mixture Models and Hidden Markov Models

  • Unsupervised clustering using GMMs
  • Introduction to Hidden Markov Models

Additional Resources

Handwritten Analytical Derivations

A recurring frustration with textbooks is that they assume too much about how intuitive each step in a derivation is for the reader — often skipping over algebraic manipulations that are non-obvious, especially to those newer to the material. These handwritten derivations are intended to provide an exhaustive, step-by-step breakdown that is (hopefully) easy to follow, with no steps omitted.

DocumentDescription
Gaussian MLE (1D)Full derivation of the maximum likelihood estimates for the mean and variance of a 1D Gaussian distribution, starting from the log-likelihood and working through each calculus step explicitly.
Linear Regression MLEStep-by-step derivation of the ordinary least squares solution via maximum likelihood estimation, including the matrix calculus needed to arrive at the normal equations.
GMM Expectation-MaximizationDetailed derivation of the EM algorithm for Gaussian Mixture Models, covering the E-step (responsibility computation) and M-step (parameter updates) with all intermediate algebra shown.