Bayesian Non-Parametric Models

Bayesian Statistics - Capstone Project

Capstone Project: Introduction
Bayesian Statistics
Capstone Project
Author

Oren Bochman

Published

July 2, 2025

Keywords

Time Series

Here is a question I raised in my Feynman Notebook:

  1. Will we learn about Gaussian Processes/Neural Networks in this course?

This is not available as a course on Coursera and isn’t a part of the specialization which ended in the last course. So this notes are my own personal notes gathered from tutorials and courses I found on the web. - Tamara Broderick’s Gaussian Processes for Regression tutorials from - 2025 slides video and code - 2024 slides - Tamara Broderick

Overview

  • In this course we learn to:
  • We will build the following skills:
    • Probability Distribution (Dirichlet, Beta)
    • Bayesian Statistics
  • There are currently five modules planned in this course:
    1. Gaussian Processes for Regression: We will focus on Gaussian processes as a flexible prior distribution for regression problems, allowing us to capture complex relationships in the data.
      1. Gaussian process model slides
      2. Gaussian process regression slides
      3. Squared exponential kernel and observation noise slides
      4. What uncertainty are we quantifying? slides
      5. A list of resources: slide
    2. Dirichlet process: We will explore the Dirichlet process as a prior distribution over probability measures, allowing for flexible modeling of unknown distributions.
    • The Beta distribution
    • The Dirichlet Distribution
    • Dirichlet process
    • Polya urn scheme
    • Stick breaking representation
    • Dirichlet process mixture models
    • Hierarchical Dirichlet processes
    1. Chinese restaurant process: We will introduce the Chinese restaurant process as a metaphor for the Dirichlet process, providing an intuitive understanding of how it works.
    2. Indian buffet process: We will discuss the Indian buffet process as a model for representing the distribution of features in a dataset, allowing for flexible and scalable modeling of complex data structures.
    3. Polya tree: We will explore the Polya tree as a nonparametric prior distribution for modeling probability distributions, allowing for flexible and adaptive modeling of complex data structures.

Prerequisite skill checklist 🗒️

Bayesian Statistics

Mixture Models

Time Series Analysis

Some References:

  1. Gaussian Processes Rasmussen and Williams (2006)
  2. Surrogates Gramacy (2020)
  3. A Tutorial on Bayesian Optimization Frazier (2018)
  • It builds a surrogate for the objective and quantifies the uncertainty in that surrogate using a Bayesian machine learning technique, Gaussian process regression, and then uses an acquisition function defined from this surrogate to decide where to sample
  • three common acquisition functions:
    • expected improvement
    • entropy search
    • knowledge gradient