Here is a question I raised in my Feynman Notebook:
- Will we learn about Gaussian Processes/Neural Networks in this course?
- This is a type of Bayesian Non-parametric and we don’t cover these in the specialization. However Abel Rodriguez, the instructor of the third course on mixture model has a short course
- Herbert Lee wrote a Bayesian Nonparametrics via Neural Networks on the subject.
- Athanasios Kottas of UCSC has made the following notes available on his website:
- Tutorial on Nonparametric Bayesian density regression: modeling methods and applications
- Short course on Applied Bayesian Nonparametric Mixture Modeling with references (16 pages) with 52 of them being his own papers.
This is not available as a course on Coursera and isn’t a part of the specialization which ended in the last course. So this notes are my own personal notes gathered from tutorials and courses I found on the web. - Tamara Broderick’s Gaussian Processes for Regression tutorials from - 2025 slides video and code - 2024 slides - Tamara Broderick
Overview
- In this course we learn to:
- We will build the following skills:
- Probability Distribution (Dirichlet, Beta)
- Bayesian Statistics
- There are currently five modules planned in this course:
- Gaussian Processes for Regression: We will focus on Gaussian processes as a flexible prior distribution for regression problems, allowing us to capture complex relationships in the data.
- Dirichlet process: We will explore the Dirichlet process as a prior distribution over probability measures, allowing for flexible modeling of unknown distributions.
- The Beta distribution
- The Dirichlet Distribution
- Dirichlet process
- Polya urn scheme
- Stick breaking representation
- Dirichlet process mixture models
- Hierarchical Dirichlet processes
- Chinese restaurant process: We will introduce the Chinese restaurant process as a metaphor for the Dirichlet process, providing an intuitive understanding of how it works.
- Indian buffet process: We will discuss the Indian buffet process as a model for representing the distribution of features in a dataset, allowing for flexible and scalable modeling of complex data structures.
- Polya tree: We will explore the Polya tree as a nonparametric prior distribution for modeling probability distributions, allowing for flexible and adaptive modeling of complex data structures.
Prerequisite skill checklist 🗒️
NotePrerequisite skill checklist
Bayesian Statistics
Mixture Models
Time Series Analysis
Some References:
- Gaussian Processes Rasmussen and Williams (2006)
- Surrogates Gramacy (2020)
- A Tutorial on Bayesian Optimization Frazier (2018)
- It builds a surrogate for the objective and quantifies the uncertainty in that surrogate using a Bayesian machine learning technique, Gaussian process regression, and then uses an acquisition function defined from this surrogate to decide where to sample
- three common acquisition functions:
- expected improvement
- entropy search
- knowledge gradient