# Bayesian Inference with MCMC

• Topics in Model Performance
• This module gives an overview of topics related to assessing the quality of models. While some of these metrics may be familiar to those with a Machine Learning background, the goal is to bring awareness to the concepts rooted in Information Theory. The course website is https://sjster.github.io/introduction_to_computational_statistics/docs/Production/BayesianInference.html. Instructions to download and run the notebooks are at https://sjster.github.io/introduction_to_computational_statistics/docs/Production/getting_started.html
• The Metropolis Algorithms for MCMC
• This module serves as a gentle introduction to Markov-Chain Monte Carlo methods. The general idea behind Markov chains are presented along with their role in sampling from distributions. The Metropolis and Metropolis-Hastings algorithms are introduced and implemented in Python to help illustrate their details. The course website is https://sjster.github.io/introduction_to_computational_statistics/docs/Production/MonteCarlo.html. Instructions to download and run the notebooks are at https://sjster.github.io/introduction_to_computational_statistics/docs/Production/getting_started.html
• Gibbs Sampling and Hamiltonian Monte Carlo Algorithms
• This module is a continuation of module 2 and introduces Gibbs sampling and the Hamiltonian Monte Carlo (HMC) algorithms for inferring distributions. The Gibbs sampler algorithm is illustrated in detail, while the HMC receives a more high-level treatment due to the complexity of the algorithm. Finally, some of the properties of MCMC algorithms are presented to set the stage for Course 3 which uses the popular probabilistic framework PyMC3. The course website is https://sjster.github.io/introduction_to_computational_statistics/docs/Production/MonteCarlo.html#gibbs-sampling.