APPA 5002: Discrete-Time Markov Chains and Monte Carlo Methods
ÌýPreview this courseÌýin the non-credit experience today!Ìý
Start working toward program admission and requirements right away.ÌýWork you complete in the non-credit experience will transfer to the for-credit experience when you upgrade and pay tuition. See How It Works for details.
- Course Type: Pathway | Breadth
- Specialization: Foundations of Probability and Statistics
- Instructor:ÌýDr.ÌýJem CorcoranÌý
- Prior knowledge needed:
- Programming languages: Intro to R programming
- Math: Calculus 1 and 2
- Technical requirements:ÌýÌý
Learning Outcomes
- Analyze long-term behavior of Markov processes for the purposes of both prediction and understanding equilibrium in dynamic stochastic systems
- Apply Markov decision processes to solve problems involving uncertainty and sequential decision-making
- Simulate data from complex probability distributions using Markov chain Monte Carlo algorithms
Course Grading Policy
Assignment | Percentage of Grade | AI Usage Policy |
---|---|---|
Ungraded Labs | 0% | Limited |
Graded Assignments | 60% (5 assignments, 12% each) | Limited |
Final Exam | 40% | No AI |
Course Content
Duration: 1 Hour
Welcome to the course! This module contains logistical information to get you started! Seven readings, 4 ungraded labs.
Duration: 6 hours
In this module we will review definitions and basic computations of conditional probabilities. We will then define a Markov chain and its associated transition probability matrix and learn how to do many basic calculations. We will then tackle more advanced calculations involving absorbing states and techniques for putting a longer history into a Markov framework!
Duration: 6 hours
What happens if you run a Markov chain out for a "very long time"? In many cases, it turns out that the chain will settle into a sort of "equilibrium" or "limiting distribution" where you will find it in various states with various fixed probabilities. In this Module, we will define communication classes, recurrence, and periodicity properties for Markov chains with the ultimate goal of being able to answer existence and uniqueness questions about limiting distributions!
Duration: 6 hours
In this Module, we will define what is meant by a "stationary" distribution for a Markov chain. You will learn how it relates to the limiting distribution discussed in the previous Module. You will also spend time learning about the very powerful "first-step analysis" technique for solving many, otherwise intractable, problems of interest surrounding Markov chains. We will discuss rates of convergence for a Markov chain to settle into its "stationary mode", and just maybe we'll give a monkey a keyboard and hope for the best!
Duration: 7 hours
In this Module we explore several options for simulating values from discrete and continuous distributions. Several of the algorithms we consider will involve creating a Markov chain with a stationary or limiting distribution that is equivalent to the "target" distribution of interest. This Module includes the inverse cdf method, the accept-reject algorithm, the Metropolis-Hastings algorithm, the Gibbs sampler, and a brief introduction to "perfect sampling".
Duration: 3 hours
In reinforcement learning, an "agent" learns to make decisions in an environment through receiving rewards or punishments for taking various actions. A Markov decision process (MDP) is reinforcement learning where, given the current state of the environment and the agent's current action, past states and actions used to get the agent to that point are irrelevant. In this Module, we learn about the famous "Bellman equation", which is used to recursively assign rewards to various states and how to use it in order to find an optimal strategy for the agent!
Duration: 3 hour
Final Exam Format: Proctored Exam
The course final examination is similar to the quizzes you have taken in this course. There are a few differences:
- The exam is proctored. You will need to arrange for a time to take the proctored exam.
- It is a three hour exam.
- You will have two attempts. You may submit your answers unlimited times within each three-hour attempt.
Notes
- Cross-listed Courses: CoursesÌýthat are offered under two or more programs. Considered equivalent when evaluating progress toward degree requirements. You may not earn credit for more than one version of a cross-listed course.
- Page Updates: This page is periodically updated. Course information on the Coursera platform supersedes the information on this page. Click theÌýView on CourseraÌýbuttonÌýabove for the most up-to-date information.