SlideShare ist ein Scribd-Unternehmen logo
1 von 6
Downloaden Sie, um offline zu lesen
Appendix for Lecture 2: Monte Carlo Methods (Basics)
Dahua Lin
1 Justification of Basic Sampling Methods
Proposition 1. Let F be the cdf of a real-valued random variable with distribution D. Let
U ∼ Uniform([0, 1]), then F−1(U) ∼ D.
Proof. Let X = F−1(U). It suffices to show that the cdf of X is F. For any t ∈ R,
P(X ≤ t) = P(F−1
(U) ≤ t) = P(U ≤ F(t)) = F(t). (1)
Here, we utilize the fact that F is non-decreasing.
Proposition 2. Samples producted using Rejection sampling has the desired distribution.
Proof. Each iteration actually generate two random variables: x and u, where u ∈ {0, 1} is the
indicator of acceptance. The join distribution of x and u is given by
˜p(dx, u = 1) = a(u = 1|x)q(dx) =
p(x)
Mq(x)
q(x)dx =
p(x)
M
µ(dx). (2)
Here, a(u|x) is the conditional distribution of u on x, and µ is the base measure. On the other
hand, we have
Pr(u = 1) = ˜p(dx, u = 1) =
p(x)
M
µ(dx) =
1
M
. (3)
Thus, the resultant distribution is
˜p(dx|u = 1) =
˜p(dx, u = 1)
Pr(u = 1)
= p(x)µ(dx). (4)
This completes the proof.
2 Markov Chain Theory
Proposition 3. When the state space Ω is countable, we have
µ − ν TV =
1
2
x∈Ω
|µ(x) − ν(x)|. (5)
Proof. Let A = {x ∈ Ω : µ(x) ≥ nu(x)}. By definition, we have
µ − ν TV ≥ |µ(A) − ν(A)| = µ(A) − ν(A), (6)
µ − ν TV ≥ |µ(Ac
) − ν(Ac
)| = ν(Ac
) − µ(Ac
). (7)
We also have
µ(A) − ν(A) =
x∈A
µ(x) − ν(x) =
x∈A
|µ(x) − ν(x)|, (8)
ν(Ac
) − µ(Ac
) =
x∈Ac
ν(x) − µ(x) =
x∈Ac
|µ(x) − ν(x)|. (9)
1
Combining the equations above results in
µ − ν TV ≥
1
2
(µ(A) − ν(A) + ν(Ac
) − µ(Ac
))
=
1
2
x∈A
|µ(x) − ν(x)| +
x∈Ac
|µ(x) − ν(x)|
=
1
2
x∈Ω
|µ(x) − ν(x)|. (10)
Next we show the inequality of the other direction. For any A ⊂ Ω, we have
|µ(Ac
) − ν(Ac
)| = |(µ(Ω) − µ(A)) − (ν(Ω) − ν(A))| = |µ(A) − ν(A)| (11)
Hence,
|µ(A) − ν(A)| =
1
2
(|µ(A) − ν(A)| + |µ(Ac
) − ν(Ac
)|)
≤
1
2
x∈A
|µ(x) − ν(x)| +
x∈Ac
|µ(x) − ν(x)|
≤
1
2
x∈Ω
|µ(x) − ν(x)|. (12)
As A is arbitrary, we can conclude that
µ − ν TV sup
A
|µ(A) − ν(A)| ≤
1
2
x∈Ω
|µ(x) − ν(x)|. (13)
This completes the proof.
Proposition 4. The total variation distance (µ, ν) → µ − ν TV is a metric.
Proof. To show that it is a metric, we verify the four properties that a metric needs to satisfy
one by one.
1. µ − ν TV is non-negative, as |µ(A) − ν(A)| is always non-negative.
2. When µ = ν, |µ(A) − ν(A)| is always zero, and hence µ − ν TV = 0. On the other
hand, when µ = ν, there exists A ⊂ S such that |µ(A) − ν(A)| > 0, and therefore
µ − ν TV ≥ |µ(A) − ν(A)| > 0. Together we can conclude that µ − ν TV = 0 iff µ = ν.
3. µ−ν TV = ν −µ TV as |µ(A)−ν(A)| = |ν(A)−µ(A)| holds for any measurable subset
A.
4. Next, we show that the total variation distance satisfies the triangle inequality, as below.
Let µ, ν, η be three probability measures over Ω:
µ − ν TV = sup
A∈S
|µ(A) − ν(A)|
= sup
A∈S
|µ(A) − η(A) + η(A) − ν(A)|
≤ sup
A∈S
(|µ(A) − η(A)| + |η(A) − ν(A)|)
≤ sup
A∈S
|µ(A) − η(A)| + sup
A∈S
|η(A) − ν(A)|
= µ − η TV + η − ν TV . (14)
2
The proof is completed.
Proposition 5. Consider a Markov chain over a countable space Ω with transition proba-
bility matrix P. Let π be a probability measure over Ω that is in detailed balance with P,
i.e. π(x)P(x, y) = π(y)P(y, x), ∀x, y ∈ Ω. Then π is invariant to P, i.e. π = πP.
Proof. With the assumption of detailed balance, we have
(πP)(y) =
x∈Ω
π(x)P(x, y) =
x∈Ω
π(y)P(y, x) = π(y)
x∈Ω
P(y, x) = π(y). (15)
Hence, π = πP, or in other words, π is invariant to P.
Proposition 6. Let (Xt) be an ergodic Markov chain Markov(π, P) where π is in detailed
balance with P, then given arbitrary sequence x0, . . . , xn ∈ Ω, we have
Pr(X0 = x0, . . . , Xn = xn) = Pr(X0 = xn, . . . , Xn = x0). (16)
Proof. First, we have
Pr(X0 = x0, . . . , Xn = xn) = π(x0)P(x0, x1) · · · P(xn−1, xn). (17)
On the other hand, by detailed balance, we have P(x, y) = π(y)P(y,x)
π(x) , and thus
Pr(X0 = xn, . . . , Xn = x0) = π(xn)P(xn, xn−1) · · · P(x1, x0)
= π(xn)
π(xn−1)P(xn−1, xn)
π(xn)
· · ·
π(x0)P(x0, x1)
π(x1)
= P(xn−1, xn) · · · P(x0, x1)π(x0). (18)
Comparing Eq.(17) and Eq.(18) results in the equality that we intend to prove.
Proposition 7. Over a measurable space (Ω, S), if a stochastic kernel P is reversible w.r.t. π,
then π is invariant to P.
Proof. Let π = πP, it suffices to show that π (A) = π(A) for every A ∈ S under the reversibility
assumption. Given any A ∈ S, let fA(x, y) := 1(y ∈ A), then we have
π (A) = π(dx)P(x, A)
= π(dx) fA(x, y)P(x, dy)
= fA(x, y)π(dx)P(x, dy)
= fA(y, x)π(dx)P(x, dy)
= 1(x ∈ A)π(dx)P(x, dy)
= 1(x ∈ A)π(dx) P(x, dy)
= 1(x ∈ A)π(dx) = π(A). (19)
This completes the proof.
3
Proposition 8. Given a stochastic kernel P and a probability measure π over (Ω, S). Suppose
both Px and π are absolutely continuous w.r.t. a base measure µ, that is, π(dx) = π(x)µ(dx)
and P(x, dy) = Px(dy) = px(y)µ(dy), then P is reversible w.r.t. π if and only if
π(x)px(y) = π(y)py(x), a.e. (20)
Proof. First, assuming detailed balance, i.e. π(x)px(y) = π(y)py(x), a.e., we show reversibility.
f(x, y)π(dx)P(x, dy) = f(x, y)π(x)px(y)µ(dx)µ(dy)
= f(x, y)π(y)py(x)µ(dx)µ(dy) ...[detailed balance]
= f(y, x)π(x)px(y)µ(dx)µ(dy) ...[exchange variables]
= f(y, x)π(dx)P(x, dy). (21)
Next, we show the converse. The definition of reversibility implies that
f(x, y)π(dx)P(x, dy) = f(x, y)π(dy)P(y, dx) (22)
Hence,
f(x, y)π(x)px(y)µ(dx)µ(dy) = f(x, y)π(y)py(x)µ(dx)µ(dy) (23)
Hence, f(x, y)π(x)px(y) = f(x, y)π(y)py(x) a.e. for arbitrary integrable function f, which im-
plies that π(x)px(y) = π(y)py(x) a.e.
Proposition 9. Given a stochastic kernel P and a probability measure π over (Ω, S). If
P(x, dy) = m(x)Ix(dy) + px(y)µ(dy) and π(x)px(y) = π(y)py(x) a.e, then P is reversible
w.r.t. π.
Proof. Under the given conditions, we have
f(x, y)π(dx)P(x, dy) = f(x, y)π(dx) (m(x)Ix(dy) + px(y)µ(dy))
= f(x, y)m(x)π(dx)Ix(dy) + f(x, y)px(y)µ(dy)
= f(x, x)m(x)π(dx) + f(x, y)px(y)π(dx)µ(dy)
= f(x, x)m(x)π(dx) + f(x, y)px(y)π(x)µ(dx)µ(dy). (24)
For the right hand side, we have
f(y, x)π(dx)P(x, dy) = f(y, x)π(dx) (m(x)Ix(dy) + px(y)µ(dy))
= f(y, x)m(x)π(dx)Ix(dy) + f(y, x)px(y)µ(dy)
= f(x, x)m(x)π(dx) + f(y, x)px(y)π(dx)µ(dy)
= f(x, x)m(x)π(dx) + f(y, x)px(y)π(x)µ(dx)µ(dy)
= f(x, x)m(x)π(dx) + f(x, y)py(x)π(y)µ(dx)µ(dy). (25)
With π(x)px(y) = π(y)py(x), we can see that the left and right hand sides are equal. This
completes the proof.
4
3 Justification of MCMC Methods
Proposition 10. Samples produced using the Metropolis-Hastings algorithm has the desired
distribution, and the resultant chain is reversible.
Proof. It suffices to show that the M-H update is reversible w.r.t. π, which implies that π is
invariant. The stochastic kernel of M-H update is given by
P(x, dy) = m(x)I(x, dy) + q(x, dy)a(x, y) = r(x)I(x, dy) + qx(y)a(x, y)µ(dy) (26)
Here, µ is the base measure, and I(x, dy) is the identity measure given by I(x, A) = 1(x ∈ A),
and m(x) is the probability that the proposal is rejected, which is given by
m(x) = 1 −
Ω
q(x, dy)a(x, y). (27)
Let g(x, y) = h(x)qx(y)a(x, y). With Proposition 9, it suffices to show that g(x, y) = g(y, x).
Here, a(x, y) = min{r(x, y), 1}. Also, from the definition r(x, y) =
h(y)qy(x)
h(x)qx(y) , it is easy to see
that r(x, y) = 1/r(y, x). We first consider the case where r(x, y) ≤ 1 (thus r(y, x) ≥ 1), then
g(x, y) = h(x)qx(y)a(x, y) = h(x)qx(y)
h(y)qy(x)
h(x)qx(y)
= h(y)qy(x), (28)
and
g(y, x) = h(y)qy(x)a(y, x) = h(y)qy(x). (29)
Hence, g(x, y) = g(y, x) when r(x, y) ≤ 1. Similarly, we can show that the equality holds when
r(x, y) ≥ 1. This completes the proof.
Proposition 11. The Metropolis algorithm is a special case of the Metropolis-Hastings algo-
rithm.
Proof. It suffices to show that when q is symmetric, i.e. qx(y) = qy(x), the acceptance rate
reduces to the form given in the Metropolis algorithm. Particularly, when qx(y) = qy(x), the
acceptance rate of the M-H algorithm is
a(x, y) = min{r(x, y), 1} = min
h(y)qy(x)
h(x)qx(y)
, 1 = min
h(y)
h(x)
, 1 . (30)
This completes the proof.
Proposition 12. The Gibbs sampling update is a special case of the Metropolis-Hastings update.
Proof. Without losing generality, we assume the sample is comprised of two components: x =
(x1, x2). Consider a proposal qx(dy) = π(dy1, x2)I(dx2). In this case, we have
r((x1, x2), (y1, x2)) =
π(y1, x2)π(x1, x2)
π(x1, x2)π(y1, x2)
= 1. (31)
This implies that the candidate is always accepted. Also, generating a sample from qx is
equivalent to drawing one from p(y|z). This completes the argument.
Proposition 13. Let K1, . . . , Km be stochastic kernels with invariant measure π, and q ∈ Rm
be a probability vector, then K = m
k=1 qiKi is also a stochastic kernel with invariant measure
π. Moreover, if K1, . . . , Km are all reversible, then K is reversible.
5
Proof. First, it is easy to see that convex combinations of probability measures remain proba-
bility measures. As an immediate consequence, Kx, a convex combination of Ki(x, ·), is also a
probability measure. Given a measurable subset A, Ki(·, A) is measurable for each i, so is their
convex combinations. Hence, we can conclude that K remains a stochastic kernel. Next, we
show that π invariant to K, as
πK = π
m
i=1
qiKi =
m
i=1
qi(πKi) =
m
i=1
qiπ = π. (32)
This proves the first statement. Then, we assume that K1, . . . , Km are reversible, then for K,
we have
f(x, y)π(dx)K(x, dy) =
m
i=1
qi f(x, y)π(dx)Ki(x, dy)
=
m
i=1
qi f(y, x)π(dx)Ki(x, dy)
= f(y, x)π(dx)K(x, dy). (33)
This implies that K is also reversible, thus completing the proof.
Proposition 14. Let K1, . . . , Km be stochastic kernels with invariant measure π. Then K =
Km ◦ · · · ◦ K1 is also a stochastic kernel with invariant measure π.
Proof. Consider K = K2 ◦ K1. To show that K is a stochastic kernel, we first show that
Kx(dy) = K(x, dy) is a probability measure. Given arbitrary measurable subset A, we have
K(x, A) = K1(x, dy)K2(y, A). (34)
As this is a bounded non-negative integration and K2(y, A) is measurable, it constitutes a
measure. Also,
K(x, Ω) = K1(x, dy)K2(y, Ω) = K1(x, dy) = 1. (35)
Hence, K(x, ·) is a probability measure, and thus K a stochastic kernel. Next, we show π is
invariant to K:
πK = π(K2 ◦ K1) = (πK1)K2 = πK2 = π. (36)
We have proved the statement for a composition of two kernels K1 ◦ K2. By induction, we can
further extend to finite composition, thus completing the proof.
6

Weitere ähnliche Inhalte

Was ist angesagt?

Multiple estimators for Monte Carlo approximations
Multiple estimators for Monte Carlo approximationsMultiple estimators for Monte Carlo approximations
Multiple estimators for Monte Carlo approximationsChristian Robert
 
Poster for Bayesian Statistics in the Big Data Era conference
Poster for Bayesian Statistics in the Big Data Era conferencePoster for Bayesian Statistics in the Big Data Era conference
Poster for Bayesian Statistics in the Big Data Era conferenceChristian Robert
 
Inference in generative models using the Wasserstein distance [[INI]
Inference in generative models using the Wasserstein distance [[INI]Inference in generative models using the Wasserstein distance [[INI]
Inference in generative models using the Wasserstein distance [[INI]Christian Robert
 
ABC based on Wasserstein distances
ABC based on Wasserstein distancesABC based on Wasserstein distances
ABC based on Wasserstein distancesChristian Robert
 
Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithms
Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithmsRao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithms
Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithmsChristian Robert
 
Lundi 16h15-copules-charpentier
Lundi 16h15-copules-charpentierLundi 16h15-copules-charpentier
Lundi 16h15-copules-charpentierArthur Charpentier
 
ABC convergence under well- and mis-specified models
ABC convergence under well- and mis-specified modelsABC convergence under well- and mis-specified models
ABC convergence under well- and mis-specified modelsChristian Robert
 
Harmonic Analysis and Deep Learning
Harmonic Analysis and Deep LearningHarmonic Analysis and Deep Learning
Harmonic Analysis and Deep LearningSungbin Lim
 
Mark Girolami's Read Paper 2010
Mark Girolami's Read Paper 2010Mark Girolami's Read Paper 2010
Mark Girolami's Read Paper 2010Christian Robert
 
Some recent developments in the traffic flow variational formulation
Some recent developments in the traffic flow variational formulationSome recent developments in the traffic flow variational formulation
Some recent developments in the traffic flow variational formulationGuillaume Costeseque
 
Solving High-order Non-linear Partial Differential Equations by Modified q-Ho...
Solving High-order Non-linear Partial Differential Equations by Modified q-Ho...Solving High-order Non-linear Partial Differential Equations by Modified q-Ho...
Solving High-order Non-linear Partial Differential Equations by Modified q-Ho...BRNSS Publication Hub
 

Was ist angesagt? (20)

Multiple estimators for Monte Carlo approximations
Multiple estimators for Monte Carlo approximationsMultiple estimators for Monte Carlo approximations
Multiple estimators for Monte Carlo approximations
 
Poster for Bayesian Statistics in the Big Data Era conference
Poster for Bayesian Statistics in the Big Data Era conferencePoster for Bayesian Statistics in the Big Data Era conference
Poster for Bayesian Statistics in the Big Data Era conference
 
Inference in generative models using the Wasserstein distance [[INI]
Inference in generative models using the Wasserstein distance [[INI]Inference in generative models using the Wasserstein distance [[INI]
Inference in generative models using the Wasserstein distance [[INI]
 
ABC based on Wasserstein distances
ABC based on Wasserstein distancesABC based on Wasserstein distances
ABC based on Wasserstein distances
 
Multiattribute utility copula
Multiattribute utility copulaMultiattribute utility copula
Multiattribute utility copula
 
Bohmian
BohmianBohmian
Bohmian
 
Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithms
Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithmsRao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithms
Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithms
 
Slides astin
Slides astinSlides astin
Slides astin
 
Lundi 16h15-copules-charpentier
Lundi 16h15-copules-charpentierLundi 16h15-copules-charpentier
Lundi 16h15-copules-charpentier
 
the ABC of ABC
the ABC of ABCthe ABC of ABC
the ABC of ABC
 
ABC convergence under well- and mis-specified models
ABC convergence under well- and mis-specified modelsABC convergence under well- and mis-specified models
ABC convergence under well- and mis-specified models
 
Slides lausanne-2013-v2
Slides lausanne-2013-v2Slides lausanne-2013-v2
Slides lausanne-2013-v2
 
Harmonic Analysis and Deep Learning
Harmonic Analysis and Deep LearningHarmonic Analysis and Deep Learning
Harmonic Analysis and Deep Learning
 
Mcgill3
Mcgill3Mcgill3
Mcgill3
 
Slides guanauato
Slides guanauatoSlides guanauato
Slides guanauato
 
Mark Girolami's Read Paper 2010
Mark Girolami's Read Paper 2010Mark Girolami's Read Paper 2010
Mark Girolami's Read Paper 2010
 
Some recent developments in the traffic flow variational formulation
Some recent developments in the traffic flow variational formulationSome recent developments in the traffic flow variational formulation
Some recent developments in the traffic flow variational formulation
 
Sildes buenos aires
Sildes buenos airesSildes buenos aires
Sildes buenos aires
 
Solving High-order Non-linear Partial Differential Equations by Modified q-Ho...
Solving High-order Non-linear Partial Differential Equations by Modified q-Ho...Solving High-order Non-linear Partial Differential Equations by Modified q-Ho...
Solving High-order Non-linear Partial Differential Equations by Modified q-Ho...
 
Matrix calculus
Matrix calculusMatrix calculus
Matrix calculus
 

Andere mochten auch

Markovian sequential decision-making in non-stationary environments: applicat...
Markovian sequential decision-making in non-stationary environments: applicat...Markovian sequential decision-making in non-stationary environments: applicat...
Markovian sequential decision-making in non-stationary environments: applicat...Emmanuel Hadoux
 
Optimization of probabilistic argumentation with Markov processes
Optimization of probabilistic argumentation with Markov processesOptimization of probabilistic argumentation with Markov processes
Optimization of probabilistic argumentation with Markov processesEmmanuel Hadoux
 
Thesis : Contraints and Observability In Markov Decision Processes
Thesis : Contraints and Observability In Markov Decision ProcessesThesis : Contraints and Observability In Markov Decision Processes
Thesis : Contraints and Observability In Markov Decision ProcessesCamille Besse
 
MLPI Lecture 0: Overview
MLPI Lecture 0: OverviewMLPI Lecture 0: Overview
MLPI Lecture 0: OverviewDahua Lin
 
Hierarchical Pomdp Planning And Execution
Hierarchical Pomdp Planning And ExecutionHierarchical Pomdp Planning And Execution
Hierarchical Pomdp Planning And Executionahmad bassiouny
 
MLPI Lecture 3: Advanced Sampling Techniques
MLPI Lecture 3: Advanced Sampling TechniquesMLPI Lecture 3: Advanced Sampling Techniques
MLPI Lecture 3: Advanced Sampling TechniquesDahua Lin
 
MLPI Lecture 1: Maths for Machine Learning
MLPI Lecture 1: Maths for Machine LearningMLPI Lecture 1: Maths for Machine Learning
MLPI Lecture 1: Maths for Machine LearningDahua Lin
 
Lecture 5: Variational Estimation and Inference
Lecture 5: Variational Estimation and InferenceLecture 5: Variational Estimation and Inference
Lecture 5: Variational Estimation and InferenceDahua Lin
 
MLPI Lecture 2: Monte Carlo Methods (Basics)
MLPI Lecture 2: Monte Carlo Methods (Basics)MLPI Lecture 2: Monte Carlo Methods (Basics)
MLPI Lecture 2: Monte Carlo Methods (Basics)Dahua Lin
 
Challenges for implementing Monte Carlo Tree Search in commercial games
Challenges for implementing Monte Carlo Tree Search in commercial gamesChallenges for implementing Monte Carlo Tree Search in commercial games
Challenges for implementing Monte Carlo Tree Search in commercial gamesMatthew Bedder
 
Programs that Play better than Us
Programs that Play better than UsPrograms that Play better than Us
Programs that Play better than UsMelvin Zhang
 
MLPI Lecture 4: Graphical Model and Exponential Family
MLPI Lecture 4: Graphical Model and Exponential FamilyMLPI Lecture 4: Graphical Model and Exponential Family
MLPI Lecture 4: Graphical Model and Exponential FamilyDahua Lin
 
Mcts ai
Mcts aiMcts ai
Mcts aiftgaic
 
Dynamic Information Retrieval Tutorial - SIGIR 2015
Dynamic Information Retrieval Tutorial - SIGIR 2015Dynamic Information Retrieval Tutorial - SIGIR 2015
Dynamic Information Retrieval Tutorial - SIGIR 2015Marc Sloan
 
Application of Monte Carlo Tree Search in a Fighting Game AI (GCCE 2016)
Application of Monte Carlo Tree Search in a Fighting Game AI (GCCE 2016)Application of Monte Carlo Tree Search in a Fighting Game AI (GCCE 2016)
Application of Monte Carlo Tree Search in a Fighting Game AI (GCCE 2016)ftgaic
 
Introduction to Reinforcement Learning
Introduction to Reinforcement LearningIntroduction to Reinforcement Learning
Introduction to Reinforcement LearningEdward Balaban
 

Andere mochten auch (20)

sg247934
sg247934sg247934
sg247934
 
Markovian sequential decision-making in non-stationary environments: applicat...
Markovian sequential decision-making in non-stationary environments: applicat...Markovian sequential decision-making in non-stationary environments: applicat...
Markovian sequential decision-making in non-stationary environments: applicat...
 
Optimization of probabilistic argumentation with Markov processes
Optimization of probabilistic argumentation with Markov processesOptimization of probabilistic argumentation with Markov processes
Optimization of probabilistic argumentation with Markov processes
 
Thesis : Contraints and Observability In Markov Decision Processes
Thesis : Contraints and Observability In Markov Decision ProcessesThesis : Contraints and Observability In Markov Decision Processes
Thesis : Contraints and Observability In Markov Decision Processes
 
dissertation
dissertationdissertation
dissertation
 
BayesianBeliefTrajectoryDiagram
BayesianBeliefTrajectoryDiagramBayesianBeliefTrajectoryDiagram
BayesianBeliefTrajectoryDiagram
 
POMDP Seminar Backup3
POMDP Seminar Backup3POMDP Seminar Backup3
POMDP Seminar Backup3
 
MLPI Lecture 0: Overview
MLPI Lecture 0: OverviewMLPI Lecture 0: Overview
MLPI Lecture 0: Overview
 
Hierarchical Pomdp Planning And Execution
Hierarchical Pomdp Planning And ExecutionHierarchical Pomdp Planning And Execution
Hierarchical Pomdp Planning And Execution
 
MLPI Lecture 3: Advanced Sampling Techniques
MLPI Lecture 3: Advanced Sampling TechniquesMLPI Lecture 3: Advanced Sampling Techniques
MLPI Lecture 3: Advanced Sampling Techniques
 
MLPI Lecture 1: Maths for Machine Learning
MLPI Lecture 1: Maths for Machine LearningMLPI Lecture 1: Maths for Machine Learning
MLPI Lecture 1: Maths for Machine Learning
 
Lecture 5: Variational Estimation and Inference
Lecture 5: Variational Estimation and InferenceLecture 5: Variational Estimation and Inference
Lecture 5: Variational Estimation and Inference
 
MLPI Lecture 2: Monte Carlo Methods (Basics)
MLPI Lecture 2: Monte Carlo Methods (Basics)MLPI Lecture 2: Monte Carlo Methods (Basics)
MLPI Lecture 2: Monte Carlo Methods (Basics)
 
Challenges for implementing Monte Carlo Tree Search in commercial games
Challenges for implementing Monte Carlo Tree Search in commercial gamesChallenges for implementing Monte Carlo Tree Search in commercial games
Challenges for implementing Monte Carlo Tree Search in commercial games
 
Programs that Play better than Us
Programs that Play better than UsPrograms that Play better than Us
Programs that Play better than Us
 
MLPI Lecture 4: Graphical Model and Exponential Family
MLPI Lecture 4: Graphical Model and Exponential FamilyMLPI Lecture 4: Graphical Model and Exponential Family
MLPI Lecture 4: Graphical Model and Exponential Family
 
Mcts ai
Mcts aiMcts ai
Mcts ai
 
Dynamic Information Retrieval Tutorial - SIGIR 2015
Dynamic Information Retrieval Tutorial - SIGIR 2015Dynamic Information Retrieval Tutorial - SIGIR 2015
Dynamic Information Retrieval Tutorial - SIGIR 2015
 
Application of Monte Carlo Tree Search in a Fighting Game AI (GCCE 2016)
Application of Monte Carlo Tree Search in a Fighting Game AI (GCCE 2016)Application of Monte Carlo Tree Search in a Fighting Game AI (GCCE 2016)
Application of Monte Carlo Tree Search in a Fighting Game AI (GCCE 2016)
 
Introduction to Reinforcement Learning
Introduction to Reinforcement LearningIntroduction to Reinforcement Learning
Introduction to Reinforcement Learning
 

Ähnlich wie Appendix to MLPI Lecture 2 - Monte Carlo Methods (Basics)

Machine learning (9)
Machine learning (9)Machine learning (9)
Machine learning (9)NYversity
 
Testing for mixtures by seeking components
Testing for mixtures by seeking componentsTesting for mixtures by seeking components
Testing for mixtures by seeking componentsChristian Robert
 
Litvinenko_RWTH_UQ_Seminar_talk.pdf
Litvinenko_RWTH_UQ_Seminar_talk.pdfLitvinenko_RWTH_UQ_Seminar_talk.pdf
Litvinenko_RWTH_UQ_Seminar_talk.pdfAlexander Litvinenko
 
Mathsclass xii (exampler problems)
Mathsclass xii (exampler problems)Mathsclass xii (exampler problems)
Mathsclass xii (exampler problems)nitishguptamaps
 
HypergroupsAssociationSchemes_leejuntaek
HypergroupsAssociationSchemes_leejuntaekHypergroupsAssociationSchemes_leejuntaek
HypergroupsAssociationSchemes_leejuntaekJun Taek Lee
 
Probability Formula sheet
Probability Formula sheetProbability Formula sheet
Probability Formula sheetHaris Hassan
 
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...mathsjournal
 
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...mathsjournal
 
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...mathsjournal
 
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...mathsjournal
 
Common fixed point theorem for occasionally weakly compatible mapping in q fu...
Common fixed point theorem for occasionally weakly compatible mapping in q fu...Common fixed point theorem for occasionally weakly compatible mapping in q fu...
Common fixed point theorem for occasionally weakly compatible mapping in q fu...Alexander Decker
 
Multivriada ppt ms
Multivriada   ppt msMultivriada   ppt ms
Multivriada ppt msFaeco Bot
 
A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”IOSRJM
 

Ähnlich wie Appendix to MLPI Lecture 2 - Monte Carlo Methods (Basics) (20)

Machine learning (9)
Machine learning (9)Machine learning (9)
Machine learning (9)
 
Slides mc gill-v4
Slides mc gill-v4Slides mc gill-v4
Slides mc gill-v4
 
Slides mc gill-v3
Slides mc gill-v3Slides mc gill-v3
Slides mc gill-v3
 
Slides erasmus
Slides erasmusSlides erasmus
Slides erasmus
 
Testing for mixtures by seeking components
Testing for mixtures by seeking componentsTesting for mixtures by seeking components
Testing for mixtures by seeking components
 
Litvinenko_RWTH_UQ_Seminar_talk.pdf
Litvinenko_RWTH_UQ_Seminar_talk.pdfLitvinenko_RWTH_UQ_Seminar_talk.pdf
Litvinenko_RWTH_UQ_Seminar_talk.pdf
 
Mathsclass xii (exampler problems)
Mathsclass xii (exampler problems)Mathsclass xii (exampler problems)
Mathsclass xii (exampler problems)
 
Microeconomics-Help-Experts.pptx
Microeconomics-Help-Experts.pptxMicroeconomics-Help-Experts.pptx
Microeconomics-Help-Experts.pptx
 
stochastic processes assignment help
stochastic processes assignment helpstochastic processes assignment help
stochastic processes assignment help
 
HypergroupsAssociationSchemes_leejuntaek
HypergroupsAssociationSchemes_leejuntaekHypergroupsAssociationSchemes_leejuntaek
HypergroupsAssociationSchemes_leejuntaek
 
Probability Formula sheet
Probability Formula sheetProbability Formula sheet
Probability Formula sheet
 
Lemh1a1
Lemh1a1Lemh1a1
Lemh1a1
 
Lemh1a1
Lemh1a1Lemh1a1
Lemh1a1
 
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
 
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
 
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
 
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
COMMON FIXED POINT THEOREMS IN COMPATIBLE MAPPINGS OF TYPE (P*) OF GENERALIZE...
 
Common fixed point theorem for occasionally weakly compatible mapping in q fu...
Common fixed point theorem for occasionally weakly compatible mapping in q fu...Common fixed point theorem for occasionally weakly compatible mapping in q fu...
Common fixed point theorem for occasionally weakly compatible mapping in q fu...
 
Multivriada ppt ms
Multivriada   ppt msMultivriada   ppt ms
Multivriada ppt ms
 
A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”
 

Kürzlich hochgeladen

IDENTIFICATION OF THE LIVING- forensic medicine
IDENTIFICATION OF THE LIVING- forensic medicineIDENTIFICATION OF THE LIVING- forensic medicine
IDENTIFICATION OF THE LIVING- forensic medicinesherlingomez2
 
GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)Areesha Ahmad
 
Conjugation, transduction and transformation
Conjugation, transduction and transformationConjugation, transduction and transformation
Conjugation, transduction and transformationAreesha Ahmad
 
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPirithiRaju
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bSérgio Sacani
 
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Servicenishacall1
 
Module for Grade 9 for Asynchronous/Distance learning
Module for Grade 9 for Asynchronous/Distance learningModule for Grade 9 for Asynchronous/Distance learning
Module for Grade 9 for Asynchronous/Distance learninglevieagacer
 
Unit5-Cloud.pptx for lpu course cse121 o
Unit5-Cloud.pptx for lpu course cse121 oUnit5-Cloud.pptx for lpu course cse121 o
Unit5-Cloud.pptx for lpu course cse121 oManavSingh202607
 
pumpkin fruit fly, water melon fruit fly, cucumber fruit fly
pumpkin fruit fly, water melon fruit fly, cucumber fruit flypumpkin fruit fly, water melon fruit fly, cucumber fruit fly
pumpkin fruit fly, water melon fruit fly, cucumber fruit flyPRADYUMMAURYA1
 
Introduction,importance and scope of horticulture.pptx
Introduction,importance and scope of horticulture.pptxIntroduction,importance and scope of horticulture.pptx
Introduction,importance and scope of horticulture.pptxBhagirath Gogikar
 
COMPUTING ANTI-DERIVATIVES (Integration by SUBSTITUTION)
COMPUTING ANTI-DERIVATIVES(Integration by SUBSTITUTION)COMPUTING ANTI-DERIVATIVES(Integration by SUBSTITUTION)
COMPUTING ANTI-DERIVATIVES (Integration by SUBSTITUTION)AkefAfaneh2
 
GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)Areesha Ahmad
 
Pests of cotton_Sucking_Pests_Dr.UPR.pdf
Pests of cotton_Sucking_Pests_Dr.UPR.pdfPests of cotton_Sucking_Pests_Dr.UPR.pdf
Pests of cotton_Sucking_Pests_Dr.UPR.pdfPirithiRaju
 
Pests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPirithiRaju
 
Dopamine neurotransmitter determination using graphite sheet- graphene nano-s...
Dopamine neurotransmitter determination using graphite sheet- graphene nano-s...Dopamine neurotransmitter determination using graphite sheet- graphene nano-s...
Dopamine neurotransmitter determination using graphite sheet- graphene nano-s...Mohammad Khajehpour
 
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptxSCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptxRizalinePalanog2
 
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...ssuser79fe74
 
dkNET Webinar "Texera: A Scalable Cloud Computing Platform for Sharing Data a...
dkNET Webinar "Texera: A Scalable Cloud Computing Platform for Sharing Data a...dkNET Webinar "Texera: A Scalable Cloud Computing Platform for Sharing Data a...
dkNET Webinar "Texera: A Scalable Cloud Computing Platform for Sharing Data a...dkNET
 

Kürzlich hochgeladen (20)

IDENTIFICATION OF THE LIVING- forensic medicine
IDENTIFICATION OF THE LIVING- forensic medicineIDENTIFICATION OF THE LIVING- forensic medicine
IDENTIFICATION OF THE LIVING- forensic medicine
 
GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)
 
Clean In Place(CIP).pptx .
Clean In Place(CIP).pptx                 .Clean In Place(CIP).pptx                 .
Clean In Place(CIP).pptx .
 
Conjugation, transduction and transformation
Conjugation, transduction and transformationConjugation, transduction and transformation
Conjugation, transduction and transformation
 
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
 
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
 
Site Acceptance Test .
Site Acceptance Test                    .Site Acceptance Test                    .
Site Acceptance Test .
 
Module for Grade 9 for Asynchronous/Distance learning
Module for Grade 9 for Asynchronous/Distance learningModule for Grade 9 for Asynchronous/Distance learning
Module for Grade 9 for Asynchronous/Distance learning
 
Unit5-Cloud.pptx for lpu course cse121 o
Unit5-Cloud.pptx for lpu course cse121 oUnit5-Cloud.pptx for lpu course cse121 o
Unit5-Cloud.pptx for lpu course cse121 o
 
pumpkin fruit fly, water melon fruit fly, cucumber fruit fly
pumpkin fruit fly, water melon fruit fly, cucumber fruit flypumpkin fruit fly, water melon fruit fly, cucumber fruit fly
pumpkin fruit fly, water melon fruit fly, cucumber fruit fly
 
Introduction,importance and scope of horticulture.pptx
Introduction,importance and scope of horticulture.pptxIntroduction,importance and scope of horticulture.pptx
Introduction,importance and scope of horticulture.pptx
 
COMPUTING ANTI-DERIVATIVES (Integration by SUBSTITUTION)
COMPUTING ANTI-DERIVATIVES(Integration by SUBSTITUTION)COMPUTING ANTI-DERIVATIVES(Integration by SUBSTITUTION)
COMPUTING ANTI-DERIVATIVES (Integration by SUBSTITUTION)
 
GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)
 
Pests of cotton_Sucking_Pests_Dr.UPR.pdf
Pests of cotton_Sucking_Pests_Dr.UPR.pdfPests of cotton_Sucking_Pests_Dr.UPR.pdf
Pests of cotton_Sucking_Pests_Dr.UPR.pdf
 
Pests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdf
 
Dopamine neurotransmitter determination using graphite sheet- graphene nano-s...
Dopamine neurotransmitter determination using graphite sheet- graphene nano-s...Dopamine neurotransmitter determination using graphite sheet- graphene nano-s...
Dopamine neurotransmitter determination using graphite sheet- graphene nano-s...
 
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptxSCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
 
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
 
dkNET Webinar "Texera: A Scalable Cloud Computing Platform for Sharing Data a...
dkNET Webinar "Texera: A Scalable Cloud Computing Platform for Sharing Data a...dkNET Webinar "Texera: A Scalable Cloud Computing Platform for Sharing Data a...
dkNET Webinar "Texera: A Scalable Cloud Computing Platform for Sharing Data a...
 

Appendix to MLPI Lecture 2 - Monte Carlo Methods (Basics)

  • 1. Appendix for Lecture 2: Monte Carlo Methods (Basics) Dahua Lin 1 Justification of Basic Sampling Methods Proposition 1. Let F be the cdf of a real-valued random variable with distribution D. Let U ∼ Uniform([0, 1]), then F−1(U) ∼ D. Proof. Let X = F−1(U). It suffices to show that the cdf of X is F. For any t ∈ R, P(X ≤ t) = P(F−1 (U) ≤ t) = P(U ≤ F(t)) = F(t). (1) Here, we utilize the fact that F is non-decreasing. Proposition 2. Samples producted using Rejection sampling has the desired distribution. Proof. Each iteration actually generate two random variables: x and u, where u ∈ {0, 1} is the indicator of acceptance. The join distribution of x and u is given by ˜p(dx, u = 1) = a(u = 1|x)q(dx) = p(x) Mq(x) q(x)dx = p(x) M µ(dx). (2) Here, a(u|x) is the conditional distribution of u on x, and µ is the base measure. On the other hand, we have Pr(u = 1) = ˜p(dx, u = 1) = p(x) M µ(dx) = 1 M . (3) Thus, the resultant distribution is ˜p(dx|u = 1) = ˜p(dx, u = 1) Pr(u = 1) = p(x)µ(dx). (4) This completes the proof. 2 Markov Chain Theory Proposition 3. When the state space Ω is countable, we have µ − ν TV = 1 2 x∈Ω |µ(x) − ν(x)|. (5) Proof. Let A = {x ∈ Ω : µ(x) ≥ nu(x)}. By definition, we have µ − ν TV ≥ |µ(A) − ν(A)| = µ(A) − ν(A), (6) µ − ν TV ≥ |µ(Ac ) − ν(Ac )| = ν(Ac ) − µ(Ac ). (7) We also have µ(A) − ν(A) = x∈A µ(x) − ν(x) = x∈A |µ(x) − ν(x)|, (8) ν(Ac ) − µ(Ac ) = x∈Ac ν(x) − µ(x) = x∈Ac |µ(x) − ν(x)|. (9) 1
  • 2. Combining the equations above results in µ − ν TV ≥ 1 2 (µ(A) − ν(A) + ν(Ac ) − µ(Ac )) = 1 2 x∈A |µ(x) − ν(x)| + x∈Ac |µ(x) − ν(x)| = 1 2 x∈Ω |µ(x) − ν(x)|. (10) Next we show the inequality of the other direction. For any A ⊂ Ω, we have |µ(Ac ) − ν(Ac )| = |(µ(Ω) − µ(A)) − (ν(Ω) − ν(A))| = |µ(A) − ν(A)| (11) Hence, |µ(A) − ν(A)| = 1 2 (|µ(A) − ν(A)| + |µ(Ac ) − ν(Ac )|) ≤ 1 2 x∈A |µ(x) − ν(x)| + x∈Ac |µ(x) − ν(x)| ≤ 1 2 x∈Ω |µ(x) − ν(x)|. (12) As A is arbitrary, we can conclude that µ − ν TV sup A |µ(A) − ν(A)| ≤ 1 2 x∈Ω |µ(x) − ν(x)|. (13) This completes the proof. Proposition 4. The total variation distance (µ, ν) → µ − ν TV is a metric. Proof. To show that it is a metric, we verify the four properties that a metric needs to satisfy one by one. 1. µ − ν TV is non-negative, as |µ(A) − ν(A)| is always non-negative. 2. When µ = ν, |µ(A) − ν(A)| is always zero, and hence µ − ν TV = 0. On the other hand, when µ = ν, there exists A ⊂ S such that |µ(A) − ν(A)| > 0, and therefore µ − ν TV ≥ |µ(A) − ν(A)| > 0. Together we can conclude that µ − ν TV = 0 iff µ = ν. 3. µ−ν TV = ν −µ TV as |µ(A)−ν(A)| = |ν(A)−µ(A)| holds for any measurable subset A. 4. Next, we show that the total variation distance satisfies the triangle inequality, as below. Let µ, ν, η be three probability measures over Ω: µ − ν TV = sup A∈S |µ(A) − ν(A)| = sup A∈S |µ(A) − η(A) + η(A) − ν(A)| ≤ sup A∈S (|µ(A) − η(A)| + |η(A) − ν(A)|) ≤ sup A∈S |µ(A) − η(A)| + sup A∈S |η(A) − ν(A)| = µ − η TV + η − ν TV . (14) 2
  • 3. The proof is completed. Proposition 5. Consider a Markov chain over a countable space Ω with transition proba- bility matrix P. Let π be a probability measure over Ω that is in detailed balance with P, i.e. π(x)P(x, y) = π(y)P(y, x), ∀x, y ∈ Ω. Then π is invariant to P, i.e. π = πP. Proof. With the assumption of detailed balance, we have (πP)(y) = x∈Ω π(x)P(x, y) = x∈Ω π(y)P(y, x) = π(y) x∈Ω P(y, x) = π(y). (15) Hence, π = πP, or in other words, π is invariant to P. Proposition 6. Let (Xt) be an ergodic Markov chain Markov(π, P) where π is in detailed balance with P, then given arbitrary sequence x0, . . . , xn ∈ Ω, we have Pr(X0 = x0, . . . , Xn = xn) = Pr(X0 = xn, . . . , Xn = x0). (16) Proof. First, we have Pr(X0 = x0, . . . , Xn = xn) = π(x0)P(x0, x1) · · · P(xn−1, xn). (17) On the other hand, by detailed balance, we have P(x, y) = π(y)P(y,x) π(x) , and thus Pr(X0 = xn, . . . , Xn = x0) = π(xn)P(xn, xn−1) · · · P(x1, x0) = π(xn) π(xn−1)P(xn−1, xn) π(xn) · · · π(x0)P(x0, x1) π(x1) = P(xn−1, xn) · · · P(x0, x1)π(x0). (18) Comparing Eq.(17) and Eq.(18) results in the equality that we intend to prove. Proposition 7. Over a measurable space (Ω, S), if a stochastic kernel P is reversible w.r.t. π, then π is invariant to P. Proof. Let π = πP, it suffices to show that π (A) = π(A) for every A ∈ S under the reversibility assumption. Given any A ∈ S, let fA(x, y) := 1(y ∈ A), then we have π (A) = π(dx)P(x, A) = π(dx) fA(x, y)P(x, dy) = fA(x, y)π(dx)P(x, dy) = fA(y, x)π(dx)P(x, dy) = 1(x ∈ A)π(dx)P(x, dy) = 1(x ∈ A)π(dx) P(x, dy) = 1(x ∈ A)π(dx) = π(A). (19) This completes the proof. 3
  • 4. Proposition 8. Given a stochastic kernel P and a probability measure π over (Ω, S). Suppose both Px and π are absolutely continuous w.r.t. a base measure µ, that is, π(dx) = π(x)µ(dx) and P(x, dy) = Px(dy) = px(y)µ(dy), then P is reversible w.r.t. π if and only if π(x)px(y) = π(y)py(x), a.e. (20) Proof. First, assuming detailed balance, i.e. π(x)px(y) = π(y)py(x), a.e., we show reversibility. f(x, y)π(dx)P(x, dy) = f(x, y)π(x)px(y)µ(dx)µ(dy) = f(x, y)π(y)py(x)µ(dx)µ(dy) ...[detailed balance] = f(y, x)π(x)px(y)µ(dx)µ(dy) ...[exchange variables] = f(y, x)π(dx)P(x, dy). (21) Next, we show the converse. The definition of reversibility implies that f(x, y)π(dx)P(x, dy) = f(x, y)π(dy)P(y, dx) (22) Hence, f(x, y)π(x)px(y)µ(dx)µ(dy) = f(x, y)π(y)py(x)µ(dx)µ(dy) (23) Hence, f(x, y)π(x)px(y) = f(x, y)π(y)py(x) a.e. for arbitrary integrable function f, which im- plies that π(x)px(y) = π(y)py(x) a.e. Proposition 9. Given a stochastic kernel P and a probability measure π over (Ω, S). If P(x, dy) = m(x)Ix(dy) + px(y)µ(dy) and π(x)px(y) = π(y)py(x) a.e, then P is reversible w.r.t. π. Proof. Under the given conditions, we have f(x, y)π(dx)P(x, dy) = f(x, y)π(dx) (m(x)Ix(dy) + px(y)µ(dy)) = f(x, y)m(x)π(dx)Ix(dy) + f(x, y)px(y)µ(dy) = f(x, x)m(x)π(dx) + f(x, y)px(y)π(dx)µ(dy) = f(x, x)m(x)π(dx) + f(x, y)px(y)π(x)µ(dx)µ(dy). (24) For the right hand side, we have f(y, x)π(dx)P(x, dy) = f(y, x)π(dx) (m(x)Ix(dy) + px(y)µ(dy)) = f(y, x)m(x)π(dx)Ix(dy) + f(y, x)px(y)µ(dy) = f(x, x)m(x)π(dx) + f(y, x)px(y)π(dx)µ(dy) = f(x, x)m(x)π(dx) + f(y, x)px(y)π(x)µ(dx)µ(dy) = f(x, x)m(x)π(dx) + f(x, y)py(x)π(y)µ(dx)µ(dy). (25) With π(x)px(y) = π(y)py(x), we can see that the left and right hand sides are equal. This completes the proof. 4
  • 5. 3 Justification of MCMC Methods Proposition 10. Samples produced using the Metropolis-Hastings algorithm has the desired distribution, and the resultant chain is reversible. Proof. It suffices to show that the M-H update is reversible w.r.t. π, which implies that π is invariant. The stochastic kernel of M-H update is given by P(x, dy) = m(x)I(x, dy) + q(x, dy)a(x, y) = r(x)I(x, dy) + qx(y)a(x, y)µ(dy) (26) Here, µ is the base measure, and I(x, dy) is the identity measure given by I(x, A) = 1(x ∈ A), and m(x) is the probability that the proposal is rejected, which is given by m(x) = 1 − Ω q(x, dy)a(x, y). (27) Let g(x, y) = h(x)qx(y)a(x, y). With Proposition 9, it suffices to show that g(x, y) = g(y, x). Here, a(x, y) = min{r(x, y), 1}. Also, from the definition r(x, y) = h(y)qy(x) h(x)qx(y) , it is easy to see that r(x, y) = 1/r(y, x). We first consider the case where r(x, y) ≤ 1 (thus r(y, x) ≥ 1), then g(x, y) = h(x)qx(y)a(x, y) = h(x)qx(y) h(y)qy(x) h(x)qx(y) = h(y)qy(x), (28) and g(y, x) = h(y)qy(x)a(y, x) = h(y)qy(x). (29) Hence, g(x, y) = g(y, x) when r(x, y) ≤ 1. Similarly, we can show that the equality holds when r(x, y) ≥ 1. This completes the proof. Proposition 11. The Metropolis algorithm is a special case of the Metropolis-Hastings algo- rithm. Proof. It suffices to show that when q is symmetric, i.e. qx(y) = qy(x), the acceptance rate reduces to the form given in the Metropolis algorithm. Particularly, when qx(y) = qy(x), the acceptance rate of the M-H algorithm is a(x, y) = min{r(x, y), 1} = min h(y)qy(x) h(x)qx(y) , 1 = min h(y) h(x) , 1 . (30) This completes the proof. Proposition 12. The Gibbs sampling update is a special case of the Metropolis-Hastings update. Proof. Without losing generality, we assume the sample is comprised of two components: x = (x1, x2). Consider a proposal qx(dy) = π(dy1, x2)I(dx2). In this case, we have r((x1, x2), (y1, x2)) = π(y1, x2)π(x1, x2) π(x1, x2)π(y1, x2) = 1. (31) This implies that the candidate is always accepted. Also, generating a sample from qx is equivalent to drawing one from p(y|z). This completes the argument. Proposition 13. Let K1, . . . , Km be stochastic kernels with invariant measure π, and q ∈ Rm be a probability vector, then K = m k=1 qiKi is also a stochastic kernel with invariant measure π. Moreover, if K1, . . . , Km are all reversible, then K is reversible. 5
  • 6. Proof. First, it is easy to see that convex combinations of probability measures remain proba- bility measures. As an immediate consequence, Kx, a convex combination of Ki(x, ·), is also a probability measure. Given a measurable subset A, Ki(·, A) is measurable for each i, so is their convex combinations. Hence, we can conclude that K remains a stochastic kernel. Next, we show that π invariant to K, as πK = π m i=1 qiKi = m i=1 qi(πKi) = m i=1 qiπ = π. (32) This proves the first statement. Then, we assume that K1, . . . , Km are reversible, then for K, we have f(x, y)π(dx)K(x, dy) = m i=1 qi f(x, y)π(dx)Ki(x, dy) = m i=1 qi f(y, x)π(dx)Ki(x, dy) = f(y, x)π(dx)K(x, dy). (33) This implies that K is also reversible, thus completing the proof. Proposition 14. Let K1, . . . , Km be stochastic kernels with invariant measure π. Then K = Km ◦ · · · ◦ K1 is also a stochastic kernel with invariant measure π. Proof. Consider K = K2 ◦ K1. To show that K is a stochastic kernel, we first show that Kx(dy) = K(x, dy) is a probability measure. Given arbitrary measurable subset A, we have K(x, A) = K1(x, dy)K2(y, A). (34) As this is a bounded non-negative integration and K2(y, A) is measurable, it constitutes a measure. Also, K(x, Ω) = K1(x, dy)K2(y, Ω) = K1(x, dy) = 1. (35) Hence, K(x, ·) is a probability measure, and thus K a stochastic kernel. Next, we show π is invariant to K: πK = π(K2 ◦ K1) = (πK1)K2 = πK2 = π. (36) We have proved the statement for a composition of two kernels K1 ◦ K2. By induction, we can further extend to finite composition, thus completing the proof. 6