140 likes | 264 Vues
This lecture covers the foundational aspects of Markov Chains, defining them as stochastic processes that guide state transitions based on probabilistic rules. It delves into graphical representations, where nodes correspond to state-space elements and edges represent transition probabilities. Key topics include irreducibility, aperiodicity, and the mixing of reversible Markov chains. The lecture also presents examples like random walks and card shuffling, outlining the convergence to a stationary distribution using the Metropolis approach. The fundamental theorem of Markov Chains is emphasized.
E N D
6.896: Probability and Computation Spring 2011 lecture 3 Constantinos (Costis) Daskalakis costis@mit.edu
Markov Chains Def: A Markov Chain on Ω is a stochastic process (X0, X1,…, Xt, …) such that distribution of Xt conditioning on X0 = x. then a. b. =:
Graphical Representation Represent Markov chain by a graph G(P): - nodes are identified with elements of the state-space Ω • there is a directed edge between states x and y if P(x, y) > 0, with edge-weight P(x,y); (no edge if P(x,y)=0;) - self loops are allowed (when P(x,x) >0) Much of the theory of Markov chains only depends on the topology of G (P), rather than its edge-weights. irreducibility aperiodicity
Mixing of Reversible Markov Chains Theorem (Fundamental Theorem of Markov Chains): If a Markov chain P is irreducible and aperiodicthen: a. it has a unique stationary distribution π =π P ; b. ptx → π as t → ∞, for all x∈Ω.
Example 1 Random Walk on undirected Graph G=(V, E) = P is irreducible iffG is connected; P is aperiodiciffG is non-bipartite; P is reversible with respect to π(x) = deg(x)/(2|E|).
Example 2 Card Shufflings Showed convergence to uniform distribution using concept of doubly stochastic matrices.
Example 3 Input: A positive weight function w : Ω → R+. Designing Markov Chains Goal: Define MC (Xt)t such that ptx → π, where The Metropolis Approach - define a graph Γ(Ω,Ε) on Ω. - define a proposal distributionκ(x,) s.t. κ(x,y ) = κ(y,x ) > 0, for all (x, y)E, xy κ(x,y ) = 0, if (x,y) E
Example 3 The Metropolis Approach - define a graph Γ(Ω,Ε) on Ω. - define a proposal distributionκ(x,) s.t. κ(x,y ) = κ(y,x ) > 0, for all (x, y)E, xy κ(x,y ) = 0, if (x,y) E the Metropolis chain: When chain is at state x - Pick neighbor y with probability κ(x,y ) - Accept move to y with probability
Example 3 the Metropolis chain: When process is at state x - Pick neighbor y with probability κ(x,y ) - Accept move to y with probability Transition Matrix: Claim: The Metropolis Markov chain is reversible wrt to π(x) = w(x)/Z. Proof: 1point exercise
Mixing of Reversible Markov Chains Theorem (Fundamental Theorem of Markov Chains): If a Markov chain P is irreducible and aperiodicthen: a. it has a unique stationary distribution π =π P ; b. ptx → π as t → ∞, for all x∈Ω. Proof: see notes.