Markov Chains and Mixing Times

From Wikipedia, the free encyclopedia
First edition

Markov Chains and Mixing Times is a book on Markov chain mixing times. The second edition was written by David A. Levin, and Yuval Peres. Elizabeth Wilmer was a co-author on the first edition and is credited as a contributor to the second edition. The first edition was published in 2009 by the American Mathematical Society,[1][2] with an expanded second edition in 2017.[3][4][5][6]

Background[edit]

A Markov chain is a stochastic process defined by a set of states and, for each state, a probability distribution on the states. Starting from an initial state, it follows a sequence of states where each state in the sequence is chosen randomly from the distribution associated with the previous state. In that sense, it is "memoryless": each random choice depends only on the current state, and not on the past history of states. Under mild restrictions, a Markov chain with a finite set of states will have a stationary distribution that it converges to, meaning that, after a sufficiently large number of steps, the probability of being in each state will close to that of the stationary distribution, regardless of the initial state or of the exact number of steps. The mixing time of a Markov chain is the number of steps needed for this convergence to happen, to a suitable degree of accuracy. A family of Markov chains is said to be rapidly mixing if the mixing time is a polynomial function of some size parameter of the Markov chain, and slowly mixing otherwise. This book is about finite Markov chains, their stationary distributions and mixing times, and methods for determining whether Markov chains are rapidly or slowly mixing.[1][4]

A classical and familiar example of this phenomenon involves shuffling decks of cards: starting from a non-random initial deck of cards, how many shuffles does it take to reach a nearly-random permutation? This can be modeled as a Markov chain whose states are orderings of the card deck and whose state-to-state transition probabilities are given by some mathematical model of random shuffling such as the Gilbert–Shannon–Reeds model. In this situation, rapid mixing of the Markov chain means that one does not have to perform a huge number of shuffles to reach a sufficiently randomized state. Beyond card games, similar considerations arise in the behavior of the physical systems studied in statistical mechanics, and of certain randomized algorithms.[1][4]

Topics[edit]

The book is divided into two parts, the first more introductory and the second more advanced.[2][6] After three chapters of introductory material on Markov chains, chapter four defines the ways of measuring the distance of a Markov chain to its stationary distribution and the time it takes to reach that distance. Chapter five describes coupling, one of the standard techniques of bounding mixing times. In this technique, one sets up two Markov chains, one starting from the given initial state and the other from the stationary distribution, with transitions that have the correct probabilities within each chain but are not independent from chain-to-chain, in such a way that the two chains become likely to move to the same states as each other. In this way, the mixing time can be bounded by the time for the two coupled chains to synchronize. Chapter six discusses a technique called "strong stationary times" with which, for some Markov chains, one can prove that choosing a stopping time randomly from a certain distribution will result in a state drawn from the stationary distribution.[6]

After a chapter on lower bounds on mixing time based on the "bottleneck ratio" and isoperimetric number,[5] the next two chapters of the first part cover two important examples: card shuffling and random walks on graphs. Chapters 10 and 11 consider two more parameters closely related to the mixing time, the hitting time at which the Markov chain first reaches a specified state, and the cover time at which it has first reached all states.[6] They also discuss time-reversible Markov chains and their connection to electrical networks.[5] The final chapter of this part discusses the connection between the spectral gap of a Markov chain and its mixing time.[6]

The second part of the book includes many more examples in which this theory has been applied, including the Glauber dynamics on the Ising model, Markov models of chromosomal rearrangement, the asymmetric simple exclusion process in which particles randomly jump to unoccupied adjacent spaces, and random walks in the lamplighter group.[6] Topics covered in the second part of the book include more on spectral graphs and expander graphs,[5] path coupling (in which a sequence of more than two Markov chains is coupled in pairs),[6] connections between coupling and the earth mover's distance, martingales,[5] critical temperatures,[2] the "cutoff effect" in which the probability distribution of the chain transitions rapidly between unmixed and mixed,[1][2][6] the evolving set process (a derived Markov chain on sets of states of the given chain),[2] Markov chains with infinitely many states, and Markov chains that operate in continuous time rather than by a discrete sequence of steps. A guest chapter by Jim Propp and David B. Wilson describes coupling from the past, a method for obtaining samples drawn exactly from the stationary distribution rather than (as one obtains from Markov chain Monte Carlo methods) approximations to this distribution.[1][2] The final chapter collects open problems in this area.[5]

Audience and reception[edit]

This book can be used either as a reference by researchers in areas using these methods, or as the basis for a graduate-level course,[1] particularly one limited to the more introductory material in the first part of the book[6] where only an undergraduate-level knowledge of probability theory and linear algebra is required.[1][2] However, reviewer Rick Durrett suggests that the book's contents would be too advanced for undergraduate courses, even at research-level universities,[6] and reviewer Takis Konstantopoulos suggests that the book's contents would be better appreciated by a reader who has already had some exposure to the material that it covers.[5]

Reviewer Olle Häggström calls the book "authoritative and highly readable".[1] Reviewer H. M. Mai writes that its explanations are careful and "well motivated", and that the writing is "lucid and clear".[2] Reviewer László Lakatos calls it "a brilliant guide to the modern theory of Markov chains". And reviewer David Aldous predicts that it "will long remain the definitive required reading" in this area.

References[edit]

  1. ^ a b c d e f g h Häggström, Olle (2010), "Review of Markov Chains and Mixing Times (1st ed.)", Mathematical Reviews, MR 2466937
  2. ^ a b c d e f g h Mai, H. M., "Review of Markov Chains and Mixing Times (1st ed.)", zbMATH, Zbl 1160.60001
  3. ^ Lakatos, László, "Review of Markov Chains and Mixing Times (2nd ed.)", zbMATH, Zbl 1390.60001
  4. ^ a b c Aldous, David (March 2019), "Review of Markov Chains and Mixing Times (2nd ed.)", The Mathematical Intelligencer, 41 (1): 90–91, doi:10.1007/s00283-018-9839-x, MR 3918079
  5. ^ a b c d e f g Konstantopoulos, Takis (2019), "Review of Markov Chains and Mixing Times (2nd ed.)", SIAM Review, 61 (3): 631–634, doi:10.1137/19N974865, MR 3989422
  6. ^ a b c d e f g h i j Durrett, Rick (September 2019), "Review of Markov Chains and Mixing Times (2nd ed.)", MAA Reviews, Mathematical Association of America

External links[edit]