Talk:Markov chain

Page contents not supported in other languages.
From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by Mike409 (talk | contribs) at 01:21, 14 March 2013 (→‎Archive). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Template:WP1.0

Biological Examples

The Leslie Matrix is a weak example, it's not even a true markov chain. They are widely used in Biology/Bioinformatics, so it's not hard to replace it with much more interesting examples (DNA, nucleotide substitution models, Wright-Fisher genetic Drift Model, Motiff finding, ion channel/protein conformation transitions... and son on) —Preceding unsigned comment added by 79.169.3.230 (talk) 21:46, 8 January 2011 (UTC)[reply]

Error in "Markov chains with a finite state space"

There is an Error in this part. It says "Since P is a stochastic matrix, always exists". Now each row of P sums to one and all elements are non-negative, still take this example: .

There is clearly no limit if you multiply this matrix with itself. This Statement seems very dangerous to me, since people could take it from here without thinking it through. 85.178.217.33 (talk) 20:55, 23 October 2009 (UTC)[reply]

I've deleted the erroneous statement. Michael Hardy (talk) 21:59, 23 October 2009 (UTC)[reply]

Graphical example (I've got one)

I think a graphical example of a finite chain wouldn't hurt. I just drew one for my thesis but it's in PNG (transparent though), not SVG. I thought I'd put it here in case anyone is feeling inspired to add it to the article. If no one volunteers, I might do it in a few days/weeks when my head's cooled off! Uploaded to Imageshack.

The corresponding transition matrix:

0 0.2 0.7 0 0
0 0 0 0.2 0.5
0 0 0 0.1 0.9
0 0 0 0 0
0 0 0 0 0

193.136.205.152 (talk) 01:34, 8 July 2009 (UTC)[reply]

dead link

removed 'Dead link' remark in 'External Links' section, because the link worked on August 7 '08 in the line

90.152.240.206 (talk) 10:56, 7 August 2008 (UTC) xymx[reply]

just clicked it, seems to work just fine! —Preceding unsigned comment added by 78.49.82.139 (talk) 20:30, 7 August 2009 (UTC)[reply]

Power

Might be a stupid question...but how do you calculate the "power" of the matrix in order to calculate the n-step transition matrix?

See matrix multiplication for the answer to the question above. (That article could use some polishing, BTW, but the answer is there.) Michael Hardy 01:49, 30 Sep 2004 (UTC)

What is meant by a _sequence_ of random variables? A r.v. is a fn. from the probability space to some measurable space, so we can think of assigning a 'number' to each elementary event in the probability space - is the sequence here a series of values from a single r.v. generated by a series of elementary events occurring, or is it actually a sequence of different functions (which would be the literal interpretation of "a sequence of random variables").--SgtThroat 10:37, 10 Dec 2004 (UTC)

Having read around I'm pretty sure that the sequence consists of a sequence of functions. Also these need to have the same domain and codomain (?). Then each point in the sample space would correspond to a sequence x_1,x_2...x_N of values of these functions. The set of all points then corresponds to the set of all sequences. Lets assume for simplicity that each point in the underlying probability space has equal probability - then the transition probability P(x_n|x_n-1) would be obtained by taking all the sequences in which X_n-1=x_n-1 and calculating the fraction of them that have X_n=x_n. Similarly P(x_n|x_n-1,x_n-2) would be obtained by calculating the same fraction among all those that have X_n-1=x_n-1 and X_n-2=x_n-2. The Markov property is the statement that these are equal for all x_n (i.e the conditional distributions are the same) and all n. Can anybody confirm/correct this understanding? --SgtThroat 12:04, 10 Dec 2004 (UTC)

In mathematics, a sequence has a first element, a second one, a third one, and so on, so a sequence of random variables has a first random variable, a second random variable, etc. And since a r.v. is a function whose domain is a probability space, a sequence of random variables is a sequence of such functions. Your remarks about "elementary" events assumes a discrete probability space, and that assumption is not warranted. Definitely all of the r.v.'s in this sequence have the same domain; that is true of any stochastic process.

each point in the sample space would correspond to a sequence x_1,x_2...x_N of values of these functions

Correct.

Lets assume for simplicity that each point in the underlying probability space has equal probability

Again, your assuming discrete without warrant. More later .... Michael Hardy 23:02, 10 Dec 2004 (UTC)

To continue:

It seems your difficulty stems mainly from the unwarranted assumption of discreteness. Consider the "uniform" probability distribution on the interval [0, 1]. The probability that a random variable with this distribution falls in, for example, the interval [0.15, 0.23] is just the length of that interval: 0.23 − 0.15 = 0.08, and similarly for any other subinterval of [0, 1]. This is a continuous, rather than discrete, probability distribution. Nota bene: the probability assigned to each one-point set is zero! Thus, knowing the probability assigned to each one-point set does not tell you what the probability distribution is. Similarly, the familiar bell-shaped curve introduced in even the most elementary statistics courses for non-mathematically inclined people is a continuous distribution, so we're not talking about anything the least bit esoteric here.

Now consider an infinite sequence of independent fair-coin tosses. What's the probability of getting one particular infinite sequence of heads and tails? It's zero, no matter which sequence you pick. So this is also not a discrete distribution. Michael Hardy 01:01, 11 Dec 2004 (UTC)

Hi Michael. Yes, I agree that the assumption of a discrete probability space is rather limited. I was trying to get the discrete case clear before moving on to trying to understand the continuous one, although I neglected to say so. I think I now have a much better understanding of some of this material thanks to your comments and to some additional reading (in particular the entry on random vector which I hadnàt read when I wrote the above. I think we ought to clarify that the sequence of random vbles is a sequence of functions from some probability space onto some domain, both of which are the same for all elements of the sequence. It could use a link to the entry on random vector - although a sequence is not quite the same as a vector, the components of a random vector form a sequence with the required properties and there is much useful content on that page to clarify. SgtThroat 12:21, 12 Dec 2004 (UTC)


Aperiodic

Did anyone notice that the definiton of aperiodic in the text is plain wrong, the condition mentioned implies that the chain is aperiodic but it's not neccesary.

This article needs an update, so I have asked for some help from the collaboration of the week project. From my submission, some todos for this page:

  • Clearer introduction for the newcomer
  • Motivation for using Markov chains
  • Clean-up, reviewing and clarification of the math
  • Better distinction between continuous and discrete state spaces
  • A clear example of a Markov model
  • Conditions for which a unique eigenvector exists, and how to use the Power method to find limit distributions
  • Scientific applications should be extended; currently the list is not explicative
  • Relation to Hidden Markov models
  • Incorporation of Markov process

Anthony Liekens 00:07, 18 September 2005 (UTC)[reply]

Agreed. I've read the article and I still have no idea what a Markov chain - I got directed here by a Slashdot comment, saying "On Usenet there are enough kooks that a simple Markov chain based text analysing program could pass." (in regards to A.I.). Having read the article, I have no idea what he means. Please clean up this article for readability! 80.6.114.153 11:02, 3 December 2005 (UTC)[reply]
Many of the complex definitions have simpler versions which suffice for the simpler special case of a finite state Markov chain. Often, when people refer to "Markov chains", they actually mean "finite state Markov chains". Perhaps we need to clarify this early in the article, and link to a new page specifically on finite state Markov chains. Jim 14159 11:17, 1 November 2007 (UTC)[reply]

Bad didactics

In writing wrongly P(X|Y) instead of P(X=x|Y=y) you doesn't make things easier to understand, just easier to write down and no more than that. If that's your purpose just don't write anything; most easy to copy.Nijdam 22:15, 4 February 2006 (UTC)[reply]


Steady state

Quite a nonsense story about "steady state" and "convenient variables".Nijdam 22:37, 23 March 2006 (UTC)[reply]

bad formatting, perhaps?

some of the equations go off the right-hand side on Safari; dunno if this is a problem with other browsers too, but it's kinda annoying...problem is, my math is not hardcore enough for me to fix this myself--help?

I am not having this problem with internet explorer or firefox. Sounds like a mac error.

Error(s) in article

Memoryless vs Markov

From the article:

The Markov property means the system is memoryless, i.e. it does not "remember" the states it was in before, just "knows" its present state, and hence bases its "decision" to which future state it will transit purely on the present, not considering the past.

This definition of "memoryless" seems incorrect to me. In information theory a "memoryless" source is a sequence of independent identically distributed random variables. A Markov chain is not in general memoryless. WikiC 22:19, 25 May 2006 (UTC)[reply]

Do you know the usage at memorylessness? Michael Hardy 17:18, 13 June 2006 (UTC)[reply]

I agree with the statements above that Markov Chains are not memoryless. Statisticians reserve the term for "truly memoryless" distributions like Geometric or Exponential. As you gain info by knowing what happened in previous states (i.e: states are not independent), I believe we should remove the mention of "memoryless" in the article. We should still keep some mention the special property of the Markov chains, called the "Markov property". Do you guys agree?-akshayaj

The concepts are very much related: in each case we're saying the conditional probability distribution of future states given the present and past states, does not depend on the past states. Michael Hardy 17:21, 13 June 2006 (UTC)[reply]

They may be similar concepts ("memoryless in exponential" vs "Markov property in Markov chains"), but using the strict definitions, you cannot use the term "memoryless" for the property of Markov chains. In fact, for the Markov chains, none of the states but the last one are needed to determine the probability of the current state. But I repeat, knowing any previous state will add information, and thus change the current state probability. Therefore, Markov probabilities are not strictly memoryless, as they "remember" all previous states just by knowing the last state. Speaking statistically, the previous states are not independent of the current one for Markov chains, like they are for Exponential states. See the wikipedia entry on "memoryless" for a perhaps better explanation-Akshayaj 19:55, 20 June 2006 (UTC)[reply]

I'm a bit unclear on the notion of independence, in regards to whether they're correct above. I still believe my main point remains, and will try to get the answer to my "independence issue" soon-Akshayaj 19:55, 20 June 2006 (UTC)[reply]

The point is that MCs are memoryless conditional on the most recent state. 128.8.168.148 17:10, 31 August 2006 (UTC)[reply]

From the article: "In other words, the past states carry no information about future states." The mutual information between the past and future (excluding - or not - the current state) is nonzero in general. I think this is misleading. I think the better statement would be: "Any information about the future that may be contained in the past is also contained in the current state." -jmahoney@cse.ucdavis.edu

Periodicity

Also this sentence is not quite correct: "A process is periodic if there exists at least one state to which the process will continually return with a fixed time period (greater than one)." Consider the transition matrix:

Isn't a Markov chain with this transition matrix periodic? I don't know how to word that sentence better. See the definition of aperiodicity at http://www.quantlet.com/mdstat/scripts/csa/html/node26.html. WikiC 01:00, 26 May 2006 (UTC)[reply]

That definition is indeed incorrect. I felt the other definitions were much too terse (eg, saying "accessible" without saying what it meant), and a few common terms were absent altogether (eg, transient). I've expanded those definitions quite a bit. Check them out for completeness :) - grubber 02:26, 21 June 2006 (UTC)[reply]

Finite vs Discrete

There was a section on "discrete state spaces," but then it described finite state spaces. I've cleaned up that section a little bit. When I have some time, I'm going to remove the integrals earlier in the article and put in some text that reinforces that a MC has a discrete state space. - grubber 19:16, 28 June 2006 (UTC)[reply]

Slight omissions

Describe S

The "S" used in the applicable form of the Chapman-Kolmogorov equation, as well as in the one-step evolution of the marginal distribution, appears to me to be the set of sates (the state space). It even seems very clear, now that I have understood ; however, S is not actually defined anywhere, as far as I can see. Might it not be an idea to add the denotation near "state space" in bold type in the definition section (for example) ? Exaton 10:53, 3 July 2006 (UTC)[reply]

Thanks. I've put an S in the definition paragraph (although it's not very prominent and could get lost maybe... not sure where else to put it tho). Also, I added a note about what "gcd" is, for completeness. - grubber 18:16, 3 July 2006 (UTC)[reply]

Related questions

Probability in the definition of a state's hitting time

A state's (next) hitting time Ti is defined as . I'm wondering what the difference is between that definition and, say, (ie. with a probabilistic notation, like in the definition of a state's period) ? Exaton 11:25, 3 July 2006 (UTC)[reply]

The difference is that the first is a random variable, and the second says something about connectivity between states and about probability distributions. Michael Hardy 14:28, 3 July 2006 (UTC)[reply]

Absorbing states

A very good and clear article to someone who learned the subject - but forgot... I missed a referance about absorbing states. The folowing link helped me in the subject (and gives a simple example for a markov chain)

http://people.hofstra.edu/faculty/Stefan_Waner/Realworld/Summary8.html

Doesn't teach

I apologize, but this page fails to teach anything to anyone who already isn't well versed in the subject. Perhaps a wonderful reference, but not very educational, even more so to the layman. How about breaking this down into more simple terms and more concrete examples? --Anonymous 68.97.215.103 03:25, 19 July 2006 (UTC)[reply]

This isn't wiki-learn-ia, it is an encyclopedia. If you are looking for an article to teach you about Markov chains, you came to the wrong place. Try Wikibooks or Wikiversity. 130.126.108.104 15:24, 21 September 2007 (UTC)[reply]
Please see WP:MTAA. Also, the corresponding article on MathWorld is much easier to understand, and it is also intended to be an encyclopedic reference, like Wikipedia. [1] 86.140.80.60 (talk) 16:42, 21 April 2009 (UTC)[reply]
So then math articles in wikipedia are written exclusively for god? what's wrong with some pedagogy? There are few enough opportunities to explain... 24.10.97.80 (talk) 12:24, 5 November 2012 (UTC)[reply]

The article at minimum needs more of the jargon/terminology to be turned into hyperlinks to articles that explain what they mean. At present, I can't understand the steady state section that I wanted to understand more about because of such terms. And an article that only tells knowledgeable people what they already know does not serve to inform anyone. 08:18, 30 September 2007 (UTC)

Ergodicity hypothesis

We can see the sentence: "Markov chains are related to Brownian motion and the ergodic hypothesis". I think Markov chains aren't always ergordic. If they were, there would only be one final class. Could someone confirm?

This is a very strange sentence. Brownian motion is a continuous time Markov process. It is not ergodic. And, Markov chains are not always ergodic.

 Spmeyn (talk) 16:21, 9 December 2007 (UTC)Sean Meyn[reply]

You are right, but I don't see a problem. This part of the article explains the history of the study of Markov chains. It says that Markov chanes are related to Brownian motion (which is true). They are certainly related, since Brownian motion in the mathematical sense is a continuous time Markov process. But I think in this case the sentence actually referrs to the physics process of Brownian motion (which is also related). Then, the claim that Markov chains are related to the ergodic hypothesis is also correct. They are related, since the question of ergodicity may be stated in the context of Markov chains. It is a hypothesis, so there is no claim that it holds for every Markov chain. I hope this clarifies the issues. Oded (talk) 15:44, 8 May 2008 (UTC)[reply]

Suggestion about Examples

For the non-mathematical reader it would be good to point out that the question of whether a physical process is a Markov process or not is not a well posed question. It cannot be answered until we say what information we are willing to include in the "state" variable(s). It would be nice to give an example of a physical process that is Markov using one definition of state and not Markov using another definition of state.

Bad Example: Coin Toss

I think coin toss is a bad example for Markov chain. The outcome does not depend on current state, it is always 50/50 (or x/(1-x) with an unfair coin) regardless of state. The coin toss can be a degenerate case of Markov chain at best.

A first-in-first-out (FIFO) queue could be a better example: The state is the length of the queue. In a given time interval, a new person (or thing) arrives at the queue with probability of p; and a person (or thing) is served and leaves the queue with probability of q. Then, if current state is i, the probability of going to states i-1, i and i+1 depend only on current state, and input probabilities p and q. —Preceding unsigned comment added by 71.163.214.29 (talk) 04:43, 2 May 2008 (UTC)[reply]


I could not agree more with this comment. A coin toss is not a Markov process! It's an iid process. I will fix this in a few days. Sunbeam44 (talk) 15:15, 10 May 2008 (UTC)[reply]

Any i.i.d. process is Markov. And a sequence of coin tosses is Markov. But it is not a good illustration of the concept of a Markov process. Oded (talk) 15:36, 10 May 2008 (UTC)[reply]

Properties of Markov chains

The section right after the intro, "Properties of Markov chains", seems to be saying that

One needs to have this kind of a property in order for the claimed formulation of the Chapman-Kolmogorov equation to apply, as well as the discussion of the marginal distribution. However, I was under the mpression that. in general, one may have

and still be able to call it a "Markov process", since the next state depends only on the current state (and the current transition probability). Or is it the case that the transition probability must be time-independent, in order for it to be called "Markov"? If the former is true, then it should be stated right up front, at the begining. If not true, then there are numerous, serious faults and errors in the article ... Sincerely confused, linas 00:42, 29 August 2006 (UTC)[reply]

Hmm, it seems inescapable that the former not the latter is intended. Is there a name for the later? In particular, I am interested in studying a system where the transition matrix is finite-dimensional, but it changes over time. It is also the case that for my system, it is not a "second order Markov proces" (i.e. repeated concatentions of the transition matrix is not periodic). linas 01:02, 29 August 2006 (UTC)[reply]
The article assumes, but does not state, that the Markov chains are "time invariant". That is the common assumption, since time-varying Markov chains are much more difficult to study. I think it would be useful for the article to mention this property once. - grubber 15:20, 29 August 2006 (UTC)[reply]
Hmm. Well, I know how to solve a certain class of time-varying Markov chains, and surely there are other classes that are also solvable, so I was surprised that this article seemed to confuse the definition Markov chains in general with stationary Markov chains. linas 15:12, 4 September 2006 (UTC)[reply]

Never mind. On closer inspection, it appears that the article does not assume that Markov chains are time invariant. The notation in the "properties" section is marvelously misleading though: it can be misread, which is what I did at first: I took the superscript (n) to mean "raised to the power of", whereas in fact the equations all hold just fine if (n) is interpreted as merely "some label". Silly me. linas 16:01, 4 September 2006 (UTC)[reply]

More problems

After going through the article more carefully, it seems that most of the article does not assume that the process is stationary (and I explicitly marked the few places where it does). However, the section on "stationary analysis" is ambiguous. I believe that lots of it can be generalized to the non-stationary case, but would require some care with the notation being used. That section needs work. linas 16:20, 4 September 2006 (UTC)[reply]

Definition of "stationary"

The article says: "A stationary Markov chain (a.k.a. time-stationary chain or time-homogenous chain) is a process where one has (*) for all n". My professor uses "stationary" to mean something quite different, namely: for all n, m, x. When I asked him, he was quite sure that his usage was standard. When he wants to express that the transition probabilities are constant with respect to time, i.e. equation (*), he uses "homogenous" or "time-homogenous", but not "stationary". I can imagine that either the article is wrong, or that there is significant variation in terminology within the field - I propose that we either fix the article, or mention the existence of different conventions, respectively. A5 23:36, 19 November 2006 (UTC)[reply]

I've updated the article, with my professor's help. He pointed out that at least some of the external links use his terminology. I hope this is the right thing to do. A5 14:57, 22 November 2006 (UTC)[reply]
It's a good point, and off the top of my head, I know I've heard "time-homogenous" for that property, and I can't recall if I've heard it called "stationary". I'm going to check my Kulkarni book when I get on campus next. For now, I think your edit is fine. - grubber 16:44, 22 November 2006 (UTC)[reply]
I got my Kulkarni book. A DTMC with a stationary distribution π is stationary iff P(Xn=j) = πj. Time-homogeneous is the proper term for independence of n. Thanks for fixing this! - grubber 19:49, 27 November 2006 (UTC)[reply]

Recurrence

Sorry, I was just puzzled about the following statement in the section on recurrence:


It can be shown that a state is recurrent if and only if



I am puzzled about this. Take for example a simple 2 states Markov chain with and , i.e. states flip between state 1 and state 2 every period. Now both states 1 and 2 are recurrent, since they are reached every second period. But and therefore , which means by the statement above no state should be recurrent. Did I get something wrong? —The preceding unsigned comment was added by 212.201.78.127 (talk) 09:45, 15 January 2007 (UTC).[reply]

I think I can help with this - the superscript in the original notation represents the number of steps or transitions in the path from back to . Thus in your example, again using the original notation and . More generally, , thus the sum over all n would indeed be infinite. Trog23 07:20, 7 February 2007 (UTC)[reply]

text manglers

Has anyone ever made a web page showing how simple markov chains are? Wheenever I look them up on the web, I find sigmas and all sorts of stuff, I know nothing about, yet I understand markov chains and can teach any fairly bright person how to create a markov chain text mangler on paper without even the aid of a computer program, in 2 hours at most. There simple. I can think of many many other frivolouse non-text uses for them and suspect they will eventually be the solution to the Turing test given time and creativity, but they are always presented as something that needs BIG math skills. Any links that explain how to do a text mangler markov chain would be appreciatedThaddeus Slamp 06:55, 9 February 2007 (UTC)[reply]

ergodic definitions

Does anyone else find these two statements in the article to be a bit contradictory? We first have

A finite states irreducible Markov chain is said to be ergodic if its states are periodic. 

and then

A state i is said to be ergodic if it is aperiodic and positive recurrent. 
If all states in a Markov chain are ergodic, then the chain is said to be ergodic.

I assume we want "aperiodic" and not "periodic" in the first statement? --Lavaka 23:22, 30 April 2007 (UTC)[reply]

0.7 pass

I have passed this article for Wikipedia 0.7. It meets the Criteria for approval by being a B-Class article of High Importance or Higher. It is of Top importance resulting in the passing of this article. Funpika 22:15, 5 June 2007 (UTC)[reply]

Discrete-parameter Markov process vs. discrete-state Markov process

(This is my first Wikipedia discussion/talk page entry. So, please be patient. ;) )

I'm somewhat puzzled by the following statement of the Markov chain page: "In mathematics, a Markov chain, named after Andrey Markov, is a discrete-time stochastic process with the Markov property."

To my best knowledge, I cannot agree here. As far as I know, stochastic processes with discrete state space are referred to as "chains". On the other hand, we use, e.g., "discrete-time" and "continuous-time" to differ between stochastic processes that have a discrete or continuous parameter, respectively.

See, e.g.,
- G. Bolch et al.: "Queueing Networks and Markov Chains", 2nd Edition, Wiley, 2006, page 53:

[...], we consider Markov processes with discrete state spaces only, that is, Markov chains, [...]

- K.S. Trivedi: "Probability and Statistics with Reliability, Queueing and Computer Science Applications", 2nd Editions, Wiley, 2001, page 296:

Although this definition applies to Markov processes with continuous state space, we will mostly concerned with discrete-state Markov processes — specifically, Markov chains. We will study both discrete-time and continuous-time Markov chains.

Of course, in Markov chains, state changes (transitions) between the different states happen instantaneously, so at time points. Still, the transitions may happen at arbitrary points in time (CTMCs). Only in DTMCs, these time points are fixed.

Thus, I am neither able to comprehend the statement given in the Markov chain article ("[...] a Markov chain [...] is a discrete-time stochastic process") nor the fact that "DTMC" is redirected to "Markov chain" and "CTMC" is redirected to "Continuous-time Markov process". Both DTMCs and CTMCs are Markov chains.

I'm happy to receive any pointer towards work that states that Markov chains always have a discrete parameter/time. MuesLee 13:00, 19 June 2007 (UTC)[reply]

I believe that your understanding is correct, MuesLee, and offer two more references that support your view:
- J.R. Norris. "Markov Chains", Cambridge University Press, 1977 (ISBN 0-521-63396-6) Introduction: "...the case where the process can assume only a finite or countable number of states, when it is usual to refer to it as a Markov Chain."
- William J Stewart. "Introduction to the Numerical Solution of Markov Chains", Princeton University Press (ISBN 0-691-03699-3) Page 5: "If the state space of a Markov Process is discrete, the Markov process is referred to as a Markov Chain."
Trog23 08:04, 24 June 2007 (UTC)[reply]

Periods

"A state i has period k if any return to state i must occur in some multiple of k time steps and k is the largest number with this property. For example, if it is only possible to return to state i in an even number of steps, then i is periodic with period 2. Formally, the period of a state is defined as" I'm a bit confused here - shouldn't it be "smallest number with this property"? The example with even numbers suggests that this is the case as well. I won't change it, in case I'm just wrong. 66.216.172.3 18:05, 12 July 2007 (UTC)[reply]

Nevermind, I'm just dense.

Removed (Bad?) link

I removed this external link:

  • A Markov text generator generates nonsense in the style of another work, because the probability of spitting out each word depends only on the n words before it

Because it didn't work when I ran it. I moved the source code into this utility and linked to that now: http://www.utilitymill.com/utility/Markov_Chain_Parody_Text_Generator —Preceding unsigned comment added by 69.255.197.177 (talk) 20:44, 2 September 2007 (UTC)[reply]

MCML - Markov Chain Markup Language

Has anyone seen a definition of this language anywhere? A Google search just found some references to it from scientific papers, but I'm hoping there's an official or semi-official site with the definition and tools to work with it. —Preceding unsigned comment added by Engmark (talkcontribs) 13:53, 11 June 2008 (UTC)[reply]

This article needs diagrams

If you're looking for a rigorous definition of something, then all that math notation is great. But when you're new to the concept, it's completely impenetrable. Some diagrams would really help to explain what's going on here. Timrb (talk) 14:47, 17 June 2008 (UTC)[reply]

Like those in Random walk? Perhaps the link to that page should be a bit more prominent and point out that there are some pictures there. Tayste (talk - contrib) 20:34, 17 June 2008 (UTC)[reply]
I don't think we are looking for the diagrams such as in Random walk. What would be appropriate would be the kind of diagrams as in finite state machine (not the one at the top, but further down), but with probability labels on the edges. At least, that would be the kind that would best illustrate the Markov chain concept in the finite setting. Oded (talk) 06:08, 18 June 2008 (UTC)[reply]
Finite state machines are good. There are also some "trellis"-style diagrams which show all the states lined up above each "step" in the FSM (like here), which I think are pretty handy for visualization, though the diagram example shown could still use some cleanup and clarification.Timrb (talk) 13:17, 20 June 2008 (UTC)[reply]

Just added a fomula, need some technical feedback on it

I just put in a formula for in the "Markov chains with a finite state space" section. The equation works whenever it is defined (I am pretty sure), but I lack the technical knowledge to provide all the proper caveats about its use.

I also thought about adding to the article a description of the most general method that can be used to find a stationary distribution, but again I am not familiar with the proper technical terms that should be used and all the special cases that should be taken into account.


"An irreducible chain has a stationary distribution if and only if all of its states are positive-recurrent" If I understood well the definitions of the words, this seems not to be true (if stationary means time-stationary).. ? —Preceding unsigned comment added by 195.113.30.252 (talk) 17:21, 6 October 2008 (UTC)[reply]

Limit of P^k

Section "Markov chains with a finite state space" states: Since P is a stochastic matrix, $\lim_{k\to\infty}\mathbf{P}^k$ always exists.

This is only true for ergodic Markov chains. --Borishollas (talk) 12:32, 24 November 2008 (UTC)[reply]

It is clearly not true in general -- take P=[0 1;1 0] (you always switch states). The limit does not exist. 128.143.1.62 (talk) 13:56, 13 April 2009 (UTC)[reply]

Steady-state analysis and limiting distributions

In the "Steady-state analysis and limiting distributions" section it might be appropriate to add one sentence of discussion concerning the periodic case (similarly to the reducible case). Randomblue (talk) 04:07, 22 March 2009 (UTC)[reply]

Ok, I tried something out, please check. Randomblue (talk) 04:47, 22 March 2009 (UTC)[reply]

Clarification requests

In the explanation of additive markov chains of order m

There needs to be a "function exists such that" or at least something that clarifies what this function is.

On a side note, is P(X=x) or Pr(X=x) the standard terminology for the probability that event X will have a value of x? Ryan Brady (talk) 20:36, 11 June 2009 (UTC)[reply]

"Event" is the wrong word here. "Random variable" can be used. The whole expression "X = x" is an "event". Both notations are fairly standard. Michael Hardy (talk) 21:40, 11 June 2009 (UTC)[reply]
That definition of "additive Markov chain" is badly written in a number of respects. I'll have to look further to figure out what it should say instead. Michael Hardy (talk) 21:52, 11 June 2009 (UTC)[reply]

Under Properties of Markov chains

What is S? Ryan Brady (talk) 20:38, 11 June 2009 (UTC)[reply]

It's the state space of the Markov chain. Michael Hardy (talk) 21:38, 11 June 2009 (UTC)[reply]

Properties,Reducibility: When is a "Communicating class" not closed

In the section "properties", subsection "reducibilities", it says: "A communicating class is closed if the probability of leaving the class is zero". If it is possible to get from state A to a state B that is not in the class, then state B is accessible, and so isn't state B in the communicating class? How can an equivalence class not be closed? ( Martin | talkcontribs 07:41, 4 August 2009 (UTC))[reply]

Inconsistency in intro and formal definition

I the formal definition is stated: .... given the present state, the future and past states are independent but in introduction: ...means that future states depend only on the present state,... so which one is it? do future states depends on the present state or not?

GA Fantastic (talk) 12:39, 2 September 2009 (UTC)[reply]

Both are true, and in fact equivalent. There is no contradiction here. But do not forget that it is not about functional dependence; it is probabilistic. Future states depend on the present state in the sense that their probabilities depend. And when you know the present state, past states cannot help in predicting the future, just because given the present, the future and past are independent. Boris Tsirelson (talk) 18:56, 12 September 2009 (UTC)[reply]

Biological applications

Markov models are also used to understand evolution. Mutation accumulation in a population follow markov chains. This has been used to model drug resistance development in HIV etc. See http://biomet.oxfordjournals.org/cgi/reprint/96/3/645 —Preceding unsigned comment added by 129.215.149.99 (talk) 16:57, 14 October 2009 (UTC)[reply]

Markov Chains and Stochastic

Right now, the introduction says:

"Being a stochastic process means that all state transitions are probabilistic (determined by random chance and thus unpredictable in detail, though likely predictable in its statistical properties). At each step the system may change its state from the current state to another state (or remain in the same state) according to a probability distribution."

Are not all Markov chains stochastic? If not, then that should be mentioned, as otherwise the above looks like a contradiction. Unless I'm misunderstanding the definition of "unpredictable in detail".

Pgn674 (talk) 22:37, 21 November 2009 (UTC)[reply]

Finite state space citation

The method to compute the stationary distribution outlined in the Finite state space section lacks a reference.

This page provides a more detailed description of that approach. —Preceding unsigned comment added by 147.162.3.236 (talk) 14:38, 26 January 2010 (UTC)[reply]


I've found these references that could be the references needed as they are used in other paper to cite the same method. But I have no access to them, can anybody check if they are correct?
Pyke, R. (1961a). Markov renewal processes: definitions and preliminary properties. Annals of Mathematical Statistics 32, 1231-42.
Pyke, R. (1961b), Markov renewal properties with finitely many states. Annals of Mathematical Statistics 32, 1243-59. Conjugado (talk) 17:41, 22 February 2011 (UTC)[reply]

Problems with random walk example

While the random walk is often used as an example of a Markov chain, it is confusing to use as a primary example. A recent wording of the random walk example on this page: "An example of a Markov chain is a random walk on the number line which starts at zero and transitions +1 or −1 with equal probability at each step. The position reached in the next transitions only depends on the present position and not on the way this present position is reached."

Problems with this example (under this wording or any other) include...

...it brings into the picture an extraneous property--namely that of the SUM of the random variable sequence in question (the total of all the moves up to "now"; eg, the sum of elements in the set {1,1,-1,1,1,1,-1})--an extraneous property that appears nowhere in the definition of a Markov chain

...the probabilities in this example's random variable do NOT depend on the previous state, where by the state, I am referring to the value of the latest coin toss--0 or 1 (for those who will say, "But it's technically STILL a Markov chain."--yes it is, but should not a primary example aim to avoid confusion and be 'typical' of the category under definition?--as a PRIMARY example it risks only generating confusion)

...the example risks confusion between 'positions' (partial sums of a RV chain) and states/transitions--what is the state? the current position OR the value that got you there (the value, +1 or -1, of the coin toss)--shouldn't the state here, to agree with its definition in this context, refer to the coin toss result and not the position?

I suppose the appeal to the RW example is it looks like a chain of sorts but for a primary example, it only served to send me wandering around the Web for a better explaination.

One problem the example can create is a confusion as to whether Markov chains refer to "the next STATE only depending on the current state" or "the PROBABILITY DISTRIBUTION (for the next state) only depending on the current state"--it is in fact the latter (from Wikipedia's own definition of the Markov property: "the conditional probability distribution of future states of the process depend only upon the present state; that is, given the present, the future does not depend on the past")--if we confuse state with position (the sum of the sequence of random variable values) in the example, we can lose sight of the fact it is the conditional probability that underlies the definition

(However to play devil's advocate and add to the confusion, in defense of this example I do see random walks (including those of the coin-toss type) touted as examples of Markov chains (?))

68.197.205.56 (talk) 22:28, 4 March 2010 (UTC)[reply]

I have edited the text slightly to deal with the point about probabilities (in regard to the random walk example) at least. For the rest ... random walks are examples of Markov Chains (which the text says) but not all Markov chains are random walks (and the text doesn't say that they are). The article has (immediately following, and later on) other examples of Markov chains, so there is really no way the random walk example should cause confusion. You need only look at the next paragraph for a different example. Melcombe (talk) 17:19, 5 March 2010 (UTC)[reply]
For the random walk example, the states are the positions. Suppose you are at state 5. The probability distribution for the transitions would be 0.5 for 5->4, 0.5 for 5->6, and 0 for everything else. After the transition, the next transition will depend only on whether you're at 4 or 6. The fact that you were once at 5 will be "forgotten." The coin toss is kind of irrelevant except to describe the setting. It could also be a die roll and the same concept would apply with different numbers. I agree the current wording is confusing because +1 and -1 are only properties of the transitions, not the transitions themselves. Maghnus (talk) 13:39, 11 March 2010 (UTC)[reply]

Social Sciences

The section on social sciences seems pretty unnecessary. It's just a bunch of buzzwords and half baked personal theories that are (very) loosely related to Markov chains. It is by no means a serious application. It is just jargon at best and simply wrong at worst. It's also unreferenced. 203.167.251.186 (talk) 01:57, 24 May 2010 (UTC)[reply]

I hope I've now addressed this issue. Yaniv256 (talk) 02:52, 29 July 2012 (UTC)[reply]

Unsuitable formula

I removed the premature formula from the introduction. It is in a cryptic shorthand form, confusing a lot of readers, and it doesn't serve any purpose there. Nijdam (talk) 21:46, 26 August 2010 (UTC)[reply]

Joint probability distribution for a reversible Markov chain

Uniform distribution not required

I reverted edits that indicated that

applies only for uniform distributions. The proof that it works generally is pretty short. The detailed balance equation is that

equals the same expression with i and j reversed. But

so it follows immediately that Since the left-hand side of this last string of equalities equals the same expression with i and j reversed, the same must be true with the right-hand side of this last string of equalities. Quantling (talk) 15:26, 14 September 2010 (UTC)[reply]

Fair call. I had misread it. Jheald (talk) 16:25, 14 September 2010 (UTC)[reply]

Stationary process not required

I reverted a sentence that indicated that the result required a process to be stationary. While reversibility does require that there be a stationary distribution, it does not also require that there be a stationary process. In fact, many MCMC algorithms rotate through a set of transition matrices—so by definition the process is not stationary—but the matrices are chosen to have a shared stationary distribution, so all works out in the end. Quantling (talk) 15:26, 14 September 2010 (UTC)[reply]

Elaboration of the lead image

The lead image needs an elaboration on what is going on in the image considering that it is an introductory image. What is it showing? For someone who just recently was introduced to the subject, this is quite hard, if not impossible, to see. --mgarde (talk) 15:46, 24 January 2011 (UTC)[reply]

Reversed Markov Process is a Markov Process too - Should we mention it?

Markov property holds for a reversed Markov Process too (a reversed Markov process is a Markov process too). As Reversible Markov Chains are described here (a much stronger concept - If I understand correctly), we should consider for this simpler fact to be stated too.

[Either here or in Markov_property with possible reference to it].

I would like to hear your opinion. —Preceding unsigned comment added by Sirovsky (talkcontribs) 11:27, 10 May 2011 (UTC)[reply]

If you think you can improve the article, then go ahead. ylloh (talk) 11:41, 10 May 2011 (UTC)[reply]

a simple proof

Simple proof:
For warming up, imagine a cycle of tube full of water of 100L with the current speed 1meter per sec. One meter of the tube is in your house and the rest is outside. If the whole tube lengths 5 meters, than there is 1/5 of the water indoor which is 20L. And it takes 5 seconds for the water to cycle back. If the tube lengths 10 meters, than there's 1/10 of the water indoor which is 10L. And it takes 10 seconds for the water to cycle back.
Now imagine there are two separate tubes each forms a circle like above. They length 5 and 10 meters and both have one meter indoor. However the combined amount of the water from two tubes is only 100L. Now if we could only compare the amount of water inside the house, we found, say, 30% of the water indoor is in the first 1-meter part of the 5-meter tube and 70% of the water indoor is in the second 1-meter part of the 10-meter tube. Then what's the amount of water inside the house?
Let us denote the answer be x. Notice that the amount of water in the first 5-meter tube is 5 times of the 1-meter part of itself which rests indoor. And the 1-meter part has 30%*x of water. Hence the first tube has 30%*x*5 of water. For the same reason, the second 10-meter tube has 70%*x*10 of water. And the total amount of water is 100L = 30%*x*5+ 70%*x*10. Thus x = 100L / (30%*5+ 70%*10) = 11.7647059 liters of water is inside the house.
We can conclude that the portion of the water indoor is 1/E[the time it takes for the water to cycle back], in this case 1/(30%*5+ 70%*10)=11.76%. q.e.d.

I came up this simple proof, but it's undid for the reason "not a "proof" of anything, may be an example but is misplaced and WP:NOR) ".

I knew it's not a complete proof, but I believe the concept is right and help me understand deeper. So are there any adjustments could be made to improve the quality of this? Maybe should I change "simple proof" into "the idea behind this"? Isn't the reason "why there's no proof in the first place?" too long or too mathematics; so why not an "idea behind the proof"? — Preceding unsigned comment added by Guanhomer (talkcontribs) 13:47, 15 July 2011 (UTC)[reply]

Steganography using Markov chains

A while back I (Shannon Larratt) created a proof of concept showing how Markov Chains could be used very effectively for steganography. Here is the explanation and some sample software: http://www.zentastic.com/blog/2005/03/03/now-im-definitely-getting-arrested/

If anyone feels this is relevant, do feel free to add it. I don't want to add my own stuff as that's tacky and I do my best to avoid it!!!Snowrail (talk) 05:26, 16 July 2012 (UTC)[reply]

Markov perfect equilibrium

I think we need a section linking to the game theory concept of Markov perfect equilibrium. I don't want to put something half-baked on the page so I thought to work on it here inside a box and get approval first. Feel free to edit inside the box and post above it. Yaniv256 (talk) 03:27, 29 July 2012 (UTC)[reply]

I now see that the stronger link is to the older concept of stochastic games, so this is now the main article. Yaniv256 (talk) 04:37, 29 July 2012 (UTC)[reply]

Game theory


Main article: stochastic games

Stochastic games generalize both Markov decision processes and repeated games. A Markov perfect equilibrium is a refinement of the concept of sub-game perfect Nash equilibrium to this environment.

Endless applications list

The applications list is far too long and also inaccurate in the following sense. In many cases the application results not from markov chains per se but from an additional construct placed on top of markov chains, such as a Markov decision process or Markov perfect equilibrium. I don't have a clear idea what should we do about it, but I'll post again if I do. Yaniv256 (talk) 19:06, 29 July 2012 (UTC)[reply]

Archive

Would anyone else like to see this talk page get shortened a little? Yaniv256 (talk) 19:20, 29 July 2012 (UTC)[reply]

YES! Maybe then we can work to polish the article. Mike409 (talk) 01:20, 14 March 2013 (UTC)[reply]