|This article is of interest to the following WikiProjects:|
Merge with Chernoff Bounds
The other page just presents a bit more technical stuff. Could probably be merged pretty easily. Thoughts?
I think Chernoff bound is just a special case of the Chernoff bounds. So, it should be presented appropriately. Definitely merged, in either case. Else, its too confusing. —Preceding unsigned comment added by 18.104.22.168 (talk) 15:46, 18 December 2007 (UTC)
Merge it. Having two articles is really confusing, and I don't see how they differ, or relate to each other. 22.214.171.124 19:43, 14 October 2007 (UTC)
Possibly old stuff
i dont understand how to apply the chernoff theorem please discuss some views on it thank you bbye
There's a mistake! The bound P(sum_i X_i>n/2)=1-epsilon>1-exp(-...) (where X_i=1 if the event E_i happened and X_i=0 otherwise) gives an UPPER bound on n, NOT a lower bound as is given in the second formula in the page! The interpretation is wrong too: it takes no more than 150 coin flips to guess the side of the biased coin with 95% accuracy. This error is also evident from the first graph (it isn't too clear what it represents, but it obviously tells us that the Chernoff bound is an upper bound to n with respect to other strategies)
I suggest the formula P(sum_i X_i>n/2)=1-epsilon>1-exp(-2n(p-1/2)^2) (where X_i=1 if the event E_i happened and X_i=0 otherwise) is added to make the bound more obvious.
Addition 13 June 2008
This seems to have left some problems:
- What is a "Poisson trail": is this defined somewhere? (link?)
- Appears to duplicate material later in article, under "Theorem (relative error)".
- Incompleteness: eg possibly needs an "=1" in it, depending on what a "Poisson trail" is.
The formula in the leader to this article seem very useful, but there is no indication as to where they come from in the main body of the article. Not being familiar with the subject, I can't add it myself, but I know I would find it rather useful to see where those formula came from. me_and (talk) 15:10, 25 November 2008 (UTC)
Where does the inequality with p(1-p) come from?
I would love to use the inequality:
However, I haven't been able to verify it (looked at other inequalities, multiple books, etc..) Is it true? Any reference / proofs?
In the intro Chernoff Bounds are described as only requiring independence (and I believe, bounded expectation). However the proof given in section 4.2 only accepts Poisson variables, rather than general random, positive variables. Does anyone know a more general proof? — Preceding unsigned comment added by Thomasda (talk • contribs) 16:26, 2 June 2014 (UTC)
Overlap with a book
The first two sections of the article are almost identical to section 17.2 of the book Algebraic and Stochastic Coding Theory by Dave and Perm Kythe. I wonder if this is appropriate for Wikipedia given that the article does not even cite this book.
On a more technical level, the Chernoff bound
in the Example section does not agree with equation (17.2.2)
from the book. Which one is correct?
- I just found this post, which states that the current version in article should be correct.
- --Marozols (talk) 16:24, 21 October 2014 (UTC)
Hey, I noticed the links (there's two) to the book "Probability and Computing: Randomized Algorithms and Probabilistic Analysis" pointing to google books pages are dead.
I looked up, but am not a 100% sure what the appropriate steps are to replace them, and how to execute these.
The book seems to be online in pdf form here http://bayanbox.ir/view/340405828745623973/probability-and-computing-mitzenmacher-upfal.pdf