Talk:Bernoulli distribution

WikiProject Statistics (Rated Start-class, Mid-importance)

This article is within the scope of the WikiProject Statistics, a collaborative effort to improve the coverage of statistics on Wikipedia. If you would like to participate, please visit the project page or join the discussion.

Start  This article has been rated as Start-Class on the quality scale.
Mid  This article has been rated as Mid-importance on the importance scale.
WikiProject Mathematics (Rated Start-class, Low-importance)
This article is within the scope of WikiProject Mathematics, a collaborative effort to improve the coverage of Mathematics on Wikipedia. If you would like to participate, please visit the project page, where you can join the discussion and see a list of open tasks.
Mathematics rating:
 Start Class
 Low Importance
Field: Probability and statistics

f(0.5), for example, is not zero

Bernoulli distribution is just flip once?

in the coin toss example, you can only flip the coin once, not multiple times?

tacticity question

does anyone have an idea if Bernoullian probability on the tacticity page is covered in wikipedia and if so where? possibly redirect? thanks in advance V8rik 23:07, 2 August 2005 (UTC)

Nope. still red links. I have no clue what that is, from context.linas (talk) 17:19, 2 August 2012 (UTC)

mode error

i think the mode of the Bernoulli distribution should be zero if p<0.5, should be 1 if p>0.5, and should be equal to "0 and 1" when p=0.5.

Thanks.

Bernoulli kurtosis excess is actually kurtosis?

The equation given for 'kurtosis excess' is correct.

http://mathworld.wolfram.com/BernoulliDistribution.html gives this formula for the excess but describes it as the kurtosis rather than kurtosis excess as elsewhere.

Although the custom in Wikipedia is to use the term kurtosis for the kurtosis_excess, it is important to avoid confusion.

IMO it would be better to give BOTH in ALL distribution tables to reduce the risk of any more similar muddles.

And to say true kurtosis for kurtosis and excess kurtosis in the text.

BUT I also note that http://mathworld.wolfram.com/Kurtosis.html gives a MUCH simpler formula

 1/(1-p) + 1/p - 6


and this would seem preferable because it is simpler and makes obvious the effect of changing p.

A formula for the 'true' kurtosis is

  1/(1-p) + 1/p - 3


(The kurtosis excess formula given currently is just the binomial with n = 1, and can be simplified).

Paul A Bristow 14:45, 13 December 2006 (UTC) Paul A Bristow

The article says that the Bernoulli distribution with $p=1/2$ has a lower kurtosis than any other probability distribution. But in fact its kurtosis of $-2$ is the same as that of any distribution that takes two distinct values, each with probability $1/2$. Fathead99 (talk) 18:23, 5 March 2008 (UTC)

derivations

I hope somebody could help me in finding derivations or how to derive the skewness and kurtosis, even link to other sites will be much appreciated. —Preceding unsigned comment added by Student29 (talkcontribs) 19:31, 16 January 2008 (UTC)

Why no median?

Is there any particular reason why the Bernoulli distribution lacks a defined median? I would think that the median is 1 if p > 0.5, 0 if p < 0.5, 0.5 if p = 0.5. 169.232.78.24 (talk) —Preceding comment was added at 21:24, 7 February 2008 (UTC)

I think the median is 1 if p > 0.5 and 0 if p < 0.5, but it does not exist of p = 0.5. -- NaBUru38 (talk) 16:23, 11 February 2008 (UTC)

CDF

I'm from Spanish Wikipedia so my english is not good. Sorry about that. I think there's a mistake in the CDF. Here it's defined the CDF for k real, but k only can be 0 or 1. I think that CDF should be (1 - p) for k = 0 and 1 for k = 1. --201.212.140.201 (talk) 18:47, 13 April 2009 (UTC)

I agree with the previous comment. Needs fixing. —Preceding unsigned comment added by 130.91.169.182 (talk) 22:44, 15 April 2010 (UTC)

Why no Simple Language?

Attach more links to concepts explained in a way that people who forgot calc 1 a decade ago can understand. Also, explain how integrals and other complex mathematical terms apply to this concept. Please feel free to edit this into a more editor-friendly lang.

Mistake in formula?

The article says the Bernoulli distribution can be expressed as

$f(k;p) = p^k (1-p)^{1-k}\!\quad \text{for }p\in\{0,1\}$.

Surely this should read

$f(k;p) = p^k (1-p)^{1-k}\!\quad \text{for }k\in\{0,1\}$.

129.173.212.177 (talk) 17:18, 20 May 2010 (UTC)

Also $k$ should be defined. — Preceding unsigned comment added by 159.54.131.7 (talk) 14:35, 10 August 2011 (UTC)

"Number of records"

Just added this section, not sure if the wording is correct — Preceding unsigned comment added by Herix (talkcontribs) 14:16, 7 October 2011 (UTC)

Bournoulli vs. Binomial

My understanding of the difference between a Bernoulli and a binomial trial, model, distribution, etc. is that in a binomial model, all the subjects are assumed to have the same outcome probability (p is the same for all k subjects), while in the more general Bernoulli model, p(i) may vary for each subject. Thus, a binomial model (etc.) is a kind of Bernoulli model, but not vice-versa unless k=1. I notice the Wikipedia articles appear to treat the two as identical. --69.115.221.42 (talk) 12:53, 12 June 2012 (UTC)

You'd have to provide a reference for that, because I don't think it makes any sense. Every book I've seen makes it clear that Bernoulli-anythings are shift-invariant, stationary processes. The only way that they can be stationary is if p(i) is exactly the same for all shifts i; that is the definition of "stationary". linas (talk) 17:16, 2 August 2012 (UTC)

Confused probability distributions and random variables

The very first sentence seems to confuse probability distributions and random variables:

In probability theory and statistics, the Bernoulli distribution, named after Swiss scientist Jacob Bernoulli, is a discrete probability distribution, which takes value 1 with success probability $p$ and value 0 with failure probability $q=1-p$.

Obviously the distribution can only be identified with a function that takes the values $p$ and $1-p$ for 1 and 0, respectively. Perhaps it would be best to just insert "... is the probability distribution of a random variable that takes ..." Seattle Jörg (talk) 13:41, 11 February 2014 (UTC)

E(xi)=p then for n tries E(s)=n*p — Preceding unsigned comment added by Ahyorya (talkcontribs) 16:17, 29 August 2014 (UTC)