(no subject)
Oct. 20th, 2004 11:04 pmSox!
(You have to admit, baseball is a lot more fun than trying to take the limit of [n!/(x!y!(n-x-y)!)]*[(p1)^x]*[(p2)^y]*[(1-p1-p2)^(n-x-y)] as n->inf.)
(You have to admit, baseball is a lot more fun than trying to take the limit of [n!/(x!y!(n-x-y)!)]*[(p1)^x]*[(p2)^y]*[(1-p1-p2)^(n-x-y)] as n->inf.)
(no subject)
Date: 2004-10-20 09:25 pm (UTC)(no subject)
Date: 2004-10-20 09:29 pm (UTC)I bet they didn't teach you about that.
(no subject)
Date: 2004-10-21 08:41 am (UTC)....If only Dave Olson were here.... ;)
(no subject)
Date: 2004-10-21 02:06 pm (UTC)It was a pain in the ass, but I eventually did get it. I think. I got something which could be easily decomposed into two univariate Poisson distributions, which was the point, so I got the right answer although I couldn't swear by my process. I think one of my two great weaknesses in math is that there are these "Fun Facts" that apparently everybody knows that are never taught anywhere - like that (1 + 1/n)^n -> e as n -> infinity - and, being that I never studied math on my own time, I didn't learn them. (The other great weakness being that I suck at integrals.)
(no subject)
Date: 2004-10-25 06:05 pm (UTC)This is probably too late to be useful, but I'll go ahead and post it since the Sox aren't playing tonight and I need some other way of procrastinating before my exam on Wednesday.
What you want to do is take a limit of the trinomial probability distribution to end up with another probability distribution. For simplicity, I'll discuss the binomial case. There are two important limits of this distribution.
If you take the limit of the binomial distribution P(x) = [n!/(x!*(n-x)!)]*p^x*(1-p)^(n-x) where n->inf and p->0 such that for all n, np=a, where a is constant, you end up with the Poisson distribution p(x) = a^x*e^(-a)/x!. This requires knowing the limit lim t->inf [1-(u/t)]^(t) = e^(-u). The key here is replacing n and p with the constant quantity a.
If you consider a limit where both n and np are much larger than 1 but finite, you can expand x about the mean in a "small" parameter u << np and get a normal approximation with mean np and variance np(1-p). (Doing this expansion requires Stirling's formula: n! ~ sqrt(2*pi*n)*(n/e)^n.) Here, if you take the limit n->inf and switch variables, you end up with a continuous distribution which is normal.
If you consider the large n limit without specifying behavior for p, a normal distribution is the natural answer. Thinking about the binomial distribution as resulting from summing n coin flip results, the central limit theorem states that for large n the distribution should approach a normal distribution with mean np and variance np(1-p).
In the multivariate case, taking the corresponding limits results in distributions which are products of the one-variable distributions.