WSU STAT 360
Class Session 3 Summary and Notes Autumn 2000
Today's topics
Probability distributions come from the following considerations. First, discrete distributions come from the contemplation of very simple processes, and simply counting various possible outcomes. Examples of such distributions that we looked at in class are the geometric and binomial distributions, which you may read about in more detail below. There are several other distributions that fall into this category that we will examine in later classes.
Second, we can obtain continuous distributions through a process of taking a discrete distribution to the limit of infinitely many trials. In this case the geometric distribution becomes an exponential distribution and the binomial distribution becomes a normal (Gaussian) distribution. In the case of the normal distribution this result can be thought of as follows. If we consider a random variable that is the outcome of many random influences acting in tandem, then the random variable that results is distributed normally.
Third, we can ask ourselves the following question. Suppose we think about a process that has many possible, equally likely outcomes, each with a miniscule probability of occuring. Suppose we also know the expected value of this process. What probability distribution will produce the expected value and also provides the least information regarding the outcome of any event? In the limit of infinitely many possible outcomes this requisite distribution is exponential.
If, on the other hand, we demand that the distribution produce a known expected value and variance, the requisite distribution is the normal or Gaussian distribution.
Thus, we conclude that many important probability distributions merely reflect that we have minimal information regarding the outcome of a process. However, the fact that the Gaussian or normal distribution appears in so many of our considerations shows something about why it is so widely studied and highly regarded. Petr Beckmann was a very brilliant electrical engineer who said that undue attention is paid to the normal distribution. He deemed this the "Gaussian" disease. Nevertheless, the normal distribution is the single most important continuous distribution you will ever encounter.
Other probability distributions are found as functions of random variables. The most important of these is the function of adding together several independent realizations from a single known distribution. This is the basis of sampling, where our individual sample outcomes are added together to make a sample mean value. Distributions such as Student's t, Chi squared, and Snedecor's F, all result from such considerations. We will examine the process of sampling in class on September 29, 2000. Be sure to attend.
The example of the murder of Sir Edmund Mallory by means of poison candy was contrived , I know, but it provided a very clear exposition of the process of Bayesian analysis. If you wish to pursue this further, you might examine another example that I have included in these notes of classes from last year. There are several points which I think are quite important about the analysis of Sir Edmund's demise.
Finally, we had some fun calculating the odds that two people in class might share the same ordinal day of the month as birthday. With 12 in class we decided the odds were almost certain (0.93) probability, and indeed we had only to query two people to find a match (Jacob and Antonio). Details about this sort of calculation are found below in notes from last year.
Discrete probability distributions and notes regarding their origin. This text book seems to just pull distributions from thin air without any explanation of where they came from. The purpose of the following list is to correct that deficiency.
P(X=k)=p*qk-1
P(X=k)=(k-1)C(r-1)*pr*q(k-r)
P(X=k)=rCk*(N-r)C(n-k)/NCn
P(X1=n1,...,Xk=nk)=N!/[n1!*...*nk!]Obviously we obtained this distribution just like we did for the binomial--we simply counted possible outcomes.
Assignment: For Friday, September 17, 1999, work the following problems. Make your work neat, use quadrile paper, and be thorough.
The in class example: Birthdays and ordinal days of the month.
The problem was to calculate the probability that at least two people in the class share the same ordinal day of the month as a birthday.
We have very little to go on, but it appears reasonable to assume that all months have only 30 days. The error in this assumption is probably small, and for the purpose we made of this in class it is also immaterial. We also assume that it is equally likely that a person's birthday falls on any of the ordinal days.
Next, it is very difficult to calculate probability that people share a ordinal day, but it is trivial to calculate the probability that no two or more people share a day. Let P=probability that at least two people share a birthday. Then Q, the probability that no two or more people share a birthday is Q=1-P. Q is easy to calculate as follows.
Q=q0*q1*...*qN
where qk is the probability that the kth person chosen will not share a birthday with any of the previous k-1 people already chosen. Obviously q1=1 because the first person chosen cannot share a birthday--no one else has been chosen already. Once this person is chosen there are only 29 days left to chose from so q2=29/30, and in general
qk = (30-k+1)/30
For the 11 persons in our class we calculated Q as being about 0.16, and P=1-0.16=0.84. Thus, it is a near certainty that at least two people in class share a ordinal month day as a birthday. Indeed, we only had to sample 6 people until we found the two in our class.
Problems 3.9 and 3.19
Note that the variance is actually E[(n-average of n)2] which you can expand to the form
E(n2)-2*E(n)*(average of n)+E[(average of n)2] or, because E(n) equals the average of n... E(n2)-E(n)2 Therefore problem 3.9 should look like this... Probability of a cast flashing problem p=0.1 Lot size = 10 castings 3.9.a Use the binomial distribution to find probabilty of two flawed casts P(X=2)= 10C2*p_squared*q_eighth_power = 0.19371 3.9.b Probability of scrapping at least one part equals 1 minus probability of no flawed casts 1-P(X=0)= 1-q_tenth_power = 0.651322 3.9.c Expected value = Sum of x*P(X=x) for x=0 to 10 Variance = Expected value of (x-average x)squared = Expected value of x^2 - (Expected value of x)^2 x P(X=x) x*P(X=x) x^2*P(X=x) 0.000 0.349 0.000 0.000 1.000 0.387 0.387 0.387 2.000 0.194 0.387 0.775 3.000 0.057 0.172 0.517 4.000 0.011 0.045 0.179 5.000 0.001 0.007 0.037 6.000 0.000 0.001 0.005 7.000 0.000 0.000 0.000 8.000 0.000 0.000 0.000 9.000 0.000 0.000 0.000 10.000 0.000 0.000 0.000 Sums 1.000 1.900 Expected value = 1 Variance= 0.900 You can see this is not much different Std.Dev= 0.949 than I calculated in class, but in other cases it might be very different. Problem 3.19 The mean number of particles per wafer is 6 Thus use the Poisson distribution with lambda=6 x P(X=x) x*P x^2*P 0.000 0.002 0.000 0.000 1.000 0.015 0.015 0.015 2.000 0.045 0.089 0.178 3.000 0.089 0.268 0.803 4.000 0.134 0.535 2.142 5.000 0.161 0.803 4.016 6.000 0.161 0.964 5.782 7.000 0.138 0.964 6.746 8.000 0.103 0.826 6.608 9.000 0.069 0.620 5.576 10.000 0.041 0.413 4.130 11.000 0.023 0.248 2.726 12.000 0.011 0.135 1.622 13.000 0.005 0.068 0.879 Sums 5.947 41.224 E(x)= 5.947 Note:this is slightly low because of truncation Var(x)= 5.856 Std.Dev= 2.420
Link forward to the next set of class notes for Friday, September 22, 2000