AP Statistics Curriculum 2007 Distrib Binomial

From Socr

(Difference between revisions)
Jump to: navigation, search
Line 15: Line 15:
===Binomial Random Variables===
===Binomial Random Variables===
-
Suppose we conduct an experiment observing an n-trial (fixed) Bernoulli process. If we are interested in the RV '''X={Number of successes in the n trials}''', then X is called a '''Binomial RV''' and its distribution is called '''Binomial Distribution'''.
+
Suppose we conduct an experiment observing an n-trial (fixed) Bernoulli process. If we are interested in the RV '''X={Number of successes in the n trials}''', then X is called a '''Binomial RV''' and its distribution is called '''Binomial Distribution'''.  
====Examples====
====Examples====
-
*Roll a standard die ten times. Let X be the number of times {6} turned up. The distribution of the random variable X is a binomial distribution with n = 10 (number of trials) and p = 1/6 (probability of "success={6}). The distribution of X may be explicitely written as (P(X=x) are rounded of, you can compute these exactly by going to [http://socr.ucla.edu/htmls/SOCR_Distributions.html SOCR Distributions and selecting Binomial):
+
*Roll a standard die ten times. Let X be the number of times {6} turned up. The distribution of the random variable X is a binomial distribution with n = 10 (number of trials) and p = 1/6 (probability of "success={6}). The distribution of X may be explicitly written as (P(X=x) are rounded of, you can compute these exactly by going to [http://socr.ucla.edu/htmls/SOCR_Distributions.html SOCR Distributions and selecting Binomial):
<center>
<center>
{| class="wikitable" style="text-align:center; width:75%" border="1"
{| class="wikitable" style="text-align:center; width:75%" border="1"
Line 30: Line 30:
<center>[[Image:SOCR_EBook_Dinov_RV_Binomial_013008_Fig1.jpg|400px]]</center>
<center>[[Image:SOCR_EBook_Dinov_RV_Binomial_013008_Fig1.jpg|400px]]</center>
-
* Suppose 10% of the human population carries the green-eye alial. If we choose 1,000 people randomly and let the RV X be the number of green-eyed people in the sample. Then the distribution of X is binomial distribution with n = 1,000 and p = 0.1 (denoted as <math>X \sim B(1,000, 0.1)</math>.
+
* Suppose 10% of the human population carries the green-eye allele. If we choose 1,000 people randomly and let the RV X be the number of green-eyed people in the sample. Then the distribution of X is binomial distribution with n = 1,000 and p = 0.1 (denoted as <math>X \sim B(1,000, 0.1)</math>.
===Binomial Modeling===
===Binomial Modeling===
-
====Exact Binomila Model====
+
====Exact Binomial Model====
The Binomial distribution (i.e., biased-coin tossing experiment) is an exact physical model for any experiment which can be characterized as a series of trials where:
The Binomial distribution (i.e., biased-coin tossing experiment) is an exact physical model for any experiment which can be characterized as a series of trials where:
* Each trial has only two outcomes: success or failure;
* Each trial has only two outcomes: success or failure;
Line 40: Line 40:
====Approximate Binomial Model====
====Approximate Binomial Model====
-
If we observe an experiment comprized of n identical trials involving a large population (of size N). Suppose the population contains a sub-population of subjects that have a characteristic of interest and the sub-population proportion is p (0<p<1). Then the distribution of X={the number of outcomes in the sample with that characteristic of interest}, is approximately Binomial(n, p). This approximation is adequate if the ratio n/N < 0.05.
+
Suppose we observe an experiment comprised of n identical trials involving a large population (of size N). Assume the population contains a sub-population of subjects that have a characteristic of interest and the sub-population proportion is p (0<p<1). Then the distribution of X={the number of outcomes in the sample with that characteristic of interest}, is approximately Binomial(n, p). This approximation is adequate if the ratio n/N < 0.05.
* Example: Polling the US population to see what proportion is/has-been married. Because we sample without replacement (can't ask the same individual twice), the second assumption of the (exact) Binomial model is (slightly) violated. Yet, the small ration of sample to population size yields that the Binomial model is approximately valid (i.e., the proportion of subjects that is/has-been married does not change significantly as we poll one subject, and therefore remove him/her from the complete pool of subjects we poll).
* Example: Polling the US population to see what proportion is/has-been married. Because we sample without replacement (can't ask the same individual twice), the second assumption of the (exact) Binomial model is (slightly) violated. Yet, the small ration of sample to population size yields that the Binomial model is approximately valid (i.e., the proportion of subjects that is/has-been married does not change significantly as we poll one subject, and therefore remove him/her from the complete pool of subjects we poll).
===Binomial Probabilities===
===Binomial Probabilities===
-
If the random variable ''X'' follows the Binomial distribution with (fixed) parameters ''n'' (sample-size) and ''p'' (probability of success at one trial), we write ''X'' ~ B(''n'', ''p''). The probability of getting exactly ''x'' successes is given by the Binomila probability (or mass) function: <math>P(X=x)={n\choose k}p^k(1-p)^{n-k}</math>, for ''x'' = 0, 1, 2, ..., ''n'', where <math>{n\choose k}=\frac{n!}{k!(n-k)!}</math> is the binomial coefficient.  
+
If the random variable ''X'' follows the Binomial distribution with (fixed) parameters ''n'' (sample-size) and ''p'' (probability of success at one trial), we write ''X'' ~ B(''n'', ''p''). The probability of getting exactly ''x'' successes is given by the Binomial probability (or mass) function: <math>P(X=x)={n\choose k}p^k(1-p)^{n-k}</math>, for ''x'' = 0, 1, 2, ..., ''n'', where <math>{n\choose k}=\frac{n!}{k!(n-k)!}</math> is the binomial coefficient.  
This probability expression has an easy and intuitive interpretation. The probability of the ''x'' successes in the ''n'' trials is (''p''<sup>''x''</sup>). Similarly, the probability of the ''n-x'' failures is (1 &minus; ''p'')<sup>''n-x''</sup>. However, the ''x'' successes can be [[AP_Statistics_Curriculum_2007_Prob_Count | arranged]] anywhere among the ''n'' trials, and there are <math>{n\choose k}=\frac{n!}{k!(n-k)!}</math> different ways of arranging the ''x'' successes in a sequence of ''n'' trials, see the [[AP_Statistics_Curriculum_2007_Prob_Count |Counting section]].
This probability expression has an easy and intuitive interpretation. The probability of the ''x'' successes in the ''n'' trials is (''p''<sup>''x''</sup>). Similarly, the probability of the ''n-x'' failures is (1 &minus; ''p'')<sup>''n-x''</sup>. However, the ''x'' successes can be [[AP_Statistics_Curriculum_2007_Prob_Count | arranged]] anywhere among the ''n'' trials, and there are <math>{n\choose k}=\frac{n!}{k!(n-k)!}</math> different ways of arranging the ''x'' successes in a sequence of ''n'' trials, see the [[AP_Statistics_Curriculum_2007_Prob_Count |Counting section]].
-
 
-
 
-
===Model Validation===
 
-
Checking/affirming underlying assumptions.
 
-
 
-
* TBD
 
-
 
-
===Computational Resources: Internet-based SOCR Tools===
 
-
* TBD
 
===Examples===
===Examples===
-
Computer simulations and real observed data.  
+
====A Game of Chance====
 +
Suppose we are offered to play a game of chance under these conditions: it costs us to play $1.50 and the awarded prices are {$1, $2, $3}. Assume the probabilities of winning each price are {0.6, 0.3, 0.1}, respectively. Should we play the game? What are our chances of winning/loosing? Let's let X=awarded price. Then X={1, 2, 3}.
-
* TBD
+
<center>
-
   
+
{| class="wikitable" style="text-align:center; width:75%" border="1"
-
===Hands-on activities===
+
|-
-
Step-by-step practice problems.  
+
| x 1 || 2 || 3
 +
|-
 +
| P(X=x) || 0.6 || 0.3 || 0.1
 +
|-
 +
| x*P(X=x) || 0.6 || 0.6 || 0.3
 +
|}
 +
</center>
 +
Then the mean of this game (i.e., expected return or expectation) is computed as the weighted (by the outcome probabilities) average of all the outcome prices: <math>E[X] = x_1P(X=x_1) + x_2P(X=x_2)+x_3P(X=x_3) = 1\times 0.6 + 2\times 0.3 + 3\times 0.1 = 1.5</math>. In other words, the expected return of this came is $1.5, which equals the entry fee, and hence the game is fair - neither the player nor the house has an advantage in this game (on the long run!) Of course, each streak of n games will produce different outcomes and may give small advantage to one side, however, on the long run, no one will make money.
-
* TBD
+
You can play similar games under different conditions for the probability distribution of the prices using the SOCR [[SOCR_EduMaterials_Activities_BinomialCoinExperiment | Binomial Coin]] or [[SOCR_EduMaterials_Activities_DiceExperiment | Die]] experiments.
<hr>
<hr>
===References===
===References===
-
* TBD
 
<hr>
<hr>

Revision as of 20:22, 30 January 2008

Contents

General Advance-Placement (AP) Statistics Curriculum - Bernoulli and Binomial Random Variables and Experiments

Bernoulli process

A Bernoulli trial is an experiment whose dichotomous outcomes are random (e.g., "success" vs. "failure", "head" vs. "tail", +/-, "yes" vs. "no", etc.) Most common notations of the outcomes of a Bernoulli process are success and failure, even though these outcome labels should not be construed literally.

  • Examples of Bernoulli trials
    • A Coin Toss. We can obverse H="heads", conventionally denoted success, or T="tails" denoted as failure. A fair coin has the probability of success 0.5 by definition.
    • Rolling a Die: Where the outcome space is binarized to "success"={6} and "failure" = {1, 2, 3, 4, 5}.
    • Polls: Choosing a voter at random to ascertain whether that voter will vote "yes" in an upcoming referendum.
  • The Bernoulli random variable (RV): Mathematically, a Bernoulli trial is modeled by a random variable X(outcome) = \begin{cases}0,& s = \texttt{success},\\
1,& s = \texttt{failure}.\end{cases} If p=P(success), then the expected value of X, E[X]=p and the standard deviation of X, SD[X] is \sqrt{p(1-p)}.
  • A Bernoulli process consists of repeatedly performing independent but identical Bernoulli trials.

Binomial Random Variables

Suppose we conduct an experiment observing an n-trial (fixed) Bernoulli process. If we are interested in the RV X={Number of successes in the n trials}, then X is called a Binomial RV and its distribution is called Binomial Distribution.

Examples

  • Roll a standard die ten times. Let X be the number of times {6} turned up. The distribution of the random variable X is a binomial distribution with n = 10 (number of trials) and p = 1/6 (probability of "success={6}). The distribution of X may be explicitly written as (P(X=x) are rounded of, you can compute these exactly by going to [http://socr.ucla.edu/htmls/SOCR_Distributions.html SOCR Distributions and selecting Binomial):
x 0 1 2 3 4 5 6 7 8 9 10
P(X=x) 0.162 0.323 0.291 0.155 0.0543 0.013 0.0022 0.00025 0.000019 8.269e-7 1.654e-8
  • Suppose 10% of the human population carries the green-eye allele. If we choose 1,000 people randomly and let the RV X be the number of green-eyed people in the sample. Then the distribution of X is binomial distribution with n = 1,000 and p = 0.1 (denoted as X \sim B(1,000, 0.1).

Binomial Modeling

Exact Binomial Model

The Binomial distribution (i.e., biased-coin tossing experiment) is an exact physical model for any experiment which can be characterized as a series of trials where:

  • Each trial has only two outcomes: success or failure;
  • P(success)=p is the same for every trial; and
  • Trials are independent.

Approximate Binomial Model

Suppose we observe an experiment comprised of n identical trials involving a large population (of size N). Assume the population contains a sub-population of subjects that have a characteristic of interest and the sub-population proportion is p (0<p<1). Then the distribution of X={the number of outcomes in the sample with that characteristic of interest}, is approximately Binomial(n, p). This approximation is adequate if the ratio n/N < 0.05.

  • Example: Polling the US population to see what proportion is/has-been married. Because we sample without replacement (can't ask the same individual twice), the second assumption of the (exact) Binomial model is (slightly) violated. Yet, the small ration of sample to population size yields that the Binomial model is approximately valid (i.e., the proportion of subjects that is/has-been married does not change significantly as we poll one subject, and therefore remove him/her from the complete pool of subjects we poll).

Binomial Probabilities

If the random variable X follows the Binomial distribution with (fixed) parameters n (sample-size) and p (probability of success at one trial), we write X ~ B(n, p). The probability of getting exactly x successes is given by the Binomial probability (or mass) function: P(X=x)={n\choose k}p^k(1-p)^{n-k}, for x = 0, 1, 2, ..., n, where {n\choose k}=\frac{n!}{k!(n-k)!} is the binomial coefficient.

This probability expression has an easy and intuitive interpretation. The probability of the x successes in the n trials is (px). Similarly, the probability of the n-x failures is (1 − p)n-x. However, the x successes can be arranged anywhere among the n trials, and there are {n\choose k}=\frac{n!}{k!(n-k)!} different ways of arranging the x successes in a sequence of n trials, see the Counting section.

Examples

A Game of Chance

Suppose we are offered to play a game of chance under these conditions: it costs us to play $1.50 and the awarded prices are {$1, $2, $3}. Assume the probabilities of winning each price are {0.6, 0.3, 0.1}, respectively. Should we play the game? What are our chances of winning/loosing? Let's let X=awarded price. Then X={1, 2, 3}.

x 1 2 3
P(X=x) 0.6 0.3 0.1
x*P(X=x) 0.6 0.6 0.3

Then the mean of this game (i.e., expected return or expectation) is computed as the weighted (by the outcome probabilities) average of all the outcome prices: E[X] = x_1P(X=x_1) + x_2P(X=x_2)+x_3P(X=x_3) = 1\times 0.6 + 2\times 0.3 + 3\times 0.1 = 1.5. In other words, the expected return of this came is $1.5, which equals the entry fee, and hence the game is fair - neither the player nor the house has an advantage in this game (on the long run!) Of course, each streak of n games will produce different outcomes and may give small advantage to one side, however, on the long run, no one will make money.

You can play similar games under different conditions for the probability distribution of the prices using the SOCR Binomial Coin or Die experiments.


References




Translate this page:

(default)

Deutsch

Español

Français

Italiano

Português

日本語

България

الامارات العربية المتحدة

Suomi

इस भाषा में

Norge

한국어

中文

繁体中文

Русский

Nederlands

Ελληνικά

Hrvatska

Česká republika

Danmark

Polska

România

Sverige

Personal tools