AP Statistics Curriculum 2007 Prob Basics

From Socr

(Difference between revisions)
Jump to: navigation, search
(Approach)
Line 7: Line 7:
A '''simple random sample''' of ''n'' items is a sample in which every member of the population has an equal chance of being selected ''and'' the members of the sample are chosen independently.
A '''simple random sample''' of ''n'' items is a sample in which every member of the population has an equal chance of being selected ''and'' the members of the sample are chosen independently.
-
* '''Example''': Consider a class of students as the population under study.  If we select a sample of size 5, each possible sample of size 5 must have the same chance of being selected. When a sample is chosen randomly it is the process of selection that is random. How could we randomly select five members from this class randomly? Random sampling from finite (or countable) populations is well-defined. On the contrary, random samplying of uncountable pupulations is only allowed under the [http://en.wikipedia.org/wiki/Axiom_of_Choice Axiom of Choice].
+
* '''Example''': Consider a class of students as the population under study.  If we select a sample of size 5, each possible sample of size 5 must have the same chance of being selected. When a sample is chosen randomly it is the process of selection that is random. How could we randomly select five members from this class randomly? Random sampling from finite (or countable) populations is well-defined. On the contrary, random sampling of uncountable populations is only allowed under the [http://en.wikipedia.org/wiki/Axiom_of_Choice Axiom of Choice].
-
* '''Random Number Generation''' using [[SOCR]]:  YOu can use [http://socr.ucla.edu/htmls/SOCR_Modeler.html SOCR Modeler] to [[SOCR_EduMaterials_Activities_RNG | construct random samples]] of any size from a large number of distribution families.  
+
* '''Random Number Generation''' using [[SOCR]]:  You can use [http://socr.ucla.edu/htmls/SOCR_Modeler.html SOCR Modeler] to [[SOCR_EduMaterials_Activities_RNG | construct random samples]] of any size from a large number of distribution families.  
   
   
* '''Questions''':  
* '''Questions''':  
**How would you go about randomly selecting five students from a class of 100?
**How would you go about randomly selecting five students from a class of 100?
-
**How representative of the population is the sample likely to be? The sample wont exactly resemble the population, there will be some chance variation.  This discrepancy is called chance '''error due to sampling'''.  
+
**How representative of the population is the sample likely to be? The sample won’t exactly resemble the population as there will be some chance variation.  This discrepancy is called chance '''error due to sampling'''.  
   
   
* '''Definition''': '''Sampling bias''' is non-randomness that refers to some members having a tendency to be selected more readily than others. When the sample is biased the statistics turn out to be poor estimates.
* '''Definition''': '''Sampling bias''' is non-randomness that refers to some members having a tendency to be selected more readily than others. When the sample is biased the statistics turn out to be poor estimates.
-
===Model Validation===
+
===Hands-on activities===
-
Checking/affirming underlying assumptions.  
+
* Monty Hall (Three-Door) Problem: Go to [http://socr.ucla.edu/htmls/SOCR_Games.html SOCR Games] and select the '''Monty Hall Game'''. Click in the ''Information'' button to get the instructions on using the applet. Run the game 10 times with one of two strategies:
 +
** '''Stay-home''' strategy - choose one card first, as the computer reveals one of the donkey cards, you always stay with the card you originally chose.
 +
** '''Swap''' strategy - choose one card first, as the computer reveals one of the donkey cards, you always swap your original guess and go with the third face-down card!
-
* TBD
+
* You can try the [http://socr.ucla.edu/htmls/SOCR_Experiments.html Monty Hall Experiment], as well. There you can run a very large number of trials automatically and empirically observe the outcomes. Notice that your chance to win doubles if you use the swap-strategy. Why is that?
-
===Computational Resources: Internet-based SOCR Tools===
+
* See the [[SOCR_EduMaterials_Activities_MontyHall | SOCR Monty Hall Activity]].
-
* TBD
+
-
===Examples===
+
===Law of Large Numbers===
-
Computer simulations and real observed data.  
+
When studying the behavior of coin tosses, the law of large numbers implies that the relative proportion (relative frequency) of heads-to-tails in a coin toss experiment becomes more and more stable as the number of tosses increases. This regards the relative frequencies, not absolute counts of #H and #T.
 +
 
 +
* There are two widely held ''misconceptions'' about what the law of large numbers about coin tosses
 +
**Differences between the actual numbers of heads & tails become more and more variable with increase of the number of tosses – a seq. of 10 heads doesn’t increase the chance of a tail on the next trial.
 +
** Coin toss results are independent and fair, and the outcome behavior is unpredictable.
 +
 
 +
* [[SOCR_EduMaterials_Activities_LawOfLargeNumbers | See the Law of large numbers activity]]
 +
 
 +
===Types of probabilities===
 +
Probability models have two essential components: sample space and probabilities.
 +
* '''Sample space''' (S) for a random experiment is the set of all possible outcomes of the experiment.
 +
** An '''event''' is a collection of outcomes.
 +
** An '''event occurs''' if any outcome making up that event occurs.
 +
* '''Probabilities''' for each event in the sample space.
 +
 
 +
Where do the outcomes and the probabilities come from?
 +
* Probabilities may come from models – say mathematical/physical description of the sample space and the chance of each event. Construct a fair die tossing game.
 +
* Probabilities may be derived from data – data observations determine our probability distribution. Say we toss a coin 100 times and the observed Head/Tail counts are used as probabilities.
 +
* Subjective Probabilities – combining data and psychological factors to design a reasonable probability table (e.g., gambling, stock market).
 +
 
 +
===Event Manipulations===
 +
Just like we develop rules for numeric arithmetic, we like to use certain event-manipulation rules (event-arithmetic).
 +
 
 +
* '''Complement''': The complement of an event A, denoted <math>A^c</math> or <math>A'</math>, occurs if and only if A  does not occur.
 +
* <math>A \cup B</math>, read "A or B", contains all outcomes in A or B (or both).
 +
* <math>A \cap B</math>, read "A and B", contains all outcomes which are in both A and B.
 +
 
 +
* Draw Ven diagram pictures of these composite events.
 +
 
 +
* '''Mutually exclusive''' events cannot occur at the same time (<math>A \cap B=0</math>.
 +
 
 +
===Axioms of probability===
 +
 
 +
* First axiom: The probability of an event is a non-negative real number: <math>P(E)\geq 0 \qquad \forall E\subseteq  \Omega</math>, where <math>\Omega</math> is the sample space.
 +
 
 +
* Second axiom: This is the assumption of '''unit measure''': that the probability that some elementary event in the entire sample space will occur is 1.  More specifically, there are no elementary events outside the sample space: <math>P(\Omega) = 1.\,</math>. This is often overlooked in some mistaken probability calculations; if you cannot precisely define the whole sample space, then the probability of any subset cannot be defined either.
 +
 
 +
* Third axiom: This is the assumption of additivity: Any [[countable]] sequence of pair-wise disjoint events <math>E_1, E_2, ...</math> satisfies <math>P(E_1 \cup E_2 \cup \cdots) = \sum_i P(E_i).</math>
 +
 
 +
* Note: For a finite sample space, a sequence of number <math>\left \{p_1, p_2, p_3, …, p_n \right \}</math> is a probability distribution for a sample space <math>S = \left \{s_1, s_2, s_3, …, s_n \right \}</math>, if  the probability of the outcome <math>s_k</math>, <math>p(s_k) = p_k</math>, for each <math>1\lek\len</math>, all <math>p_k \ge0</math> and <math>\sum_{k=1}^n{p_k}=1</math>.
-
* TBD
 
-
 
-
===Hands-on activities===
 
-
Step-by-step practice problems.
 
-
* TBD
 
<hr>
<hr>
===References===
===References===
-
* TBD
+
* [http://www.stat.ucla.edu/%7Edinov/courses_students.dir/07/Fall/STAT13.1.dir/STAT13_notes.dir/lecture03.pdf Probability Lecture Notes]
<hr>
<hr>

Revision as of 00:55, 29 January 2008

Contents

General Advance-Placement (AP) Statistics Curriculum - Fundamentals of Probability Theory

Fundamentals of Probability Theory

Probability theory plays role in all studies of natural processes across scientific disciplines. The need for a theoretical probabilistic foundation is obvious since natural variation effects all measurements, observations and findings about different phenomena. Probability theory provides the basic techniques for statistical inference.

Random Sampling

A simple random sample of n items is a sample in which every member of the population has an equal chance of being selected and the members of the sample are chosen independently.

  • Example: Consider a class of students as the population under study. If we select a sample of size 5, each possible sample of size 5 must have the same chance of being selected. When a sample is chosen randomly it is the process of selection that is random. How could we randomly select five members from this class randomly? Random sampling from finite (or countable) populations is well-defined. On the contrary, random sampling of uncountable populations is only allowed under the Axiom of Choice.
  • Questions:
    • How would you go about randomly selecting five students from a class of 100?
    • How representative of the population is the sample likely to be? The sample won’t exactly resemble the population as there will be some chance variation. This discrepancy is called chance error due to sampling.
  • Definition: Sampling bias is non-randomness that refers to some members having a tendency to be selected more readily than others. When the sample is biased the statistics turn out to be poor estimates.

Hands-on activities

  • Monty Hall (Three-Door) Problem: Go to SOCR Games and select the Monty Hall Game. Click in the Information button to get the instructions on using the applet. Run the game 10 times with one of two strategies:
    • Stay-home strategy - choose one card first, as the computer reveals one of the donkey cards, you always stay with the card you originally chose.
    • Swap strategy - choose one card first, as the computer reveals one of the donkey cards, you always swap your original guess and go with the third face-down card!
  • You can try the Monty Hall Experiment, as well. There you can run a very large number of trials automatically and empirically observe the outcomes. Notice that your chance to win doubles if you use the swap-strategy. Why is that?

Law of Large Numbers

When studying the behavior of coin tosses, the law of large numbers implies that the relative proportion (relative frequency) of heads-to-tails in a coin toss experiment becomes more and more stable as the number of tosses increases. This regards the relative frequencies, not absolute counts of #H and #T.

  • There are two widely held misconceptions about what the law of large numbers about coin tosses
    • Differences between the actual numbers of heads & tails become more and more variable with increase of the number of tosses – a seq. of 10 heads doesn’t increase the chance of a tail on the next trial.
    • Coin toss results are independent and fair, and the outcome behavior is unpredictable.

Types of probabilities

Probability models have two essential components: sample space and probabilities.

  • Sample space (S) for a random experiment is the set of all possible outcomes of the experiment.
    • An event is a collection of outcomes.
    • An event occurs if any outcome making up that event occurs.
  • Probabilities for each event in the sample space.

Where do the outcomes and the probabilities come from?

  • Probabilities may come from models – say mathematical/physical description of the sample space and the chance of each event. Construct a fair die tossing game.
  • Probabilities may be derived from data – data observations determine our probability distribution. Say we toss a coin 100 times and the observed Head/Tail counts are used as probabilities.
  • Subjective Probabilities – combining data and psychological factors to design a reasonable probability table (e.g., gambling, stock market).

Event Manipulations

Just like we develop rules for numeric arithmetic, we like to use certain event-manipulation rules (event-arithmetic).

  • Complement: The complement of an event A, denoted Ac or A', occurs if and only if A does not occur.
  • A \cup B, read "A or B", contains all outcomes in A or B (or both).
  • A \cap B, read "A and B", contains all outcomes which are in both A and B.
  • Draw Ven diagram pictures of these composite events.
  • Mutually exclusive events cannot occur at the same time (A \cap B=0.

Axioms of probability

  • First axiom: The probability of an event is a non-negative real number: P(E)\geq 0 \qquad \forall E\subseteq  \Omega, where Ω is the sample space.
  • Second axiom: This is the assumption of unit measure: that the probability that some elementary event in the entire sample space will occur is 1. More specifically, there are no elementary events outside the sample space: P(\Omega) = 1.\,. This is often overlooked in some mistaken probability calculations; if you cannot precisely define the whole sample space, then the probability of any subset cannot be defined either.
  • Third axiom: This is the assumption of additivity: Any countable sequence of pair-wise disjoint events E1,E2,... satisfies P(E_1 \cup E_2 \cup \cdots) = \sum_i P(E_i).
  • Note: For a finite sample space, a sequence of number Failed to parse (lexing error): \left \{p_1, p_2, p_3, …, p_n \right \}
is a probability distribution for a sample space Failed to parse (lexing error): S = \left \{s_1, s_2, s_3, …, s_n \right \}

, if the probability of the outcome sk, p(sk) = pk, for each Failed to parse (unknown function\lek): 1\lek\len , all p_k \ge0 and \sum_{k=1}^n{p_k}=1.



References




Translate this page:

(default)

Deutsch

Español

Français

Italiano

Português

日本語

България

الامارات العربية المتحدة

Suomi

इस भाषा में

Norge

한국어

中文

繁体中文

Русский

Nederlands

Ελληνικά

Hrvatska

Česká republika

Danmark

Polska

România

Sverige

Personal tools