# AP Statistics Curriculum 2007 IntroVar

(Difference between revisions)
 Revision as of 23:00, 13 June 2007 (view source)IvoDinov (Talk | contribs)← Older edit Current revision as of 17:42, 31 March 2013 (view source)IvoDinov (Talk | contribs) m (→The Nature of Data & Variation: typo) (22 intermediate revisions not shown) Line 1: Line 1: - ==This is an Outline of a General Advance-Placement (AP) Statistics Curriculum== + [[AP_Statistics_Curriculum_2007 | General Advance-Placement (AP) Statistics Curriculum]] - Introduction to Statistics - ===Outline=== + ==The Nature of Data & Variation== - Each topic discussed in the SOCR AP Curricumum should contain the following subsections: + No matter how controlled are the environment, the protocol or the design, virtually any repeated measurement, observation, experiment, trial, study or survey is bounded to generate data that varies because of intrinsic (internal to the system) or extrinsic (due to the ambient environment) effects. - * '''Motivation/Problem''': A real data set and fundamental challenge. + - * '''Approach''': Models & strategies for solving the problem, data understanding & inference. + - * '''Model Validation''': Checking/affirming underlying assumptions. + - * '''Computational Resources''': Internet-based SOCR Tools (including offline resources, e.g., tables). + - * '''Examples''': computer simulations and real observed data. + - * '''Hands-on activities''': Step-by-step practice problems. + - ===Introduction to Statistics=== + For example, the UCLA's [[AP_Statistics_Curriculum_2007_IntroVar#References | study of Alzheimer’s disease*]] analyzed the data of 31 Mild Cognitive Impairment (MCI) and 34 probable Alzheimer’s disease (AD) patients. The investigators made every attempt to control as many variables as possible. Yet, the demographic information they collected from the outcomes of the subjects contained unavoidable variation. The same study found variation in the MMSE cognitive scores even in the same subject. The table below shows the demographic characteristics for the subjects and patients included in this study, where the following notation is used M: male; F: female; W: white; AA: African American; A: Asian: - ====[[AP_Statistics_Curriculum_2007_IntroVar | The Nature of Data & Variation]]==== + - No mater how controlled the environment, the protocol or the design, virtually any repeated measurement, observation, experiment, trial, study or survey is bound to generate data that varies because of intrinsic (internal to the system) or extrinsic (due to the ambient environment) effects. + - ====Uses and Abuses of Statistics ==== +
- ====Design of Experiments ==== + {| class="wikitable" style="text-align:center; width:75%" border="1" - ====Statistics with Calculators and Computers==== + |- - + | '''Variable''' || '''Alzheimer’s disease''' || '''MCI''' || '''Test statistics''' || '''Test score''' || '''P-value''' - ===Describing, Exploring, and Comparing Data=== + |- - ====Summarizing data with Frequency Tables ==== + | '''Age (years)''' || 76.2 (8.3) range 52–89 || 73.7 (7.4) range 57–84 || Student’s T  || $t_o = 1.284$ || ''p=0.21'' - ====Pictures of Data ==== + |- - ====Measures of Central Tendency ==== + | '''Gender (M:F)''' || 15:19 || 15:16 || Proportion || $z_o = -0.345$  || ''p=0.733'' - ====Measures of Variation ==== + |- - ====Measures of Position ==== + | '''Education (years)''' || 14.0 (2.1) range 12–19 || 16.23 (2.7) range 12–20 || Wilcoxon rank sum || $w_o = 773.0$ || ''p<0.001'' - ====Exploratory Data Analysis ==== + |- - + | '''Race (W:AA:A)''' || 29:1:4 || 26:2:3 || $\chi_{(df=2)}^2$ || $\chi_{(df=2)}^2=1.18$ || 0.55 - ===Probability=== + |- - ====Fundamentals==== + | '''MMSE''' || 20.9 (6.3) range 4–29 || 28.2 (1.6) range 23–30 || Wilcoxon rank-sum || $w_o= 977.5$  || ''p<0.001'' - ====Addition Rule ==== + |} - ====Multiplication Rule ==== +
- ====Probabilities through Simulations ==== + - ====Counting ==== + - + - ===Probability Distributions=== + - ====Random Variables ==== + - ====Bernoulli & Binomial Experiments ==== + - ====Geometric, HyperGeometric & Negative Binomial==== + - ====Mean, Variance, and Standard Deviation for the Binomial Distribution ==== + - ====Poisson Distribution==== + - + - ===Normal Probability Distributions=== + - ====The Standard Normal Distribution ==== + - ====Nonstandard Normal Distribution: Finding Probabilities ==== + - ====Nonstandard Normal Distributions: Finding Scores ==== + - ===Relations Between Distributions=== + ==Approach== - ====The Central Limit Theorem ==== + Models and strategies for solving  problems and understanding data and inferences. - ====Law of Large Numbers==== + - ====Normal Distribution as Approximation to Binomial Distribution ==== + * Once we accept that all natural phenomena are inherently variant and there are no completely deterministic processes, we need to look for models and techniques that allow us to study such acquired data in the presence of variation, uncertainty and chance. - ====Poisson Approximation to Binomial Distribution ==== + * '''Statistics''' is the data science that investigates natural processes and allows us to quantify variation to make population inferences based on limited observations. - ====Binomial Approximation to HyperGeometric==== + - ====Normal Approximation to Poisson==== + ==Model Validation== + Checking/affirming underlying assumptions. + + * Each model or technique for data exploration, analysis and understanding relies on a set of assumptions, which always need to be validated before the model or analysis tool is employed to study real data (observations or measurements that are perceived or detected by the investigator). + + * Such prior model conjectures or presumptions could take the form of mathematical constraints about the properties of the underlying process, restrictions on the study design or demands on the data acquisition protocol. + + * Common assumptions include (statistical) independence of the measurements, specific limitations on the shape of the observed distribution, restrictions on the parameters of the processes being studied, etc. + + ==Computational Resources: Internet-based SOCR Tools== + * The [[SOCR]] resource contains a variety of educational materials, demonstration applets and learning resources that illustrate data generation, experimentation, exploratory and statistical data analysis. + * [[SOCR_EduMaterials_Activities_RNG | (Numeric Pseudo-Random) Data Generation]] + * [[SOCR_EduMaterials_ExperimentsActivities | Interactive SOCR Experimentation]] with computer generated models of natural phenomena + **  [[SOCR_EduMaterials_Activities_DieCoinExperiment | Bivariate Die-Coin Experiment]] + * [[SOCR_EduMaterials_Activities_Histogram_Graphs | Exploratory Data Analysis]] + * [[SOCR_EduMaterials_AnalysisActivities_ANOVA_1 | Statistical Data Analysis]] + + ==Datasets== + There are [[SOCR_Data | a number of large, natural, useful and demonstrative datasets]] that are provided as part of this statistics [[EBook]]. Many of these data collections are intentionally selected to be large and complex. This choice is driven by the need of emphasizing the symbiosis between driving challenges, statistical concepts, mathematical derivations and the use of technology to solve relevant research problems. + + ==Examples== + Computer simulations and real observed data. + + * For example, [[SOCR_EduMaterials_Activities_Histogram_Graphs | exploratory data analysis using data histograms]]. This SOCR activity illustrates the generation and interpretation of the histogram of quantitative data. - ===Estimates and Sample Sizes=== + ==Hands-on activities== - ====Estimating a Population Mean: Large Samples ==== + Step-by-step practice problems. - ====Estimating a Population Mean: Small Samples ==== + - ====Estimating a Population Proportion ==== + * [[SOCR_EduMaterials_Activities_Histogram_Graphs | Histograms and Frequency Graphs Activity]] - ====Estimating a Population Variance==== + * [[SOCR_EduMaterials_Activities_CardsCoinsSampling | Bivariate Cards and Coins Meta-Activity]] - + - ===Hypothesis Testing=== + ==[[EBook_Problems_EDA_IntroVar|Problems]]== - ====Fundamentals of Hypothesis Testing ==== + - ====Testing a Claim about a Mean: Large Samples ==== +
- ====Testing a Claim about a Mean: Small Samples ==== + ==References== - ====Testing a Claim about a Proportion ==== + * Apostolova LG, Dinov ID, Dutton RA, Hayashi KM, Toga AW, Cummings JL, Thompson PM. [http://brain.oxfordjournals.org/cgi/reprint/awl274v1.pdf 3D comparison of hippocampal atrophy in amnestic mild cognitive impairment and Alzheimer's disease.] Brain. 2006 Nov; 129(Pt 11):2867-73. - ====Testing a Claim about a Standard Deviation or Variance==== + - + - ===Inferences from Two Samples=== + - ====Inferences about Two Means: Dependent Samples ==== + - ====Inferences about Two Means: Independent and Large Samples ==== + - ====Comparing Two Variances ==== + - ====Inferences about Two Means: Independent and Small Samples==== + - ====Inferences about Two Proportions ==== + - + - ===Correlation and Regression=== + - ====Correlation ==== + - ====Regression ==== + - ====Variation and Prediction Intervals ==== + - ====Multiple Regression ==== + - + - ===Multinomial Experiments and Contingency Tables=== + - ====Multinomial Experiments: Goodness-of-Fit ==== + - ====Contingency Tables: Independence and Homogeneity==== + - + - ===Statistical Process Control=== + - ====Control Charts for Variation and Mean ==== + - ====Control Charts for Attributes==== +

## Current revision as of 17:42, 31 March 2013

General Advance-Placement (AP) Statistics Curriculum - Introduction to Statistics

## The Nature of Data & Variation

No matter how controlled are the environment, the protocol or the design, virtually any repeated measurement, observation, experiment, trial, study or survey is bounded to generate data that varies because of intrinsic (internal to the system) or extrinsic (due to the ambient environment) effects.

For example, the UCLA's study of Alzheimer’s disease* analyzed the data of 31 Mild Cognitive Impairment (MCI) and 34 probable Alzheimer’s disease (AD) patients. The investigators made every attempt to control as many variables as possible. Yet, the demographic information they collected from the outcomes of the subjects contained unavoidable variation. The same study found variation in the MMSE cognitive scores even in the same subject. The table below shows the demographic characteristics for the subjects and patients included in this study, where the following notation is used M: male; F: female; W: white; AA: African American; A: Asian:

 Variable Alzheimer’s disease MCI Test statistics Test score P-value Age (years) 76.2 (8.3) range 52–89 73.7 (7.4) range 57–84 Student’s T to = 1.284 p=0.21 Gender (M:F) 15:19 15:16 Proportion zo = − 0.345 p=0.733 Education (years) 14.0 (2.1) range 12–19 16.23 (2.7) range 12–20 Wilcoxon rank sum wo = 773.0 p<0.001 Race (W:AA:A) 29:1:4 26:2:3 $\chi_{(df=2)}^2$ $\chi_{(df=2)}^2=1.18$ 0.55 MMSE 20.9 (6.3) range 4–29 28.2 (1.6) range 23–30 Wilcoxon rank-sum wo = 977.5 p<0.001

## Approach

Models and strategies for solving problems and understanding data and inferences.

• Once we accept that all natural phenomena are inherently variant and there are no completely deterministic processes, we need to look for models and techniques that allow us to study such acquired data in the presence of variation, uncertainty and chance.
• Statistics is the data science that investigates natural processes and allows us to quantify variation to make population inferences based on limited observations.

## Model Validation

Checking/affirming underlying assumptions.

• Each model or technique for data exploration, analysis and understanding relies on a set of assumptions, which always need to be validated before the model or analysis tool is employed to study real data (observations or measurements that are perceived or detected by the investigator).
• Such prior model conjectures or presumptions could take the form of mathematical constraints about the properties of the underlying process, restrictions on the study design or demands on the data acquisition protocol.
• Common assumptions include (statistical) independence of the measurements, specific limitations on the shape of the observed distribution, restrictions on the parameters of the processes being studied, etc.

## Datasets

There are a number of large, natural, useful and demonstrative datasets that are provided as part of this statistics EBook. Many of these data collections are intentionally selected to be large and complex. This choice is driven by the need of emphasizing the symbiosis between driving challenges, statistical concepts, mathematical derivations and the use of technology to solve relevant research problems.

## Examples

Computer simulations and real observed data.

## Hands-on activities

Step-by-step practice problems.