# AP Statistics Curriculum 2007 GLM Regress

## General Advance-Placement (AP) Statistics Curriculum - Regression

As we discussed in the Correlation section, many applications involve the analysis of relationships between two, or more, variables involved in the process of interest. Suppose we have bivariate data (X and Y) of a process and we are interested on determining the linear relation between X and Y (e.g., determining a straight line that best fits the pairs of data (X,Y)). A linear relationship between X and Y will give us the power to make predictions - i.e., given a value of X predict a corresponding Y response. Note that in this design, data consists of paired observations (X,Y) - for example, the Longitude and Latitude of the SOCR Earthquake dataset.

### Lines in 2D

There are 3 types of lines in 2D planes - Vertical Lines, Horizontal Lines and Oblique Lines. In general, the mathematical representation of lines in 2D is given by equations like aX + bY = c, most frequently expressed as Y = aX + b, provides the line is not vertical.

Recall that there is a one-to-one correspondence between any line in 2D and (linear) equations of the form

If the line is vertical (X1 = X2): X = X1;
If the line is horizontal (Y1 = Y2): Y = Y1;
Otherwise (oblique line): ${Y-Y_1 \over Y_2-Y_1}= {X-X_1 \over X_2-X_1}$, (for $X_1\not=X_2$ and $Y_1\not=Y_2$)

where (X1,Y1) and (X2,Y2) are two points on the line of interest (2-distinct points in 2D determine a unique line).

Y=2X+1
Y=-3X-5

### Linear Modeling - Regression

There are two contexts for regression:

• Y is an observed variable and X is specified by the researcher - e.g., Y is hair growth after X months, for individuals at certain dose levels of hair growth cream.

Suppose we have n pairs (X,Y), {$X_1, X_2, X_3, \cdots, X_n$} and {$Y_1, Y_2, Y_3, \cdots, Y_n$}, of observations of the same process. If a scatterplot of the data suggests a general linear trend, it would be reasonable to fit a line to the data. The main question is how to determine the best line?

#### Airfare Example

We can see from the scatterplot that greater distance is associated with higher airfare. In other words airports that tend to be further from Baltimore tend to be more expensive airfare. To decide on the best fitting line, we use the least-squares method to fit the least squares (regression) line.

#### Estimating the Best Linear Fit

The parameters of the linear regression line, Y = a + bX, can be estimated using Least Squares. The least squares technique finds the line that minimizes the sum of the squares of the regression residuals, $\hat{\varepsilon_i}=\hat{y}_{i}-y_i$, $\sum_{i=1}^N {\hat{\varepsilon_i}^2} = \sum_{i=1}^N (\hat{y}_{i}-y_i)^2$, where yi and $\hat{y}_{i}=a+bx_i$ are the observed and the predicted values of Y for xi, respectfully.

The minimization problem can be solved using calculus, by finding the first order partial derivatives and setting them equal to zero. The solution gives the slope and y-intercept of the regressions line:

• Regression line Slope:
$\hat{b} = \frac {\sum_{i=1}^{N} (x_{i} - \bar{x})(y_{i} - \bar{y}) } {\sum_{i=1}^{N} (x_{i} - \bar{x}) ^2}$
$\hat{b} = \frac {\sum_{i=1}^{N} {(x_{i}y_{i})} - N \bar{x} \bar{y}} {\sum_{i=1}^{N} (x_{i})^2 - N \bar{x}^2} = \rho_{X,Y} \frac {s_y}{s_x}$, where ρX,Y is the correlation coefficient.
• Y-intercept:
$\hat{a} = \bar{y} - \hat{b} \bar{x}$
• Properties of the least square line:
• The line goes through the point $(\bar{X},\bar{Y})$
• The sum of the residuals is equal to zero
• The estimates are unbiased (their expected values are equal to the real slope and intercept values).

### Regression Coefficients Inference

If the error terms are Normally distributed, the estimate of the slope coefficient has a normal distribution with mean equal to b and standard error given by:

$SE(\hat{b}) = s_ \hat{b} = \sqrt { {1\over (N-2)} \frac {\sum_{i=1}^N \hat{\varepsilon_i}^2} {\sum_{i=1}^N (x_i - \bar{x})^2} }$.
$[ \hat{b} - s_ \hat{b} t_{(\alpha/2, N-2)},\hat{b} + s_ \hat{b} t_{(\alpha/2, N-2)}]$
• The corresponding test for the regression slope coefficient b is analogously computed (Ho:b = bo vs. $H_a: b \not= b_0$, frequently b0 = 0)
$t_o = {\hat{b} - b_o \over s_ \hat{b} } \sim T_{df=N-2}.$

### Hands-on Example

Suppose we have 3 sample of points {(1,-1),(2,4),(6,3)}. The mean of X is 3 and the mean of Y is 2. The slope coefficient estimate is given by:

$\hat{b} = \frac {(1 - 3)((-1) - 2) + (2 - 3)(4 - 2) + (6 - 3)(3 - 2)} {(1 - 3)^2 + (2 - 3)^2 + (6 - 3)^2 } = 7/14 = 0.5$

The standard error of the slope coefficient (b) is 0.866. A 95% confidence interval is given by:

CI(b): [0.5 - 0.866 x 12.7062, 0.5 + 0.866 x 12.7062] = [-10.504, 11.504].

### Earthquake Example

Use the SOCR Earthquake Dataset to determine the best-leaner fit between the Longitude and the Latitude of the California Earthquakes since 1900. What is the interpretation of this regression line (San Andreas fault?). You can see the SOCR Geomap of these Earthquakes. The image below shows how to use the Simple Linear regression in SOCR Analyses to calculate the regression line:

$Latitude = -40.149 -0.6457\times Longitude$

### Hands-on Example

Compute by hand the regression line between X and Y and validate your results using the SOCR Regression Analysis.

 X Y in cm $x_i-\bar{x}$ $y_i-\bar{y}$ $(x_i-\bar{x})^2$ $(y_i-\bar{y})^2$ $(x_i-\bar{x})(y_i-\bar{y})$ -1 0 -3 -0.5 9 0.25 1.5 2 -1 0 -1.5 0 2.25 0 3 1 1 0.5 1 0.25 0.5 4 2 2 1.5 4 2.25 3 $\bar{x}$ $\bar{y}$