Types of Regression Models (Sec. 11-1)

Linear Regression:

• - Outcome of Dependent Variable (response) for ith experimental/sampling unit

• - Level of the Independent (predictor) variable for ith experimental/sampling unit

• - Linear (systematic) relation between Yi and Xi (aka conditional mean)

• - Mean of Y when X=0 (Y-intercept)

• - Change in mean of Y when X increases by 1 (slope)

• - Random error term

Note that and are unknown parameters. We estimate them by the least squares method.

Polynomial (Nonlinear) Regression:

This model allows for a curvilinear (as opposed to straight line) relation. Both linear and polynomial regression are susceptible to problems when predictions of Y are made outside the range of the X values used to fit the model. This is referred to as extrapolation.

Least Squares Estimation (Sec. 11-2)

1. Obtain a sample of n pairs (X1,Y1)…(Xn,Yn).

2. Plot the Y values on the vertical (up/down) axis versus their corresponding X values on the horizontal (left/right) axis.

3. Choose the line that minimizes the sum of squared vertical distances from observed values (Yi) to their fitted values ( ) Note:

4. b0 is the Y-intercept for the estimated regression equation

5. b1 is the slope of the estimated regression equation

Measures of Variation (Sec. 11-3)

Sums of Squares

Total sum of squares = Regression sum of squares + Error sum of squares

Total variation = Explained variation + Unexplained variation

Total sum of squares (Total Variation):

Regression sum of squares (Explained Variation):

Error sum of squares (Unexplained Variation):

Coefficients of Determination and Correlation

Coefficient of Determination

Proportion of variation in Y “explained” by the regression on X

Coefficient of Correlation

Measure of the direction and strength of the linear association between Y and X

Standard Error of the Estimate (Residual Standard Deviation)

Estimated standard…