Chapter 7: Linear Regression Models Flashcards
(12 cards)
Linear Regression Equation
ŷ = a + bx
ŷ = predicted y-value
a = y-intercept
b = slope
Extrapolation
Predictions made outside of the interval of current data’s x-values (often not reliable)
Residual
The difference between the actual response value and the model’s predicted response value (y-ŷ)
Positive = underestimation
Negative = overestimation
Residual Plot
Visualizes and accentuates the residuals, allowing us to assess our model’s fit
Good Fit Vs. Bad Fit For Residual Plots
Good: Apparent randomness, centered at 0, no clear patterns
Bad: Curved pattern
Line Of Best Fit
The line that minimizes the sum of the squares of the residuals
(Contains mean)
Coefficient Of Determination (r^2)
The percentage of variation in the response variable that is explained by the explanatory variable in the model
“____% of the variation in response variable can be explained by the linear relationship with explanatory variable.”
(0<=r^2<=1)
Closer to 0 = Weaker
Closer to 1 = Stronger
Reading Computer Regression Tables (LSRL)
Column 2, Row 2 = y-intercept = a
Column 2, Row 3 = slope = b
Column 1, Row 3 = x variable
Row 4, Value 1 = SD of residuals
Row 4, Value 2 = r^2 in % form
Residual Interpretation
“The actual [y-context] was [residual] [above/below] the predicted value when [x-context] = [#].”
Slope Interpretation
“The predicted [y-context] [increases/decreases] by [slope] for each additional [x-context].”
y-Intercept Interpretation
“The predicted [y-context] when [x=0 context] is [y-intercept].”
Standard Deviation Of Residuals (s) Interpretation
“The actual [y-context] is typically about [s] away from the value predicted by the LSRL.”