READING 10 SIMPLE LINEAR REGRESSION Flashcards
(68 cards)
What is the primary purpose of simple linear regression?
(A) To determine if two variables are related, without quantifying the relationship.
(B) To explain the variation in a dependent variable using the variation in a single independent variable.
(C) To predict future values of an independent variable based on a dependent variable.
(D) To analyze the correlation between multiple independent variables.
(B) To explain the variation in a dependent variable using the variation in a single independent variable.
Simple linear regression aims to model how changes in one variable (independent) are associated with changes in another (dependent), thus explaining the dependent variable’s variation.
In a simple linear regression model, the variable whose variation is being explained is called the:
(A) Independent variable.
(B) Explanatory variable.
(C) Dependent variable.
(D) Predictor variable.
(C) Dependent variable.
The dependent variable is the outcome or response variable that we are trying to understand or predict. Its variation is what we are modeling.
In a simple linear regression model, the variable used to explain the variation in the dependent variable is called the:
(A) Response variable.
(B) Dependent variable.
(C) Endogenous variable.
(D) Independent variable.
(D) Independent variable.
The independent variable is the variable that is believed to influence or explain the changes in the dependent variable. It’s the “cause” in our simple linear model.
“Variation” in the context of linear regression refers to:
(A) The difference between the highest and lowest values of a variable.
(B) The standard deviation of a variable.
(C) The degree to which a variable differs from its mean value.
(D) The correlation between two variables.
(C) The degree to which a variable differs from its mean value.
Variation, often quantified by the sum of squared deviations from the mean, describes the spread or dispersion of the data points around the average.
Suppose you are trying to predict a company’s stock price using its earnings per share (EPS). In this scenario, the dependent variable is:
(A) Earnings per share (EPS).
(B) The relationship between stock price and EPS.
(C) The prediction error.
(D) The company’s stock price.
(D) The company’s stock price.
We are trying to predict the stock price, so it is the variable being explained (dependent). EPS is the factor we are using for the prediction (independent).
Another term often used to describe the independent variable in a regression analysis is:
(A) Residual.
(B) Intercept.
(C) Explanatory variable.
(D) Coefficient of determination.
(C) Explanatory variable.
The independent variable is used to explain the changes in the dependent variable, hence the term “explanatory variable.”
Another term often used to describe the dependent variable in a regression analysis is:
(A) Slope.
(B) Regressor.
(C) Predicted variable.
(D) Error term.
(C) Predicted variable.
The dependent variable is the one we are trying to predict using the regression model.
Understanding the difference between the dependent and independent variable is crucial because:
(A) It determines the scale of the regression coefficients.
(B) It dictates the direction of the hypothesized causal relationship being modeled.
(C) It affects the calculation of the correlation coefficient.
(D) It is only important for multiple regression, not simple linear regression.
(B) It dictates the direction of the hypothesized causal relationship being modeled.
The choice of which variable is dependent and which is independent reflects the assumed direction of influence. We are modeling how X affects Y, not the other way around.
In the simple linear regression model Yi = b0+b1(Xi)+ϵi, the term ϵi represents the:
(A) Predicted value of the dependent variable.
(B) Slope of the regression line.
(C) Residual or error term for the i-th observation.
(D) Intercept of the regression line.
(C) Residual or error term for the i-th observation.
The primary goal of the Ordinary Least Squares (OLS) method in linear regression is to:
(A) Maximize the correlation between the independent and dependent variables.
(B) Minimize the sum of the absolute errors.
(C) Minimize the sum of the squared errors.
(D) Ensure that the residuals are normally distributed.
(C) Minimize the sum of the squared errors.
OLS estimates the regression coefficients by finding the line that minimizes the sum of the squared differences between the actual and predicted values of the dependent variable (the sum of squared errors, SSE)
The intercept (b0) in a simple linear regression represents the:
(A) Change in the dependent variable for a one-unit change in the independent variable.
(B) Predicted value of the dependent variable when the independent variable is zero.
(C) Average value of the dependent variable.
(D) Standard deviation of the dependent variable.
(B) Predicted value of the dependent variable when the independent variable is zero.
The intercept (b0) is the estimated value of the dependent variable (Y) when the independent variable (X) is equal to zero.
The slope coefficient (b1) in a simple linear regression represents the:
(A) Predicted value of the dependent variable when the independent variable is one.
(B) Baseline value of the dependent variable when the independent variable is zero.
(C) Change in the dependent variable for a one-unit change in the independent variable.
(D) Average change in the independent variable.
(C) Change in the dependent variable for a one-unit change in the independent variable.
The slope coefficient (b1) quantifies the change in the dependent variable (Y) associated with a one-unit increase in the independent variable (X).
The sum of the residuals (∑(Yi − Y^i)) in an OLS regression is typically:
(A) Minimized and always positive.
(B) Maximized.
(C) Equal to zero.
(D) Equal to the sum of squared errors.
(C) Equal to zero.
One of the properties of OLS is that the sum of the residuals (the differences between the actual and predicted values) is equal to zero.
If the slope coefficient in a simple linear regression is positive, it indicates that:
(A) An increase in the independent variable is associated with a decrease in the dependent variable.
(B) There is no linear relationship between the two variables.
(C) An increase in the independent variable is associated with an increase in the dependent variable.
(D) The intercept of the regression line is also positive.
(C) An increase in the independent variable is associated with an increase in the dependent variable.
A positive slope coefficient signifies a direct, positive linear relationship: as the independent variable increases, the dependent variable tends to increase as well.
The formula for the estimated slope coefficient is directly related to the:
(A) Variance of the dependent variable.
(B) Covariance between the independent and dependent variables and the variance of the independent variable.
(C) Correlation coefficient squared.
(D) Sum of squared errors.
(B) Covariance between the independent and dependent variables and the variance of the independent variable.
The intercept term in a regression model should be interpreted with caution when:
(A) The slope coefficient is statistically significant.
(B) The independent variable never takes on a value close to zero within the observed data range.
(C) The correlation between the variables is high.
(D) The sample size is large.
(B) The independent variable never takes on a value close to zero within the observed data range.
If the value of zero for the independent variable is far outside the range of the observed data, the intercept may not have a meaningful real-world interpretation. The linear relationship might not hold true at such extreme values.
The difference between the observed value of the dependent variable (Yi) and the predicted value (Y^i) is known as the:
(A) Explained variation.
(B) Total variation.
(C) Residual or error.
(D) Regression coefficient.
(C) Residual or error.
In a simple linear regression, if there is a perfect positive linear relationship between the independent and dependent variables, the sum of squared errors (SSE) will be:
(A) Positive and large.
(B) Positive and small.
(C) Equal to zero.
(D) Equal to the total variation.
(C) Equal to zero.
A perfect linear relationship means all data points lie exactly on the regression line. In this ideal scenario, there are no deviations between the actual and predicted values, resulting in a sum of squared errors (SSE) of zero.
In a simple linear regression where a stock’s return is regressed against the market return, a slope coefficient of 1.2 indicates that for every 1% increase in the market return, the stock’s return is expected to:
(A) Decrease by 1.2%.
(B) Increase by 1.2%.
(C) Remain unchanged.
(D) Increase by 0.2%.
(B) Increase by 1.2%.
The slope coefficient represents the change in the dependent variable (stock’s return) for a one-unit change in the independent variable (market return). A slope of 1.2 means a 1% increase in the market return is associated with a 1.2% expected increase in the stock’s return.
In a regression of bond yield on the policy interest rate, a slope coefficient of 0.8 implies that a 100 basis point increase in the policy interest rate is expected to lead to a change in the bond yield of:
(A) An increase of 0.8 basis points.
(B) An increase of 80 basis points.
(C) A decrease of 0.8 basis points.
(D) A decrease of 80 basis points.
(A) An increase of 0.8 basis points.
The slope of 0.8 means for every one-unit (1 basis point in this case) increase in the policy interest rate, the bond yield is expected to increase by 0.8 units (0.8 basis points). Therefore, a 100 basis point increase would lead to an expected increase of 0.8 * 100 = 80 basis points.
When interpreting a regression intercept, it is most important to consider:
(A) The magnitude of the slope coefficient.
(B) Whether the value of zero for the independent variable is within the relevant data range.
(C) The correlation coefficient between the variables.
(D) The statistical significance of the slope coefficient.
(B) Whether the value of zero for the independent variable is within the relevant data range
If the independent variable rarely or never takes on a value near zero in the observed data, the intercept may not have a practical or meaningful interpretation within the context of the model.
An intercept of -0.5% in a regression of a company’s sales on advertising expenditure suggests that if advertising expenditure is zero, the company’s sales are expected to be:
(A) 0.5% higher than average.
(B) Zero.
(C) 0.5% lower than average.
(D) -0.5%.
(D) -0.5%.
The intercept is the predicted value of the dependent variable (sales) when the independent variable (advertising expenditure) is zero. Therefore, expected sales would be -0.5%. Note that in a real-world scenario, negative sales might not be economically meaningful, highlighting the caution needed in interpreting intercepts.
A slope coefficient of -0.3 in a regression of product demand on price indicates that a $1 increase in price is expected to lead to a:
(A) Decrease in demand of 0.3 units.
(B) Increase in demand of 0.3 units.
(C) No change in demand.
(D) Decrease in demand of 3 units.
(A) Decrease in demand of 0.3 units.
A negative slope indicates an inverse relationship. A slope of -0.3 means that for every $1 increase in price, the demand for the product is expected to decrease by 0.3 units.
The magnitude of the slope coefficient in a simple linear regression directly indicates the:
(A) Strength of the linear relationship.
(B) Statistical significance of the relationship.
(C) Sensitivity of the dependent variable to a one-unit change in the independent variable.
(D) Proportion of the total variation in the dependent variable explained by the model.
(C) Sensitivity of the dependent variable to a one-unit change in the independent variable.
The slope coefficient’s magnitude quantifies how much the dependent variable is expected to change for each unit change in the independent variable. It reflects the sensitivity of Y to X.