What are the key assumptions of linear regression?

Master CRINQ's Descriptive, Inferential, and Clinical Statistics with our practice test. Tackle multiple choice questions, each with detailed explanations, to ensure you're fully prepared. Ready for your exam!

Multiple Choice

What are the key assumptions of linear regression?

Explanation:
Linear regression relies on a specific set of assumptions about how the data relate to the model and how the residuals behave. The relationship between the predictors and the outcome should be linear, so the expected value of the outcome given the predictors is described by a straight-line (or plane in multiple dimensions) in the parameterized form. Observations must be independent, meaning one case does not provide information about another. The residuals—the differences between observed and predicted values—should have constant variance across all levels of the fitted values (homoscedasticity). The residuals should be approximately normally distributed, which supports valid confidence intervals and hypothesis tests for the coefficients, especially with smaller samples. Finally, the predictor variables should not be excessively correlated with each other (no multicollinearity); high correlation among predictors makes coefficient estimates unstable and inflates standard errors. These conditions together explain why this set of assumptions is the correct one. The other options mix up heteroscedasticity for homoscedasticity, claim normality of predictors (which isn’t an assumption), or state that linearity is the only requirement, all of which misstate what linear regression needs for reliable estimates and inference.

Linear regression relies on a specific set of assumptions about how the data relate to the model and how the residuals behave. The relationship between the predictors and the outcome should be linear, so the expected value of the outcome given the predictors is described by a straight-line (or plane in multiple dimensions) in the parameterized form. Observations must be independent, meaning one case does not provide information about another. The residuals—the differences between observed and predicted values—should have constant variance across all levels of the fitted values (homoscedasticity). The residuals should be approximately normally distributed, which supports valid confidence intervals and hypothesis tests for the coefficients, especially with smaller samples. Finally, the predictor variables should not be excessively correlated with each other (no multicollinearity); high correlation among predictors makes coefficient estimates unstable and inflates standard errors.

These conditions together explain why this set of assumptions is the correct one. The other options mix up heteroscedasticity for homoscedasticity, claim normality of predictors (which isn’t an assumption), or state that linearity is the only requirement, all of which misstate what linear regression needs for reliable estimates and inference.

Subscribe

Get the latest from Examzify

You can unsubscribe at any time. Read our privacy policy