Category:1000 Data Analysis MCQDate: Published: November 8, 2025Posted by: MCQs Generator
1 min read
[flat_pm id="7169"]
These 50 MCQs covers fundamental concepts in regression analysis, including linear and multiple regression, assumptions, diagnostics, and interpretation. Ideal for students and professionals in data analysis to test understanding of predictive modeling techniques.
1. What does linear regression model the relationship between?
a) Two categorical variables
b) A dependent variable and one or more independent variables
c) Only independent variables
d) Time series only
✅ Correct Answer: b) A dependent variable and one or more independent variables
📝 Explanation:
Linear regression predicts a continuous outcome (Y) as a linear function of predictors (X).
2. The slope coefficient in simple linear regression represents
a) The intercept
b) Change in Y for a one-unit change in X
c) The correlation
d) The error term
✅ Correct Answer: b) Change in Y for a one-unit change in X
📝 Explanation:
β1 = ΔY / ΔX, holding other factors constant.
3. R-squared measures
a) Total variance
b) Proportion of variance in Y explained by X
c) Residual sum of squares
d) Standard error
✅ Correct Answer: b) Proportion of variance in Y explained by X
📝 Explanation:
Ranges from 0 to 1; higher values indicate better fit.
4. The assumption of linearity in regression means
a) Constant variance
b) Relationship between X and Y is linear
c) No multicollinearity
d) Independent errors
✅ Correct Answer: b) Relationship between X and Y is linear
📝 Explanation:
Verified via scatter plots or residual plots.
5. Homoscedasticity refers to
a) Constant variance of residuals
b) Normal distribution of errors
c) No autocorrelation
d) Linear trend
✅ Correct Answer: a) Constant variance of residuals
📝 Explanation:
Tested with Breusch-Pagan; violations suggest heteroscedasticity.
6. In multiple regression, multicollinearity is detected using
a) VIF (Variance Inflation Factor)
b) R-squared
c) F-test
d) t-test
✅ Correct Answer: a) VIF (Variance Inflation Factor)
📝 Explanation:
VIF > 5-10 indicates high collinearity among predictors.
7. The intercept in regression is
a) Expected Y when all X=0
b) Slope
c) Error
d) Correlation
✅ Correct Answer: a) Expected Y when all X=0
📝 Explanation:
β0; may lack interpretation if X=0 is outside range.
8. Residuals are
a) Predicted values
b) Observed minus predicted values
c) Independent variables
d) Coefficients
✅ Correct Answer: b) Observed minus predicted values
📝 Explanation:
Used for diagnostics; should be randomly distributed.
9. The F-test in regression tests
a) Overall model significance
b) Individual coefficients
c) Linearity
d) Normality
✅ Correct Answer: a) Overall model significance
📝 Explanation:
H0: all β=0; low p-value indicates model explains variance.
10. t-test for coefficients tests
a) H0: β=0
b) H0: R²=0
c) Intercept only
d) Variance equality
✅ Correct Answer: a) H0: β=0
📝 Explanation:
Significance of individual predictors.
11. Adjusted R-squared accounts for
a) Number of predictors
b) Sample size
c) Both a and b
d) Residuals
✅ Correct Answer: c) Both a and b
📝 Explanation:
Penalizes adding irrelevant variables; better for model comparison.
12. Outliers in regression can be detected using
a) Leverage and Cook's distance
b) R-squared
c) Slope
d) Intercept
✅ Correct Answer: a) Leverage and Cook's distance
📝 Explanation:
High values indicate influential points affecting fit.
13. The standard error of the estimate is
a) Root mean squared error
b) Average residual
c) Variance of Y
d) Correlation
✅ Correct Answer: a) Root mean squared error
📝 Explanation:
Measures prediction accuracy; √(SSE/(n-k-1)).
14. Autocorrelation in residuals is tested with
a) Durbin-Watson test
b) Breusch-Pagan
c) Shapiro-Wilk
d) VIF
✅ Correct Answer: a) Durbin-Watson test
📝 Explanation:
Values near 2 indicate no serial correlation; common in time series.
15. Logistic regression is used for
a) Continuous outcomes
b) Binary or categorical outcomes
c) Time series
d) Clustering
✅ Correct Answer: b) Binary or categorical outcomes
📝 Explanation:
Models log-odds; uses sigmoid function.
16. In ridge regression, the penalty is on
a) Sum of squared coefficients
b) Absolute coefficients
c) Interactions
d) Residuals
✅ Correct Answer: a) Sum of squared coefficients
📝 Explanation:
L2 regularization; reduces multicollinearity.
17. Lasso regression uses
a) L1 penalty
b) L2 penalty
c) No penalty
d) Quadratic penalty
✅ Correct Answer: a) L1 penalty
📝 Explanation:
Sum of absolute coefficients; performs variable selection.