True. Forward stepwise selection builds up models by sequentially adding predictors. Thus, the (k+1)-variable model must contain all predictors from the k-variable model plus one additional predictor.
False. Backward stepwise selection starts with all predictors and removes one at a time. Hence, the (k+1)-variable model may contain different predictors than the k-variable model.
False. Forward and backward stepwise selections do not necessarily produce nested models across techniques.
False. Again, models from forward and backward selection are not necessarily nested in each other.
False. Best subset selection evaluates all combinations and does not require that a k-variable model be nested within a (k+1)-variable model.
Correct Answer: iii. Less flexible and hence will give improved
prediction accuracy when its increase in bias is less than its decrease
in variance.
Explanation: The lasso is less flexible than least squares because it constrains the sum of the absolute values of the coefficients. This increases bias but decreases variance. Improved prediction accuracy occurs when the increase in bias is less than the decrease in variance.
Correct Answer: iii. Less flexible and hence will give improved prediction accuracy when its increase in bias is less than its decrease in variance.
Explanation: Like lasso, ridge regression adds a penalty term to the loss function and is also less flexible than least squares. The same reasoning about bias-variance tradeoff applies.
Correct Answer: i. More flexible and hence will give improved prediction accuracy when its increase in bias is less than its decrease in variance.
Explanation: Non-linear methods are more flexible than least squares, often reducing bias at the cost of increased variance. Prediction improves when the bias decrease outweighs the variance increase.
Correct Answer: iv. Steadily decrease.
Explanation: As s increases, the constraint on the sum of the absolute values of coefficients is relaxed, allowing more flexibility in fitting the training data, hence training RSS will steadily decrease.
Correct Answer: ii. Decrease initially, and then eventually start increasing in a U shape.
Explanation: Initially, increasing s improves the fit and reduces bias, lowering test RSS. Beyond a point, increased model complexity leads to over-fitting and increased variance, making test RSS increase again. This forms a U-shaped curve.
Correct Answer: iv. Steadily decrease.
Explanation: Increasing s leads to more model flexibility, which increases the model variance steadily.
Correct Answer: iv. Steadily decrease.
Explanation: Increasing s reduces the penalty, which lowers bias. Hence, squared bias steadily decreases.
Correct Answer: v. Remain constant
Explanation: Irreducible error is due to inherent noise in the data and does not depend on the model. It remains constant regardless of s.