Please enable JavaScript.
Coggle requires JavaScript to display documents.
Subset Selection, Regularization/ Shrinkage Methods, Dimension Reduction,…
Subset Selection
-
Best Subset Selection
Algorithm
- For k = 1, 2,...p:
(a) Fit all (p k) models that contain k predictors.
(b) Pick the best among (p k) models, and call it Mk.
best = having the smallest RSS, or largest R2.
- M0=null model with no predictors predicting the sample mean for each observation.
- Select single best model from M0,...,Mp using cross-validated prediction error, Cp (AIC), BIC, adjusted R-squared.
Deviance Metric
-
The smaller the deviance, the better the fit
-
-
Stepwise Selection
Forward
Algorithm
- "" ~ Best subset selection;
- For k = 0,...,p − 1:
(a) Consider all p − k models that augment predictors in Mk with one additional predictor.
(b) Choose the best among p − k models, and call it M(k+1)
best = having smallest RSS or highest R^2.
- "" ~ Best subset selection;
Not guaranteed to find the best possible model out of all 2^p models containing subsets of p predictors
e.g. Best possible 1-variable model contains X1 but
best possible 2-variable model contains X2 & X3 =>
Forward Stepwise Selection Fails to select best possible 2-variable model cos it must contain X1 w. 1 additional variable
-
Hybrid approach
After adding each new variable, Remove any variables that no longer provide an improvement in the model fit
Backward
Algorithm
- Mp=full model with all p predictors
- For k = p, p − 1,..., 1:
(a) Consider all k models that contain all but one of the predictors in Mk, for a total of k − 1 predictors.
(b) Choose the best among k models, called M(k−1)
best = having smallest RSS or highest R-squared.
- "" ~ Best subset selection
-
-
-
Model Selection Metrics
Cp estimate of test MSE
-
σˆ2=estimate of variance of model error w. all predictors
d = number of predictors; n = number of observations
Adds increasing penalty of 2dˆσ^2 to the decreasing training RSS as more predictors are added not to underestimate the test error
-
-
-
-
-
Method 1: Indirectly estimate test error by making adjustment to the training error to account for the bias due to overfitting.
-
-
Method 2: Directly estimate the test error, using either validation set approach or cross-validation approach
-
-
-
-
-
-
-