Boosting and Greedy Algorithm (III):

Josh Day
March 24, 2014

Sections 12.6 and 12.7

Introduction

  • Review L2 Boosting
  • Nonparametric curve estimation
    • L2 Boosting using kernel estimators
  • L2 Boosting for high-dimensional linear models
    • Connection with LASSO
  • Forward Selection and Orthogonal Matching Pursuit

L2 Boosting

Review - What is L2 Boosting?

  • Functional gradient descent using the squared error loss.
  • This amounts to repeated fitting of residuals. [1]

L2 Boosting Example

library(MASS)          # Data (Boston)
library(rpart)         # Trees
library(ipred)         # Bagging
library(randomForest)  # Random Forest
library(gbm)           # L2 Boosting

n <- dim(Boston)[1]
train <- sample(1:n, .66*n)
Boston.train <- Boston[train,]
Boston.validate <- Boston[-train,]

Orthogonal Matching Pursuit

[1] [1]

Forward Selection and Orthogonal Matching Pursuit

  • OMP is “Boosting version” of Forward Selection
    • Both methods fit \( p-|S^{[m-1]}| \) models in each iteration
      • Forward Selection refits full models
      • OMP fits univariate predictors to residuals
  • OMP is much more efficient in computation
  • OMP has properties comparable to L1 penalty methods

References

  1. Bühlmann, Peter, and Geer. Statistics for High-Dimensional Data. Heidelberg New York: Springer, 2011.
  2. Hastie, Trevor, Robert Tibshirani, and J. H. Friedman. The Elements of Statistical Learning. New York: Springer, 2009.