C. Donovan
02 March 2018
combining models be good!
This makes some sense:
Our best model wasn't all that special, so why settle for one?
This idea also aligns with evidence/ideas elsewhere:
Sooo. today we'll:
We have seen and used resampling
We've also briefly discussed OOB validation
Concept is simple
Let's look at trees we've fitted already. How stable were they?
I'm neither surprised nor filled with confidence.
Bootstrapping was originally developed to help us with our logical statistical problem. Different samples will give different estimates! Statistics is all about quantifying uncertainty (and hence assessing confidence in “answers”)
Each row of our \( \mathbf{y} \) and \( \mathbf{X} \) will be chosen with probability \( 1/n \)
Claim: this mimics the original sampling of \( \mathbf{y} \) and \( \mathbf{X} \) from some population
What you need to know
A high-level description of the method is simple:
Note we have two classes of predictive model in mind: a quantitative response (i.e. regression type problems) or a categorical response (classification problems).
Following the original notation of Brieman 1996:
What?: bootstrap your data \( M \) times, fit to each to give \( M \) models. Want a prediction? You now get \( M \).
Note - the type of model is not specified, it can be anything.
(of course, they only published good results yeah?)
Requires only:
How many? You can use a validation dataset to check the generalisation error has plateaued for your \( M \)
Let's do one…
In a nutshell:
The aggregate of lots of unstable models is better than the, almost random, best one you get from one fit to all the data.