This is an old version, view current version.

30.4 Bagging

When bootstrapping is carried through inference it is known as bootstrap aggregation, or bagging, in the machine-learning literature (Breiman 1996). In the simplest case, this involves bootstrapping the original data, fitting a model to each bootstrapped data set, then averaging the predictions. For instance, rather than using an estimate ˆσ from the original data set, bootstrapped data sets yboot(1),,yboot(N) are generated. Each is used to generate an estimate ˆσboot(n). The final estimate is ˆσ=1NNn=1ˆσboot(n). The same would be done to estimate a predictive quantity ˜y for as yet unseen data. ˆ˜y=1NNn=1ˆ˜yboot(n). For discrete parameters, voting is used to select the outcome.

One way of viewing bagging is as a classical attempt to get something like averaging over parameter estimation uncertainty.

References

Breiman, Leo. 1996. “Bagging Predictors.” Machine Learning 24 (2): 123–40.