Cross-validation pitfalls when selecting and assessing regression and classification models

resampling
cross-validation
articles
Author

Max Kuhn

Published

April 10, 2014


Damjan Krstajic and friends have a great paper on pitfalls of cross-validation. Although the paper uses chemistry data, the meat of the article is broadly applicable. It does a great job of illustrating different resampling approaches and I learned more about double and nested cross-validation.

Figure 10 surprised me; I assumed that the precision in resampled estimates is mostly driven by the number of resamples. For example, a resampled estimate of the RMSE using 64 resamples has a standard error of sd/8 which is twice as good as one using 16 resamples (i.e. sd/4). In their work, the variation in 50 repeats of 10-fold CV are much better than 50 repeats of 10-fold nested CV.

Finally, the article has an excellent historical summary of the pivotal papers on this subject and does a great job of labeling and articulating the different goals that one might have when resampling predictive models.

(This article was originally posted at http://appliedpredictivemodeling.com)