Matthieu Bloch
April 9, 2020
| Method | Parameter |
|---|---|
| Polynomial regression | polynomial degree \(d\) |
| Ridge regression | \(\lambda\) |
| SVMs | margin violation constraint \(C\) |
| Kernel methods | kernel choice |
| \(K\) nearest neighbor | \(K\) |
In addition to training data \(\calD\), suppose we have access to another validation set \(\calV\eqdef\{(\bfx_i,y_i\}_{i=1}^K\)
Assume \(h\) selected from training set \(\calD\) and use the validation to form an estimate \[\widehat{R}_{\textsf{val}}(h)\eqdef\frac{1}{K}\sum_{i=1}^K\ell(h(\bfx_i),y_i)\]
Using validation for model selection
Effect of bias
Dilemma: we would like \(R(h)\approx R(h^{-})\approx\widehat{R}_{\textsf{val}}(h^{-})\)