⇤ ← Revision 1 as of 2017-01-12 20:02:29
Size: 349
Comment:
|
← Revision 2 as of 2017-01-15 00:11:07 ⇥
Size: 1292
Comment: Fleshed out notes
|
Deletions are marked like this. | Additions are marked like this. |
Line 7: | Line 7: |
* Where (in the domain of the data) is the model useful? * Prediction versus inference. One opinion is that a good prediction model also ought to be explainable like a good inference model. * Parametric versus non-parametric. Parametric is great whenever you can determine the shape of your data; otherwise, non-parametric may be the only way to go. Had a side-discussion about Lasso. David had read a paper demonstrating that using Lasso to select the best variables to include in a model uses one degree of freedom per variable, in order, within about 6 significant figures. This is contrary to the belief that selecting variables would be like conducting consecutive experiments, and require some ''p''-value adjustments (see the Green Jelly Bean XKCD comic). Back to chapter 2: * ISLR's treatment of model accuracy in train versus test is top-notch. * Its discussion about the bias/variance trade-off is also good. |
Jan 12, 2017
Review of Chapter 2
Chatted about various highlights of chapter 2. Went through several of the R examples that we hadn't seen before, especially contour, persp, and identify.
- Where (in the domain of the data) is the model useful?
- Prediction versus inference. One opinion is that a good prediction model also ought to be explainable like a good inference model.
- Parametric versus non-parametric. Parametric is great whenever you can determine the shape of your data; otherwise, non-parametric may be the only way to go.
Had a side-discussion about Lasso. David had read a paper demonstrating that using Lasso to select the best variables to include in a model uses one degree of freedom per variable, in order, within about 6 significant figures. This is contrary to the belief that selecting variables would be like conducting consecutive experiments, and require some p-value adjustments (see the Green Jelly Bean XKCD comic).
Back to chapter 2:
- ISLR's treatment of model accuracy in train versus test is top-notch.
- Its discussion about the bias/variance trade-off is also good.
Next Time
- Thursday, February 9, 2017
Be ready to discuss Chapter 3 of AnIntroductionToStatisticalLearning