Model Evaluation
Last edited: October 10, 2025Some ideas of model validation
Cross Validation
Hold-out cross-validation
For instance, you can do:
- 70% for training
- 30% hold out cross validation for testing
But at very large dataset scales, the validation size can be capped at a fixed size (so you can hold out like 0.1% or something but still have 10k samples).
k-fold cross validation
- shuffle the data
- divide the data into \(k\) equal sized pieces
- repeatedly train the algorithm on 4/5 of the data, test on remaining 1/5
In practice people do 10 folds.
Model Selection
Last edited: October 10, 2025Model selection:
- choose some parameters
- evaluate every model using a Model Evaluation method of some kind
- and in production…
- k-fold or LOOCV: retrain best model on all data
- Hold-out cross-validation: optionally retrain, if you have time
A special case of model selection is feature selection:
- choose a subset of the most relevant features to train on
- note that power set is \(2^{m}\) in size; so instead of doing this we train \(O\qty(n)\) by starting out with an empty set, and then adding features sequentially that would give us the best performance
overfitting
Last edited: October 10, 2025consider something like a polynomial interpolation:
Interpolating polynomial (or most ML models in general) are smooth, and so interpolating between points will result in “overshooting” regional points and “bouncing around”
…as a function of parameters
At a fixed dataset, just increasing the number of parameters will increase
SU-CS229 OCT062025
Last edited: October 10, 2025Key Sequence
Notation
New Concepts
Important Results / Claims
Questions
Interesting Factoids
Decisions.jl
Last edited: October 10, 2025A general formulation of decision networks; track:
- track how distributions affect each other (i.e. generalize your problem into a Bayes Net)
- apply transformation of the edges to your new problem type structure
- transform back into your new formulation’s solver
- solve
