Explores overfitting, cross-validation, and regularization in machine learning, emphasizing model complexity and the importance of regularization strength.
Covers the basics of Machine Learning, including recognizing hand-written digits, supervised classification, decision boundaries, and polynomial curve fitting.
Covers overfitting, regularization, and cross-validation in machine learning, exploring polynomial curve fitting, feature expansion, kernel functions, and model selection.
Delves into the trade-off between model flexibility and bias-variance in error decomposition, polynomial regression, KNN, and the curse of dimensionality.