Explores computing density of states and Bayesian inference using importance sampling, showcasing lower variance and parallelizability of the proposed method.
Delves into the trade-off between model flexibility and bias-variance in error decomposition, polynomial regression, KNN, and the curse of dimensionality.
Covers overfitting, regularization, and cross-validation in machine learning, exploring polynomial curve fitting, feature expansion, kernel functions, and model selection.