Treffer: Overfitting, Model Tuning, and Evaluation of Prediction Performance

Title:
Overfitting, Model Tuning, and Evaluation of Prediction Performance
Contributors:
Bill and Melinda Gates Foundation
Source:
Multivariate Statistical Machine Learning Methods for Genomic Prediction ; page 109-139 ; ISBN 9783030890094 9783030890100
Publisher Information:
Springer International Publishing
Publication Year:
2022
Document Type:
Buch book part
Language:
English
ISBN:
978-3-030-89009-4
978-3-030-89010-0
3-030-89009-0
3-030-89010-4
DOI:
10.1007/978-3-030-89010-0_4
Accession Number:
edsbas.F3E08ADE
Database:
BASE

Weitere Informationen

The overfitting phenomenon happens when a statistical machine learning model learns very well about the noise as well as the signal that is present in the training data. On the other hand, an underfitted phenomenon occurs when only a few predictors are included in the statistical machine learning model that represents the complete structure of the data pattern poorly. This problem also arises when the training data set is too small and thus an underfitted model does a poor job of fitting the training data and unsatisfactorily predicts new data points. This chapter describes the importance of the trade-off between prediction accuracy and model interpretability, as well as the difference between explanatory and predictive modeling: Explanatory modeling minimizes bias, whereas predictive modeling seeks to minimize the combination of bias and estimation variance. We assess the importance and different methods of cross-validation as well as the importance and strategies of tuning that are key to the successful use of some statistical machine learning methods. We explain the most important metrics for evaluating the prediction performance for continuous, binary, categorical, and count response variables.