Title | Model Evaluation | ![]() |
Duration | 60 | |
Module | A | |
Lesson Type | Tutorial | |
Focus | Technical - Foundations of AI | |
Topic | Foundations of AI |
model evaluation, cross-validation, hyperparameter optimization,
None.
None.
None.
None.
The materials of this learning event are available under CC BY-NC-SA 4.0.
Prepare a Jupyter notebook environment with pandas, matplotlib, numpy and scikit-learn packages
Duration (min) | Description | Concepts |
---|---|---|
5 | Introduction to model evaluation | empirical error, predictive and generalization performance |
5 | Training a simple classifier | MLP, hyperparameters |
10 | Evaluating a classifier | confusion matrix, accuracy, TPR, FPR, precision, misclassification rate, F1 score |
10 | ROC/PR curves and their interpretation | decision boundary, ROC curve, PR curve, AUC |
10 | Underfitting and overfitting | training and test error |
10 | Cross-validation and hyperparameter optimization | validation set, validation error, 5-fold cross-validation |
10 | Evaluation of regression models | MSE, RMSE, MAE |
Click here for an overview of all lesson plans of the master human centred AI
Please visit the home page of the consortium HCAIM
![]() |
The Human-Centered AI Masters programme was co-financed by the Connecting Europe Facility of the European Union Under Grant №CEF-TC-2020-1 Digital Skills 2020-EU-IA-0068. The materials of this learning event are available under CC BY-NC-SA 4.0
|
The HCAIM consortium consists of three excellence centres, three SMEs and four Universities |