Course detail

Machine Learning

FSI-VSC Acad. year: 2026/2027 Summer semester

Learning outcomes of the course unit

Prerequisites

Planned learning activities and teaching methods

Assesment methods and criteria linked to learning outcomes

Language of instruction

Czech

Aims

The aim of the course is to familiarize students with machine learning methods and their applications in classification, regression, and clustering. Students will learn about both parametric and non-parametric classification and regression models, as well as key concepts such as error metrics, regularization, cross-validation, gradient descent, and modern approaches, including boosting and Gaussian mixture models. The course bridges theory and practice, focusing on the design and implementation of machine learning models.

Specification of controlled education, way of implementation and compensation for absences

The study programmes with the given course

Programme N-AIŘ-P: Applied Computer Science and Control, Master's
branch ---: no specialisation, 5 credits, compulsory

Type of course unit

 

Lecture

26 hours, optionally

Syllabus


  1. Introduction to Machine Learning: the machine learning model life cycle, learning paradigms and task types, model generalization, cross-validation.

  2. Data Preprocessing: missing values, outliers, normalization/standardization, categorical encoding, variable transformations, basic feature selection.

  3. Regression: linear and polynomial regression, the least squares method, gradient descent, loss functions, performance metrics, redundant variables (multicollinearity), regularization.

  4. Linear Regression-Based Classification, Logistic and Regularized Logistic Regression: basic loss functions and performance metrics for classifiers, the impact of class imbalance on training, classifier performance and evaluation.

  5. Support Vector Machines for Classification and Regression, Kernel Functions.

  6. Perceptron, Multilayer Perceptron, Extreme Learning Machine, Forward and Backpropagation, Gradient Descent Variants, Methods for Reducing Overfitting.

  7. Tree-Based Methods for Classification and Regression: decision trees, splitting criteria, pruning, ensemble methods.

  8. Memory-Based Methods for Classification and Regression: k-nearest neighbors, distance metrics, choice of k and weighting.

  9. Introduction to Probability Theory, Bayesian Classifier, Gaussian Discriminant Analysis, Naive Bayes Classifier.

  10. Gaussian Mixture Models, EM Algorithm.

  11. Clustering: k-means clustering, Gaussian mixture–based clustering, density-based clustering.

  12. Dimensionality Reduction, Boosting.

  13. Review.

Computer-assisted exercise

26 hours, compulsory

Syllabus


  1. Introduction to the programming environment.

  2. Data preprocessing.

  3. Linear and polynomial regression using least squares and regularized least squares.

  4. Classification of linearly separable data.

  5. Classification of nonlinearly separable data.

  6. Using multilayer perceptrons for classification and regression tasks.

  7. Using decision trees for classification and regression tasks.

  8. Using the k-nearest neighbors method for classification and regression tasks.

  9. Bayesian classifier, Gaussian discriminant analysis, and Naive Bayes classifier.

  10. Using Gaussian mixture models for classification and regression tasks.

  11. Data clustering.

  12. Dimensionality reduction, Boosting.

  13. Final assessment.