Course detail
Neural networks and Machine Learning
FSI-VSC Acad. year: 2025/2026 Summer semester
The course provides an introduction to the theory and methods of machine learning, focusing on their application in solving classification, regression, and clustering tasks.
Supervisor
Department
Learning outcomes of the course unit
Prerequisites
Basic knowledge of statistics, optimization, and programming is expected.
Planned learning activities and teaching methods
Assesment methods and criteria linked to learning outcomes
Knowledge and skills are verified by credit and examination. Credit requirements: elaboration of given tasks. Attendance at lectures is recommended, while attendance at practical sessions is mandatory. Practical sessions that a student is unable to attend in the regular term can be made up during a substitute term. The exam is oral and covers the entire course material.
Language of instruction
Czech
Aims
The aim of the course is to familiarize students with machine learning methods and their applications in classification, regression, and clustering. Students will learn about both parametric and non-parametric classification and regression models, as well as key concepts such as error metrics, regularization, cross-validation, gradient descent, and modern approaches, including boosting and Gaussian mixture models. The course bridges theory and practice, focusing on the design and implementation of machine learning models.
Specification of controlled education, way of implementation and compensation for absences
The study programmes with the given course
Programme N-AIŘ-P: Applied Computer Science and Control, Master's
branch ---: no specialisation, 5 credits, compulsory
Type of course unit
Lecture
26 hours, optionally
Syllabus
- Introduction, supervised and unsupervised learning, regression vs. classification, dataset splitting, error metrics, loss functions, cross-validation, overfitting, regularization.
- Linear regression, least squares method, gradient descent, regularized least squares method.
- Linear classification, logistic regression, regularized logistic regression.
- Non-parametric models, nearest neighbors method.
- Decision trees for classification and regression problems.
- Generative models for classification, Bayes classifier, Gaussian discriminant analysis.
- Naive Bayes classifier.
- Support vector machines, kernel functions.
- Clustering, k-means clustering.
- Gaussian mixture models.
- Dimensionality reduction, boosting.
- Mathematical model of a neuron, activation functions, multilayer perceptron, forward and backward propagation.
- Feedforward and multilayer neural networks, recurrent networks, topologically organized neural networks.
Computer-assisted exercise
26 hours, compulsory
Syllabus
- Introduction to the programming environment.
- Least squares method and regularized least squares method.
- Linear classification, logistic regression, regularized logistic regression.
- Nearest neighbors method.
- Decision trees for classification and regression problems.
- Bayes classifier, Gaussian discriminant analysis.
- Naive Bayes classifier.
- Support vector machines.
- K-means clustering.
- Gaussian mixture models.
- Dimensionality reduction, boosting.
- Multilayer perceptron.
- Final assessment.