Machine Learning 2
Data is displayed for the academic year: 2025./2026.
Lecturers
Course Description
Machine learning is a branch of artificial intelligence that deals with the design of algorithms that improve their efficiency based on empirical data. The key challenge of machine learning is how to ensure that the model generalizes well. The theory of statistical machine learning studies the generalization capabilities of algorithms with the aim of better understanding existing algorithms and developing better algorithms. The course provides an introduction to the theory of statistical learning and an overview of its applications to classical and modern machine learning algorithms. The course content is divided into three parts. The first part studies the basic concepts and tools of statistical learning theory. In the second part, these tools are applied to classical classification and regression algorithms. The third part studies more advanced techniques of statistical learning theory and applications to deep learning models.
Prerequisites
Learning outcomes of the course "Machine Learning 1" (FER) or a similar course. Learning outcomes of the courses "Mathematical Analysis 1", "Mathematical Analysis 2", "Discrete Mathematics", "Linear Algebra", "Probability and Statistics", and "Information Theory", or similar courses. Good knowledge of programming in Python. Desirable, but not essential: learning outcomes of the course "Deep Learning 1" (FER) or a similar course.
Study Programmes
University graduate
[FER3-EN] Control Systems and Robotics - profile
Elective courses
(3. semester)
Learning Outcomes
- Explain the fundamental concepts of statistical learning theory, including PAC learnability, empirical risk minimization, and the bias-variance tradeoff
- Apply core statistical learning theory principles to explain the behavior and properties of regularization techniques and optimization algorithms
- Analyze the generalization capabilities of ensemble methods like boosting and random forests, relating them to concepts such as stability and complexity
- Differentiate between various complexity measures (e.g., VC dimension, Rademacher complexity) and derive their application in establishing generalization bounds for learning models
- Evaluate advanced theoretical frameworks such as PAC-Bayes theory and compression bounds, assessing their relevance for understanding generalization in complex learning scenarios
- Explain the theoretical assumptions, advantages, and limitations of sparse kernel machines
- Explain the main approaches to semisupervised machine learning and list their advantages and disadvantages
- Synthesize current theoretical understanding of deep learning phenomena, including training dynamics, and implicit regularization
- Explain open problems in deep learning generalization and efficiency, informed by advanced statistical learning theory
Forms of Teaching
Lectures
Lectures are given for 13 weeks in one three-hour session per week.
Seminars and workshopsPresenting one selected scientific paper.
ExercisesRecitations are given in one-hour sessions as needed, spanning 13 weeks.
Independent assignmentsPreparatory reading of textbook chapters. Reading selected scientific papers.
LaboratoryProgramming assignments, demonstrated to the instructor or teaching assistant.
Week by Week Schedule
- Introduction and course organization. Overview of statistical learning theory. Course goals and expectations.
- PAC learning framework. Generalization bounds. Sample complexity. No free lunch theorem.
- Empirical risk minimization. Bias-variance tradeoff. Model selection. Uniform convergence.
- VC dimension. The fundamental theorem of statistical learning.
- Regularization techniques. L1 and L2 regularization. Early stopping. Overfitting and underfitting revisited.
- Optimization for machine learning. Gradient descent variants. Stochastic gradient descent. Convergence properties.
- Kernel methods and reproducing kernel Hilbert spaces (RKHS). Theoretical foundations of kernels. Feature maps.
- Midterm exam
- Weak learnability. Boosting. AdaBoost algorithm. Gradient boosting.
- Decision trees. Random forests. Ensemble methods. Sample complexity.
- Rademacher complexity. VC dimension. Generalization bounds for deep learning.
- PAC-Bayes theory. Derivations and applications. Bayesian neural networks.
- Compression bounds. Minimum description length (MDL) principle. Learnability from a compression perspective.
- Theoretical foundations of modern deep learning. Training dynamics. Implicit regularization.
- Final exam
Literature
General
ID 223750
Winter semester
5 ECTS
L1 e-Learning
30 Lectures
0 Seminar
15 Exercises
15 Laboratory exercises
0 Project laboratory
0 Physical education excercises
Grading System
Excellent
Very Good
Good
Sufficient
Pristupačnost