Statistical Learning Theory, Spring Semester 2019
Instructor
Prof. Dr. J.M. BuhmannAssistants
Luca CorinziaViktor Wegmayr
Alina Dubatovka
Carlos Cotrini Jimenez
Paolo Penna
Fabian Laumer
Ivan Ovinnikov
General Information
The ETHZ Course Catalogue information can be found here.
The course covers advanced methods of statistical learning. The fundamentals of Machine Learning as presented in the course "Introduction to Machine Learning" and "Advanced Machine Learning" are expanded and, in particular, the following topics are discussed:
- Theory of estimators: How can we measure the quality of a statistical estimator? We already discussed bias and variance of estimators very briefly, but the interesting part is yet to come.
- Variational methods and optimization: We consider optimization approaches for problems where the optimizer is a probability distribution. Concepts we will discuss in this context include:
- Maximum Entropy
- Information Bottleneck
- Deterministic Annealing
- Clustering: The problem of sorting data into groups without using training samples. This requires a definition of ``similarity'' between data points and adequate optimization procedures;
- Model Selection: We have already discussed how to fit a model to a data set in ML I, which usually involved adjusting model parameters for a given type of model. Model selection refers to the question of how complex the chosen model should be. As we already know, simple and complex models both have advantages and drawbacks alike;
- Statistical physics models: approaches for large systems approximate optimization, which originate in the statistical physics (free energy minimization applied to spin glasses and other models); sampling methods based on these models
Time and Place
Type | Time | Place | |
---|---|---|---|
Lectures | Mon 14-16, Tue 9-10 | HG | E 5 |
Exercises | Mon 16-18 | HG | E 5 |
Piazza website
Material
Past written Exams:
Exam 2016
Exam 2017
Exam 2018
Projects
Projects are small coding exercises that concern the implementation of an algorithm taught in the lecture/exercise class.There will be eight coding exercises, with a time span of two weeks per coding exercise. Each one of them will be graded as not passed or with a passing grade ranging from 4 to 6. The project part is passed if the student receives a passing grade in at least five coding exercises, and in that case the grade of the project part is the average of the five best coding exercises.
In order to be admitted to the exam the student has to pass the project part, and the final grade for the whole class is the weighted average 0.7 exam + 0.3 project. More details at the project repository.
Reading
- Prelimiary course script (ver. Mar 2019). This script has not been fully checked and thus comes without any guarantees, however, is good for getting oriented in the material.
- Duda, Hart, Stork: Pattern Classification, Wiley Interscience, 2000.
- Hastie, Tibshirani, Friedman: The Elements of Statistical Learning, Springer, 2001.
- L. Devroye, L. Gyorfi, and G. Lugosi: A probabilistic theory of pattern recognition. Springer, New York, 1996
Projects from the ISE group
Proposals
Web Acknowledgements
The web-page code is based (with modifications) on the one of the course on Machine Learning (Fall Semester 2013; Prof. A. Krause).