Institute of Computer Science
  1. Courses
  2. 2018/19 spring
  3. Machine Learning II (LTAT.02.004)
ET
Log in

Machine Learning II 2018/19 spring

Previous years: 2008 » 2012 » 2013 » 2014

  • Main
  • Lectures
  • Exercise sessions
  • Grading
  • Piazza forum
  • Upload

XIII. Elements of Statistical Learning Theory

Given by Sven Laur

Brief summary: Bias-variance dilemma revisited. Training and testing data as iid samples from the distribution of future challenges. Confidence bounds on cost estimations via Monte-Carlo integration. Why is does the training error underestimate future costs. Case study for the finite function set. Bias in training error and its connection to union-bound and multiple hypothesis testing. Consistency and identifiability properties. VC-dimension as a way to estimate bias in training error. Rademacher complexity and its connection to the bias in the training error. Limitations of statistical learning theory.

Slides: PDF

Video: UTTV (2016) UTTV (2015) UTTV (2014)

Literature:

  • Cristianini & Shawe-Taylor: Support Vector Machines: Generalisation Theory (Chapter 4)
  • Bartlett & Mendelson: Rademacher and Gaussian Complexities: Risk Bounds and Structural Results
  • David MacKay: Information Theory, Inference, and Learning Algorithms: Capacity of a Single Neuron

Complementary exercises:

  • Estimate the difference between training and test errors for different classifiers
    • Draw data from linearly separable model with some gaussian random shifts
    • Try various linear and non-linear classifiers
    • Plot the discrepancy as a function of training sample size
    • Draw data form more complex model that cannot be represented by predictor classes
    • Repeat the procedure
    • Estimate VC and Rademacher complexities and see if SLT bounds coincide with practice
  • Estimate the difference between training and test errors for different prediction algorithms
    • Draw the data form a linear model
    • Try various linear and non-linear predictors
    • Plot the discrepancy as a function of training sample size
    • Draw data form more complex model that cannot be represented by predictor classes
    • Repeat the procedure
    • Estimate VC and Rademacher complexities and see if SLT bounds coincide with practice

Free implementations:

  • None that we know
  • Institute of Computer Science
  • Faculty of Science and Technology
  • University of Tartu
In case of technical problems or questions write to:

Contact the course organizers with the organizational and course content questions.
The proprietary copyrights of educational materials belong to the University of Tartu. The use of educational materials is permitted for the purposes and under the conditions provided for in the copyright law for the free use of a work. When using educational materials, the user is obligated to give credit to the author of the educational materials.
The use of educational materials for other purposes is allowed only with the prior written consent of the University of Tartu.
Terms of use for the Courses environment