All lecture locations are listed on p. 4 of the first set of slides.

Course announcements will be posted on the **mailing list**.

This page will contain slides and detailed notes for the kernel part of the course. The assignment may also be found here (at the bottom of the page). Note that the slides will be updated as the course progresses, and I modify them to answer questions I get in the classes. I'll put the date of last update next to each document - be sure to get the latest one. Let me know if you find errors.

There are sets of practice exercises and solutions further down the page (after the slides).

See David Silver's **page** for the reinforcement learning part of the course.

- Definition of a kernel, how it relates to a feature space
- Combining kernels to make new kernels
- The reproducing kernel Hilbert space
- Applications: difference in means, kernel PCA, kernel ridge regression

Lectures 4, 5, 6, and 7 **slides** and **notes**, last modified 23 Feb 2016

- Distance between means in RKHS, integral probability metrics, the maximum mean discrepancy (MMD), two-sample tests
- Choice of kernels for distinguishing distributions, characteristic kernels
- Covariance operator in RKHS: proof of existence, definition of norms (including HSIC, the Hilbert-Schmidt independence criterion)
- Application of HSIC to independence testing
- Application of HSIC to feature selection, taxonomy discovery.
- Introduction to independent component analysis, kernel ICA

Lecture 8 **slides** and **notes**, last modified 26 Jan 2016

- Introduction to convex optimization
- The representer theorem
- Large margin classification, support vector machines for clasification

Lecture 9 **slides**, lecture 10 **slides** , and **notes**, last modified 20 Mar 2013

- Metric, normed, and unitary spaces, Cauchy sequences and completion, Banach and Hilbert spaces
- Bounded linear operators and the Riesz Theorem
- Equivalent notions of an RKHS: existence of reproducing kernel, boundedness of the evaluation operator
- Positive definiteness of reproducing kernels, the Moore-Aronszajn Theorem
- Mercer's Theorem for representing kernels

Supplementary lecture **slides**, last modified 22 Mar 2012

- Loss and risk, estimation and approximation error, a new interpretation of MMD
- Why use an RKHS: comparison with other function classes (Lipschitz and bounded Lipschitz)
- Characteristic kernels and universal kernels

Dates:

- Free schedule

Included in selections:

Deep Learning

Good materials on deep learning.

Good materials on deep learning.

Machine Learning

Machine learning: from the basics to advanced topics. Includes statistics...

Machine learning: from the basics to advanced topics. Includes statistics...

More from 'Computer Science':

Capstone Exam in Statistics and Data Science

Solidify and demonstrate your knowledge and abilities in probability, data analysis...

Solidify and demonstrate your knowledge and abilities in probability, data analysis...

Network Science

The course is an interdisciplinary course, focused on the emerging science of...

The course is an interdisciplinary course, focused on the emerging science of...

Gameplay Programming for Video Game Designers

Learn how to create the best gameplay by understanding algorithmic thinking...

Learn how to create the best gameplay by understanding algorithmic thinking...

Introducing Robotics: Making Robots Move

The world needs people who understand how to get robots moving.

The world needs people who understand how to get robots moving.

Python in High Performance Computing

Learn how to analyse Python programmes and identify performance barriers to...

Learn how to analyse Python programmes and identify performance barriers to...

© 2013-2019