GI13/4C60: Advanced Topics in Machine Learning (Spring 2005)


Class Times: Tuesday, 14:00-17:00, Friday 10:00-13:00
Location: Computer Science Dept, Room 104
Instructors: Massimiliano Pontil
Email Contact : gi13@cs.ucl.ac.uk

Course description

The course presents the elements of kernel-based methods from a machine learning perspective. It introduces the theoretical basis for studying these methods (theory of positive definite kernels, associated reproducing kernel Hilbert spaces and techniques to construct kernel functions) and present selected topics in this area. This includes learning algorithms such as regularization networks, support vector machines, kernel principal component analysis, kernel canonical correlation analysis, anomaly detection, etc., as well as a discussion of the value of these algorithms for applications. The material is primarily based on a recent book and on research publications.

Students should gain an in-depth familiarity with kernel methods and be able to individually read and discuss in class research papers in the field.

Prerequisites

A good background in university-level mathematics (calculus, elements of linear algebra and optimization), Supervised Learning (4C55/GI01).

Grading

The course has the following assessment components: 1) Written Examination (2.5 hours, 50%); 2) Orally assessed coursework (24 hours, 25%) Reports (24 hours, 25%).

To pass this course, students must obtain at least 40% on the coursework component and an average of at least 50% when the coursework and exam components of a course are weighted together.

In-class presentations

Each student needs to choose a topic, either based on a chapter/section from the course book or a research paper (see list below). S/he will read and then prepare a presentation. The presentation's aim is to describe and explain the topic to the class (allowing for questions). The following are the requirements:

1. Topics are allocated to students on a first-come first-get basis. To be allocated, a student needs to send a request to the lecturer stating the name of the chosen topic. The list will be updated online to indicate which topics are still available. Once a topic has been allocated, there will be no changes.

2. All requests for topic allocation must be sent to the lecturer no later than 12:00am, Monday Jan 24, 2005. The first group of presentations will be on Feb 4. The second group on Feb 11.

3. The presentation needs to be done using power-point on a laptop (will be provided) or on a transparency-projector (preferably using Latex but handwriting is o.k. provided it is clear). The amount of material per slide should be equivalent to approximately 2/4 minutes of presentation time per slide. The time allocated for each presentation is 30-50 minutes.

4. The presentation material needs to be prepared in electronic format having 4 slides per page (preferably in .pdf or .ps but PowerPoint is also possible) and sent to the lecturer at least 48 hours prior to the presentation date. The dates/times of each presentation will be announced in class.

Title Who When
Singular value decomposition, Ch. 6.1 and proof of Prop. 6.12 Hassan Feb 4
Direction of max covariance and generalized eigenvalue problem, Ch. 6.3 6.4, pp 155-164 Feng Yuan Feb 4
Canonical correlation analysis, Ch. 6.5, pp 164-175 Xi Chen Feb 11
Fisher discriminant analysis, Ch. 5.4, pp 132-137 Christopher Feb 11
Kernel perceptron algorithm, Ch. 7.4, pp 241-248 Phil Feb 4
Novelty detection, Ch 7.1.1, 7.1.3 Oliver Feb 4
nu-support vector machine, Ch 7.2 Stefania Feb 11
Support vector machine regression, Ch 7.3.3 Jia Zhou Feb 4
Regularized multi-task learning Sylvain Feb 11
Probabilistic outputs for SVMs Cheng Yuan Feb 11
SMO Method to train SVMs Available  
Learning multiple parameters with SVMs Available  
Reduced sets methods Available  
Kernel PCA paper Tao Feb 4
Geometry of SVM classifiers Available  

Syllabus

The tentative schedule of the course is listed below. Follow the link for each class to find a detailed description, suggested readings, and lecture slides.


Date Title
Tuesday, January 11 Kernels in Machine Learning

-- Ridge Regression
-- Feature maps
-- Positive definite kernels
-- kernel construction
-- kernel on Euclidean spaces

Friday, January 12 Kernel-Based Learning Algorithms

-- Simple operation in feature space
-- Kernel PCA
-- Novelty detection
-- Optimal separating hyperplane
-- Soft margin separation
-- Support vector machines
-- Extensions

Friday, January 21 Reproducing Kernel Hilbert Spaces

-- Hilbert spaces
-- Reproducing kernel Hilbert space
-- Mercer theorem
-- Regularization, representer theorem

Friday, February 4 Individual Presentations I

-- TBA

Friday, February 11 Individual Presentations II

-- TBA

Reading List