Presentation - SVM & KM - May 2009

24
Support Vector Machines and Kernel methods by Lucian Huluta 06/15/2009 6/20/2009

Transcript of Presentation - SVM & KM - May 2009

Page 1: Presentation - SVM & KM - May 2009

Support Vector Machinesand

Kernel methods

byLucian Huluta

06/15/2009

6/2

0/2

009

Page 2: Presentation - SVM & KM - May 2009

Support Vector Machine (SVM)

What is Support Vector Machine?

A statistical tool, essentially used for NONLINEAR

classification/regression.

A SUPERVISED LEARNING mechanism like

neural networks.

An quick and adaptive method for PATTERN

ANALYSIS.

A fast and flexible approach for learning COMPLEX

SYSTEMS.

2

Page 3: Presentation - SVM & KM - May 2009

Support Vector Machine (SVM)

Strengths

Few parameters required for tuning the learning

machine

Learning involves optimisation of a convex

function

It scales relatively well to high dimensional

data

Weaknesses

Training large data still difficult

Need to choose a “good” kernel function

3

Page 4: Presentation - SVM & KM - May 2009

- weights ( - dimensional vector),

- bias

Binary classification problem

- input space

SVM: linear classification

More than one solution for the decision function!

4

Page 5: Presentation - SVM & KM - May 2009

Generalization region:

SVM: Generalization capacity

Generalization ability

5

Page 6: Presentation - SVM & KM - May 2009

SVM: Hard margin

Training data must satisfy:

Quadratic optimization problem

subject to:

minimize

with constraint:

6

Page 7: Presentation - SVM & KM - May 2009

SVM: Primal form

Convert the constrained problem => unconstrained problem:

We obtain:

Solving the for and

where is nonnegative Lagrange multipliers.

7

Page 8: Presentation - SVM & KM - May 2009

The dual form of the cost function consists of inner products.

SVM: Dual form

Solve QP with following problem:

The SVM is called

hard-margin support vector machine

8

Page 9: Presentation - SVM & KM - May 2009

The modified QP minimizes following cost function:

subject to the constraints:

SVM: L1-soft margin problem

: trade-off between the maximization of the margin and minimization of the classification error.

9

Page 10: Presentation - SVM & KM - May 2009

SVM: L2-soft margin problem

The modified QP minimizes following cost function:

subject to the constraints:

10

Page 11: Presentation - SVM & KM - May 2009

Decision function:

We assume that all the training

data are within the tube with radius ε

named insensitive loss function

Slack variables:

SVM: Regression

11

Page 12: Presentation - SVM & KM - May 2009

Cost function with slack variables:

SVM: Regression

If p=1: L1 soft-margin

If p=2: L2 soft-margin

subject to the constraints:

12

Page 13: Presentation - SVM & KM - May 2009

SVM: Linear inseparability

1. data are NOT linear separable.

2. feature space is HIGH DIMENSIONAL, hence QP

takes long time to solve

3. nonlinear function approximation problems can

NOT be solved13

Page 14: Presentation - SVM & KM - May 2009

If the feature space is Hilbert space, i.e., where inner product applies…

SVM: Linear inseparability

…,we can simplify the optimization problem by a TRICK!!!

14

Page 15: Presentation - SVM & KM - May 2009

Kernel Trick = is a method for using a linear classifier algorithm to solve a non-linear problem by

The Kernel “trick”

Kernel trick avoids computing inner product of two vectors in feature space.

choosing appropriate KERNEL FUNCTIONS

15

Page 16: Presentation - SVM & KM - May 2009

Consider a two-dimensional input space together with the feature map:

Numerical Example

Kernel function

16

Page 17: Presentation - SVM & KM - May 2009

Choose kernel function:

Maximize:

Compute bias term:

Classify data using decision function:

SVM with Kernel: Steps

17

Page 18: Presentation - SVM & KM - May 2009

Linear:

Polynomial:

Radial Basis Function:

Others: design kernels suitable for target applications

Kernels

18

Page 19: Presentation - SVM & KM - May 2009

Demo

19 To see video demo please visit this link.

Page 20: Presentation - SVM & KM - May 2009

Breast cancer diagnosis and prognosis

Handwritten digit recognition

On-line Handwriting Recognition

Text Categorization

3-D Object Recognition Problems

Function Approximation and Regression

Detection of Remote Protein Homologies

Gene Expression

Vast number of applications…

• andFault diagnosis in chemical processes

Applications of SVM

20

Page 21: Presentation - SVM & KM - May 2009

Application aspects of SVM – Belousov et.al., 2002, Journal of Chemometrics

Current developments: SVM

About Kernel latent variables approaches and SVM– Czekaj et.al., 2005, Journal of Chemometrics

Kernel based orthogonal projections to latent structures– Rantalainen et.al., 2007, Journal of Chemometrics

Performance assessment of a novel fault diagnosis system based on SVM – Yelamos et.al., 2009, Computers and Chemical Engineering

SVM and its application in chemistry – Li et.al., 2009, Chemometrics and intelligent Laboratory

21

Page 22: Presentation - SVM & KM - May 2009

Identification of MIMO Hammerstein systems with LS-SVM, Goethals et.al., 2005, Automatica

Current developments: SVM

An online support vector machine for abnormal event detection, Davy et.al., 2006, Signal Processing

Support vector machine for quality, monitoring in a plastic injection molding process, Ribeiro, 2005, IEEE System Man and Cybernetics

Fault prediction for nonlinear system based on Hammerstein model and LS-SVM, Jiang et.al., 2009,IFAC Safeprocess

22

Page 23: Presentation - SVM & KM - May 2009

My future work

Finish my diploma project

Study the role of various “Tuning” parameters on classification results

Apply SVM to Tennessee Eastman benchmark that involves 20 pre-defined faults.

Apply of SVM based classification algorithm to small academic example

Study support vector machine based classification for “one-against-one” and “one-against-all” problems

23

Page 24: Presentation - SVM & KM - May 2009

Thank you..

24

&Answers