Local SVM Classification Based on Triangulation

Download Report

Transcript Local SVM Classification Based on Triangulation

CS558 Project
Local SVM Classification based
on triangulation
(on the plane)
Glenn Fung
Outline of Talk
 Classification problem on the plane
 All of the recommended stages were applied:
 Sampling
 Ordering:
 Clustering
 Triangulation
 Interpolation (Classification)
SVM: Support vector Machines
 Optimization: Number of training points increased
 Evaluation:
Checkerboard dataset
Spiral dataset
Classification Problem in R 2
 Given m points in 2 dimensional space
 Represented by an m-by-2 matrix A
 Membership of each A i in class +1 or –1
SAMPLING:
1000 randomly sampled points
ORDERING:
Clustering
 A Fuzzy-logic based clustering algorithm was used.
 32 cluster centers were obtained
250
200
150
100
50
0
-50
-50
0
50
100
150
200
250
ORDERING:
Delaunay Triangulation
 Algorithms to triangulate and to get the Delaunay triangulation from HWKs 3
and 4 were used.
 Given a point,the random point approach is used to localize the triangle that
contains it.
200
150
100
50
0
0
50
100
150
200
Interpolation:
SVM
 SVM : Support Vector Machine Classifiers
 A different nonlinear Classifier is used for each triangle
 The triangle structure is efficiently used for both training
and testing phases and for defining a “simple” and fast
nonlinear classifier.
What is a Support Vector Machine?
 An optimally defined surface
 Typically nonlinear in the input space
 Linear in a higher dimensional space
 Implicitly defined by a kernel function
What are Support Vector Machines
Used For?
 Classification
 Regression & Data Fitting
 Supervised & Unsupervised Learning
(Will concentrate on classification)
Support Vector Machines
Maximizing the Margin between Bounding
Planes
w
x 0w = í + 1
A+
A-
x 0w = í à 1
2
jj wjj 2
The Nonlinear Classifier
 The nonlinear classifier:
K (x 0; A 0)D u = í
K (A; A 0) : R m â n â R n â m7
à ! R mâ m
 Where K is a nonlinear kernel, e.g.:
 Gaussian (Radial Basis) Kernel :
0
K ( A; A ) i j =
"
à ö kA i à A j k 22
; i; j = 1; . . .; m
 The i j -entry of K (A; A 0) represents the “similarity”
of data points A i and A j
Reduced Support Vector Machine Algorithm
öu
Nonlinear Separating Surface: K (x 0; Aö0)D
ö= í
(i) Choose a random subset matrix A 2 R mâ n of
entire data matrix A 2 R mâ n
(ii) Solve the following problem by the Newton
method with corresponding D ú D :
min
(u; í ) 2 R
÷
kp(e à
2
m+ 1
öu
D (K (A; A 0)D
ö à eí ); ë) k 22 + 12ku
ö; í k 22
(iii) The separating surface is defined by the optimal
solution ( u; í ) in step (ii):
öu
K (x 0; Aö0)D
ö= í
How to Choose A in RSVM?
 A is a representative sample of the entire dataset
 Need not be a subset of A
 A good selection of A may generate a classifier using
very small m
 Possible ways to choose A :
 Choose m random rows from the entire dataset A
 Choose A such that the distance between its rows
exceeds a certain tolerance
 Use k cluster centers of A + and A à as A
Obtained Bizarre “Checkerboard”
Optimization: More sampled points
Training parameters adjusted
Result: Improved Checkerboard
Nonlinear PSVM: Spiral Dataset
94 Red Dots & 94 White Dots
Next:Bascom Hill
Some Questions
 Would it work for B&W pictures (regression instead of
classification?
 Aplications?