Dimension Reduction and Classication Methods for Object...

38
Dimension Reduction and Classification Methods for Object Recognition in Vision Charles B OUVEYRON LMC-IMAG & INRIA Rhône-Alpes Dimension Reduction and Classification Methods for Object Recognition in Vision – p.1/23

Transcript of Dimension Reduction and Classication Methods for Object...

Page 1: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

Dimension Reduction andClassification Methods

for Object Recognition in Vision

Charles BOUVEYRON

LMC-IMAG & INRIA Rhône-Alpes

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.1/23

Page 2: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

Framework

Goal : approach the capability of training and visualrecognition of the human eye.

locate an object on natural images with complexbackground,recognize the object class (require a learning stage).

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.2/23

Page 3: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

Framework

Goal : approach the capability of training and visualrecognition of the human eye.

locate an object on natural images with complexbackground,recognize the object class (require a learning stage).

Tools :

local descriptors of the image,statistical methods:

dimension reduction,Bayesian classification.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.2/23

Page 4: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

Some applications

Object recognition and localisation inimages or videos,

Automatic localisation of tumors in MRIimages,

Identification of peoples for accesscontrol,

Interpretation of aerial images(applications in biology),

Content-based image retrieval (likeGoogle Image Search).

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.3/23

Page 5: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

Outline of the talk

1 – Object recognition:

Learning the modelRecognition

2 – Why to reduce the dimension ?

3 – Previous works

4 – Our approach

5 – Experimental results

6 – Conclusion and further work

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.4/23

Page 6: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

1 – Object recognition

Learning

Projection in thelearning space

(a) Interest points detection

(b) Object features selection

(c) Computelocal descriptors

Object features recognized and object located Statistical model

Dimensionreduction

Test image

(a) + (c)

Bayesianclassification

Decision

Learning:

detection and description,

dimension reduction,

learning the densities.

Recognition:

detection and description,

projection in learning space,

classification using Bayesian

rules.Dimension Reduction and Classification Methods for Object Recognition in Vision – p.5/23

Page 7: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

1 – Learning: detection

The Harris-Laplace operator allows to detectcharacteristic points of the image structure with theircharacteristic scale.

Fig. 1 – Interest points detection using the Harris-Laplace operator.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.6/23

Page 8: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

1 – Learning: detection

The Harris-Laplace operator allows to detectcharacteristic points of the image structure with theircharacteristic scale.

Harris filter:

based on the autocorrelation matrix A computed ona neighborhood of the keypoint x:

A =

(

IxIx

IxIy∑

IyIx

IyIy

)

.

eigenvalues of A:2 strong eigenvalues → x is an interest point,1 strong eigenvalue → x is a contour,0 strong eigenvalue → x is in an homogeneous region.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.6/23

Page 9: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

1 – Learning: detection

The Harris-Laplace operator allows to detectcharacteristic points of the image structure with theircharacteristic scale.

Laplace operator: detection of the caracteristic scale[MS03a]

computation of the Laplacian for several scalesaround the keypoints,caracteristic scale: s∗ = argmaxs{∆x,y(s)}

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.6/23

Page 10: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

1 – Learning: description

SIFT descriptor [Low04] : a robust descriptor [MS03b]

based on the gradient values in 8 directions of 4 × 4patches around the interest point→ descriptor 128-dimensional.invariant with luminosity changes (and geometrictransformations).

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.7/23

Page 11: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

1 – Learning: statistical model

We consider a statistical model made of 6 classes:

1 background class modeled by a multidimensionalGaussian density f0,5 motorbike parts. Each of them is also modeled bya Gaussian density fi, i = 1, ..., 5.

Fig. 2 – Statistical model for 4 object classes and 1 background class.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.8/23

Page 12: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

1 – Recognition: classification

We used the three following classification methods:

(R3) Bayesian classification (require to estimate thefull covariance matrix Σi for each class i),(R2) Naive Bayesian classification(∀i = 1, ..., 5, Σi = Σ),(R1) Independence assumption (∀i = 1, ..., 5, Σi

diagonal matrix).

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.9/23

Page 13: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

1 – Recognition: classification

We used the three following classification methods:

(R3) Bayesian classification (require to estimate thefull covariance matrix Σi for each class i),(R2) Naive Bayesian classification(∀i = 1, ..., 5, Σi = Σ),(R1) Independence assumption (∀i = 1, ..., 5, Σi

diagonal matrix).

The Bayesian classification rule consists in affecting thepoint x to the class Ci with the maximal a posterioriprobability (MAP) P (Ci|x):

x ∈ Ci, if i = argmaxj=0,...,5{pjfj(x)}

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.9/23

Page 14: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

2 – Why to reduce the dimension ?

Dimension reduction:

find a new feature space,with a dimension significantly smaller,which contains a large part of the originalinformation.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.10/23

Page 15: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

2 – Why to reduce the dimension ?

Dimension reduction:

find a new feature space,with a dimension significantly smaller,which contains a large part of the originalinformation.

Why ?

find a more discriminant representation of the data.overcome the curse of the dimensionality(particularly for classification methods).this also allows to denoise the data.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.10/23

Page 16: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

3 – Previous works

Eigen-Images [TP91] :

Descriptor based on grayscale histogram.Dimension reduction using PCA for face recognition.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.11/23

Page 17: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

3 – Previous works

Eigen-Images [TP91] :

Descriptor based on grayscale histogram.Dimension reduction using PCA for face recognition.

PCA-SIFT [KS03] : show that dimension reductionincreases recognition results.

SIFT in dimension 3042 then apply PCA to obtain amodified descriptor in dimension 20 (empiricallycomputed)Results (in image retrieval application) :

using SIFT in dimension 3042 : 40% corrects,using PCA-SIFT in dimension 20 : 90% corrects.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.11/23

Page 18: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

4 – Our approach: goals

We will compare the following dimension reductionmethods :

Principal Component Analysis (PCA),Discriminant Analysis (LDA),Nonlinear method: Locally Linear Embedding (LLE).

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.12/23

Page 19: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

4 – Our approach: goals

We will compare the following dimension reductionmethods :

Principal Component Analysis (PCA),Discriminant Analysis (LDA),Nonlinear method: Locally Linear Embedding (LLE).

Main goals:

find a more discriminant data representation in orderto increase recognition results,accelerate the computing times (in order to be usedin real-time applications),

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.12/23

Page 20: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

4 – Our approach: PCA, LDA and LLE

PCA: Principal Component Analysis

find a new data representation in a subspace ofsmaller size and maximizing the variance.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.13/23

Page 21: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

4 – Our approach: PCA, LDA and LLE

PCA: Principal Component Analysis

find a new data representation in a subspace ofsmaller size and maximizing the variance.

LDA: Linear Discriminant Analysis

find the (k − 1) discriminant axes maximizing theinterclass variance and minimizing the intraclassvariance.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.13/23

Page 22: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

4 – Our approach: PCA, LDA and LLE

PCA: Principal Component Analysis

find a new data representation in a subspace ofsmaller size and maximizing the variance.

LDA: Linear Discriminant Analysis

find the (k − 1) discriminant axes maximizing theinterclass variance and minimizing the intraclassvariance.

LLE: Locally Linear Embedding [SR01]

find an embedding which preserves the localgeometry in a linear neighborhood of each point inthe original space.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.13/23

Page 23: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

4 – Our approach: data

We chose to work with a set of 200 motorbike images,

We are interested by 5 features of the object.

(a) Examples of image database:

(b) The 5 object features:

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.14/23

Page 24: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

5 – Results: projection using LDA

Fig. 3 – Supervised dimension reduction using LDA without background.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.15/23

Page 25: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

5 – Results: protocol

For each dimension d = 1, ..., 128:

We reduced the dimension of the learning set usingPCA, LDA and LLE,then, we learned the densities of each classfollowing the model.Next, we projected test descriptors in the learningspace,and computed the a posteriori probability for eachclass according to (R1),(R2) and (R3).Finally, affected each test descriptor to the class withthe maximum a posteriori probability.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.16/23

Page 26: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

5 – Results: PCA vs LDA

0 20 40 60 80 100 120 1400.45

0.5

0.55

0.6

0.65

0.7

Dimension

Cla

ssifi

catio

n ra

te

R1R2R3

0 20 40 60 80 100 120 1400.45

0.5

0.55

0.6

0.65

0.7

Dimension

Cla

ssifi

catio

n ra

te

R1R2R3

(a) (b)

Fig. 4 – Classification results with (a) PCA, (b) LDA using the threeclassifiers.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.17/23

Page 27: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

5 – Results: discussion

Dimension reduction by PCA:

increases recognition results, particularly with therule (R3),the better results are obtained in 45 dimensions.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.18/23

Page 28: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

5 – Results: discussion

Dimension reduction by PCA:

increases recognition results, particularly with therule (R3),the better results are obtained in 45 dimensions.

Dimension reduction by LDA:

also increases recognition results, particularly withthe rule (R2) in low dimension,very good results are obtained in fixed dimension:(k − 1) dimensions (here k = 6),computing times are smaller.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.18/23

Page 29: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

5 – Results: discussion

Dimension reduction by PCA:

increases recognition results, particularly with therule (R3),the better results are obtained in 45 dimensions.

Dimension reduction by LDA:

also increases recognition results, particularly withthe rule (R2) in low dimension,very good results are obtained in fixed dimension:(k − 1) dimensions (here k = 6),computing times are smaller.

Dimension reduction by LLE:

recognition results are worse than other methods.Dimension Reduction and Classification Methods for Object Recognition in Vision – p.18/23

Page 30: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

6 – Conclusion

Our work highlights that:

dimension reduction increases recognition results,which allows to use efficiently complex classificationmethodsparticularly, LDA gives a very discriminantrepresentation in low and fixed dimension,this dimension is known in advance: d = (k − 1).

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.19/23

Page 31: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

6 – Further work

Statistical model:

use a model based on Gaussian mixture (object andbackground),find a method for automatically detecting objectfeatures.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.20/23

Page 32: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

6 – Further work

Statistical model:

use a model based on Gaussian mixture (object andbackground),find a method for automatically detecting objectfeatures.

Dimension reduction:

use kernel based methods (KPCA and KLDA),use dimension reduction for each class.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.20/23

Page 33: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

6 – Further work

Statistical model:

use a model based on Gaussian mixture (object andbackground),find a method for automatically detecting objectfeatures.

Dimension reduction:

use kernel based methods (KPCA and KLDA),use dimension reduction for each class.

Include other informations:

model spatial information or neighborhoodrelationship using Bayesian network or hiddenMarkov models.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.20/23

Page 34: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

6 – Dimension reduction for each class

The idea:

The idea is to reduce the dimension for each classindependently,and, for a test descriptor x, compute the a posterioriprobability P (Ci|x) in each space,finally, affect x to the class with the MAP.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.21/23

Page 35: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

6 – Dimension reduction for each class

The idea:

The idea is to reduce the dimension for each classindependently,and, for a test descriptor x, compute the a posterioriprobability P (Ci|x) in each space,finally, affect x to the class with the MAP.

Using PCA or LLE:

reduce the dimension in k different spaces,

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.21/23

Page 36: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

6 – Dimension reduction for each class

The idea:

The idea is to reduce the dimension for each classindependently,and, for a test descriptor x, compute the a posterioriprobability P (Ci|x) in each space,finally, affect x to the class with the MAP.

Using PCA or LLE:

reduce the dimension in k different spaces,

Using LDA:

reduce the dimension in (k − 1) different spaces,in each of k − 1 spaces, compare the MAP of oneobject feature class to the background class.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.21/23

Page 37: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

6 – Dimension reduction for each class

Reduce the dimension in (5 − 1) different spaces,

Projection of test data of the class n◦ 1 in differentspaces,

Compute the MAP for each test descriptor and affect tothe best class.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.22/23

Page 38: Dimension Reduction and Classication Methods for Object ...lear.inrialpes.fr/~bouveyron/work/presentation_SSIAB.pdf · dimension reduction increases recognition results, which allows

References

References[KS03] Y. Ke and R. Sukthankar. Pca-sift: A more distinctive representation for local

image descriptors. Technical report, INTEL Research, 2003.

[Low04] D. Lowe. Distinctive image features from scale-invariant keypoints. accepted forpublication in the International Journal of Computer Vision, 2004.

[MS03a] K. Mikolajczk and C. Schmid. Indexing based on scale invariant interest points. InInternational Conference on Computer Vision, pages 525–531, 2003.

[MS03b] K. Mikolajczk and C. Schmid. A performance evaluation of local descriptors. InIEEE Conference on Computer Vision and Pattern Recognition, 2003.

[SR01] L. K. Saul and S. T. Roweis. An introduction to locally linear embedding. Technicalreport, Gatsby Computational Neuroscience Unit, UCL, 2001.

[TP91] M. Turk and A. Pentland. Eigenfaces for recognition. Cognitive Neuroscience,3(1), 1991.

Dimension Reduction and Classification Methods for Object Recognition in Vision – p.23/23