Canonical Correlation Analysis: An overview with application to learning methods By David R....
-
date post
21-Dec-2015 -
Category
Documents
-
view
213 -
download
0
Transcript of Canonical Correlation Analysis: An overview with application to learning methods By David R....
Canonical Correlation Canonical Correlation Analysis: An overview with Analysis: An overview with
application to learning application to learning methodsmethods
By David R. Hardoon, Sandor Szedmak, John Shawe-TaylorBy David R. Hardoon, Sandor Szedmak, John Shawe-TaylorSchool of Electronics and Computer Science, University of School of Electronics and Computer Science, University of
SouthamptonSouthamptonPublished in Neural Computaion, 2004Published in Neural Computaion, 2004
Presented by:Presented by:Shankar BhargavShankar Bhargav
Measuring the linear relationship between Measuring the linear relationship between two multi dimensional variablestwo multi dimensional variables
Finding two sets of basis vectors such that Finding two sets of basis vectors such that the correlation between the projections of the correlation between the projections of the variables onto these basis vectors is the variables onto these basis vectors is maximizedmaximized
Determine Correlation CoefficientsDetermine Correlation Coefficients
Canonical Correlation AnalysisCanonical Correlation Analysis
More than one canonical correlations will More than one canonical correlations will be found each corresponding to a different be found each corresponding to a different set of basis vectors/Canonical variatesset of basis vectors/Canonical variatesCorrelations between successively Correlations between successively extracted canonical variates are smaller extracted canonical variates are smaller and smallerand smallerCorrelation coefficients : Proportion of Correlation coefficients : Proportion of correlation between the canonical variates correlation between the canonical variates accounted for by the particular variable. accounted for by the particular variable.
Canonical Correlation AnalysisCanonical Correlation Analysis
Differences with CorrelationDifferences with Correlation
Not dependent on the coordinate system Not dependent on the coordinate system of variablesof variables
Finds direction that yield maximum Finds direction that yield maximum correlationscorrelations
Find basis vectors for two sets of variables x, y Find basis vectors for two sets of variables x, y such that the correlations between the such that the correlations between the projections of the variables onto these basis projections of the variables onto these basis vectorvector
Sx = (x.wx) and Sy = (y.wy)
ρ = E[Sx Sy ]
√ E[Sx2] E[Sy
2]
ρ = E[(xT wx yT wy)]
√E[(xT wx xT wx) ] E[(yT wy yT wy)]
ρ = max wx wy E[wxTx
yT wy]
√E[wxTx xT wx ] E[wy
T y yT wy]
ρ = max wx wy wxTCxy
wy
√ wxTCxxwx wy
TCyy wy
Solving this
with constraint with constraint wxTCxxwx =1
wyTCyy
wy=1
Cxx-1CxyCyy
-1Cyx wx = ρ2 wx
Cyy-1CyxCxx
-1Cxy wy= ρ2 wy
Cxy wy = ρλx Cxx wx
Cyx wx = ρλy Cyy wy
λx=λy-1= wy
TCyywy
√√ wxTCxxwx
CCA in MatlabCCA in Matlab
[ A, B, r, U, V ] = canoncorr(x, y)[ A, B, r, U, V ] = canoncorr(x, y)
x, y : set of variables in the form of matricesx, y : set of variables in the form of matrices Each row is an observationEach row is an observation Each column is an attribute/featureEach column is an attribute/feature
A, B: Matrices containing the correlation coefficientA, B: Matrices containing the correlation coefficientr : Column matrix containing the canonical r : Column matrix containing the canonical
correlations (Successively decreasing)correlations (Successively decreasing)U, V: Canonical variates/basis vectors for A,B U, V: Canonical variates/basis vectors for A,B
respectivelyrespectively
Interpretation of CCAInterpretation of CCA
Correlation coefficient represents unique Correlation coefficient represents unique contribution of each variable to relationcontribution of each variable to relationMulticollinearity may obscure relationshipsMulticollinearity may obscure relationshipsFactor Loading : Correlations between the Factor Loading : Correlations between the canonical variates (basis vector) and the canonical variates (basis vector) and the variables in each setvariables in each setProportion of variance explained by the Proportion of variance explained by the canonical variates can be inferred by canonical variates can be inferred by factor loadingfactor loading
Redundancy CalculationRedundancy Calculation
Redundancy Redundancy leftleft =[ ∑ (loadings =[ ∑ (loadingsleftleft22)/p]*R)/p]*Rcc
22
Redundancy Redundancy rightright =[ ∑ (loadings =[ ∑ (loadingsrightright22)/q]*R)/q]*Rcc
22
p – Number of variable in the first (left) set of variablesp – Number of variable in the first (left) set of variables q – Number of variable in the second (right) set of q – Number of variable in the second (right) set of
variablesvariablesRc2 – Respective squared canonical correlationRc2 – Respective squared canonical correlation
Since successively extracted roots are uncorrelated we Since successively extracted roots are uncorrelated we can sum the redundancies across all correlations to can sum the redundancies across all correlations to get a single index of redundancy.get a single index of redundancy.
ApplicationApplication
Kernel CCA can be used to find non linear Kernel CCA can be used to find non linear relationships between multi variatesrelationships between multi variates
Two views of the same semantic object to Two views of the same semantic object to extract the representation of the semanticsextract the representation of the semantics Speaker Recognition – Audio and Lip Speaker Recognition – Audio and Lip
movementmovement Image retrieval – Image features (HSV, Image retrieval – Image features (HSV,
Texture) and Associated textTexture) and Associated text
Use of KCCA in cross-modal Use of KCCA in cross-modal retrievalretrieval
400 records of JPEG images for each class 400 records of JPEG images for each class with associated text and a total of 3 classeswith associated text and a total of 3 classes
Data was split randomly into 2 parts for Data was split randomly into 2 parts for training and testtraining and test
Features Features Image – HSV Color, Gabor textureImage – HSV Color, Gabor texture
Text – Term frequenciesText – Term frequencies Results were taken for an average of 10 runsResults were taken for an average of 10 runs
Cross-modal retrievalCross-modal retrieval
Content based retrieval: Retrieve images Content based retrieval: Retrieve images in the same classin the same classTested with 10 and 30 images setsTested with 10 and 30 images sets
where countwhere countjjkk = 1 if the image k in the set is of = 1 if the image k in the set is of
the same label as the text query present in the same label as the text query present in the set, else countthe set, else count jj
kk = 0. = 0.
Comparison of KCCA (with 5 and 30 Eigen Comparison of KCCA (with 5 and 30 Eigen vectors) with GVSMvectors) with GVSM
Content based retrievalContent based retrieval
Mate based retrievalMate based retrieval
Match the exact image among the Match the exact image among the selected retrieved imagesselected retrieved images
Tested with 10 and 30 images setsTested with 10 and 30 images sets
where countwhere countjj = 1 if the exact matching image = 1 if the exact matching image
was present in the set else it is 0was present in the set else it is 0
Comparison of KCCA (with 30 and 150 Eigen Comparison of KCCA (with 30 and 150 Eigen vectors) with GVSMvectors) with GVSMMate based retrievalMate based retrieval
CommentsComments
The goodThe good Good explanation of CCA and KCCAGood explanation of CCA and KCCA Innovative use of KCCA in image retrieval applicationInnovative use of KCCA in image retrieval application
The badThe bad The data set and the number of classes used The data set and the number of classes used
were smallwere small The image set size is not taken into account The image set size is not taken into account
while calculating accuracy in Mate based while calculating accuracy in Mate based retrievalretrieval
Could have done cross-validation testsCould have done cross-validation tests
Limitations and Assumptions of Limitations and Assumptions of CCACCA
At least 40 to 60 times as many cases as At least 40 to 60 times as many cases as variables is recommended to get relliable variables is recommended to get relliable estimates for two roots– estimates for two roots– BarciKowski &BarciKowski & Stevens(1986)Stevens(1986)
Outliers can greatly affect the canonical Outliers can greatly affect the canonical correlationcorrelation
Variables in two sets should not be Variables in two sets should not be completely redundantcompletely redundant