Confused about Karhunen-Loeve Transform

17 次查看(过去 30 天)
Hi all, I've read lots of documents about KLT, but still get confused on how to correctly apply it to a data set. My references:
To sum up about the first link, the step to calculate the KLT of a matrix x = [x1(t); x2(t); x3(t), ... xn(t)] in which xn(t) is a time series
Step1: auto-correlation matrix--> R = x*x';
Step2: calculate transformation matrix Phi_H--> [V,D] = eig( R); Phi_H = V';
Step3: transformed matrix--> y = Phi_H * x
Then apply this algorithm to the example at link2, to transform a matrix x = [1 2 4; 2 3 10]; The transformation matrix is not right according to that done by Mathematica, neither does the final result.
Could anyone please revise my algorithm applying to transform a time-series matrix?
Thank you for any suggestions Kyle
  2 个评论
Daniel Shub
Daniel Shub 2011-9-26
I always confuse the differences between KLT, SVD, and PCA. Do you need to subtract the mean in KLT?
Kyle
Kyle 2011-9-27
They are really similar, wiki states that KLT is "also called" SVD. In some code SVD removes the mean, but I think KLT doesn't need this step.

请先登录,再进行评论。

采纳的回答

UJJWAL
UJJWAL 2011-9-26
编辑:Walter Roberson 2015-8-24
Ok.... Let me put it down one by one...
Karhunen Loeve Transform relies on the covariance matrix of a set of observation vectors. The basic Algorithm is clearly explained in the first link you have posted. I am below giving the code for the KLT for the same example as given in the MATHEMATICA Example which you have mentioned :-
clc
clear all
y=[1,2,4;2,3,10];
y=y' %Reasons for transposing will become clear when you will read the second point given below.
[V,D]=eig(cov(y))
KLT = V' * y';
Note the following points :-
a) For a given Eigenvalue there can be multiple EigenVectors which will be multiples of each other. Now if you run the program then please observe the following points :-
(i) Take a look at the transformation matrix returned by mathematica and the matrix V(which is the same as the transformation matrix returned by MATLAB). They are essentially the same except that one of the eigenvectors has been multiplied by -1 which is perfectly acceptable because even by multiplication by -1 it is still an eigenvector. It is this difference in the transformation matrices used by the two softwares that results into different values of the KVT. Also note that one of the eigenvectors in both the softwares is the same and so is one of the rows in the KVT. So Essentially the difference is there just because of the multiplication by -1 and there is nothing to worry about these results. If mathematics allows it then nothing can refute that.
(ii) Also note that in MATHEMATICA the observations are taken row-wise while in MATLAB they are taken Column-wise.It is because of that I have transposed y in the above code.
Please remember that each software calculates a quantity by a certain algorithm and minute differences may arise. The only important thing is mathematical consistency. Hope this clears your doubt.
Happy To help.
UJJWAL
  2 个评论
SGUNITN
SGUNITN 2020-10-30
I was reading some paper where KLT is used for template protection.
Could you please share your view if it can be used for protecting a template like below.
[481.28 404.27 845.21 311.36 363.94 92.913 375.61 2647.8 2479.8 99.213 13.647 407.16 795.75 3121]
Jaime De La Mota Sanchis
As far as I understand, the KLT should be orthogonal, but as far as I can tell, they aren't. Am I wrong?

请先登录,再进行评论。

更多回答(0 个)

类别

Help CenterFile Exchange 中查找有关 Operating on Diagonal Matrices 的更多信息

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by