The two minimization problems below are equivalent:
min{trace(AXTBX):XXT=In}=min{trace(AQT˜BQ):QQT=Im}, where A,˜B and Q are square matrices of the same size, A,B are also p.s.d and X is a m×n,(m>n) rectangular matrix.
The solution to this latter minimization problem is well known: the extrema of the trace are the extrema of {∑mi=1λi(A)λσ(i)(˜B):σ∈Sm}, where λi(M) denotes the i-th eigenvalue of a real symmetric matrix M and Sm is the symmetric group of order m.
Suppose A and ˜B are orthogonally diagonalized as A=UΛUT and ˜B=VΣVT, where Λ=diag(λ1(A),λ2(A),…,λm(A)) contains the eigenvalues of A arranged in ascending order and Σ is analogously defined, but the eigenvalues are arranged in descending order. That is, if λ1(B),…,λn(B) are arranged in ascending order, and for Σ=diag(λn(B),…,λ1(B),0,…,0)
Then am looking for a 'detailed proof' that the minima is reached at an X∗ given by:
X∗=VUT[In0(m−n)×n].
for my better understanding.
Answer
It is a bit easy to show for the square case. For the rectangular case, it becomes cumbersome due to heavy notational stuff. You need to have the following inequality for the positive (semi) definite matrices
trace(AB)≥N∑i=1αiβN−i+1
where α1≥α2...≥αN are the eigenvalues of A and β1≥β2...≥βN are eigenvalues of B. The lower bound is achieved when A and B commute and the corresponding eigenvalues are in order (ie ascending for A and descending for B).
Now X is orthogonal, Define ˆB=XHBX. Note that B and ˆB will have same eigenvalues. This implies trace(AˆB)≥∑Ni=1αiβN−i+1. Thus it is enough to design X such that we can attain the (universal) lower bound for every X. This is possible only if X=VUHP. Here U and V comes from eigen decomposition of A and B. P is a suitable permutation (which is orthogonal) matrix which rearranges the eigenvalues. Since you have already assumed the required order, it should be identity.
No comments:
Post a Comment