Thursday, 21 November 2013

linear algebra - Change of basis matrix - part of a proof



I'm trying to understand a proof from Comprehensive Introduction to Linear Algebra (page 244)



I can't really figure out what steps have been taken to get from eq. 1. to eq. 2. It's just overcomplicated there. Different proofs of that theorem are very easy and obvious for me, multiplication of matrix by its inverse produces an identity matrix etc. But this particular case here is causing problems for me. Now I'm not trying to understand why this theorem is true, because I already do. The thing I want is to understand step from eq. 1. to eq. 2. Thank you for help.




Here's what I'm talking about


Answer



The collection of equations



ai=nj=1pijbji=1,,n



is equivalent to the matrix equation a=Pb where aT=(a1,,an), bT=(b1,,bn) and P is the n×n matrix with (i,j)th element pij.



If P is invertible, then we can rewrite this as b=P1a. Denoting the (i,j)th element of P1 by p1ij (not great notation in my opinion), the matrix equation b=P1a is equivalent to the collection of equations




bi=nj=1p1ijaji=1,,n.



Swapping the roles of the indicies i and j, you get the following collection of equations:



bj=ni=1p1jiaij=1,,n.


No comments:

Post a Comment

real analysis - How to find limhrightarrow0fracsin(ha)h

How to find limh0sin(ha)h without lhopital rule? I know when I use lhopital I easy get $$ \lim_{h\rightarrow 0}...