Wednesday, April 6, 2016

linear algebra - Change of basis matrix - part of a proof




I'm trying to understand a proof from Comprehensive Introduction to Linear Algebra (page 244)



I can't really figure out what steps have been taken to get from eq. 1. to eq. 2. It's just overcomplicated there. Different proofs of that theorem are very easy and obvious for me, multiplication of matrix by its inverse produces an identity matrix etc. But this particular case here is causing problems for me. Now I'm not trying to understand why this theorem is true, because I already do. The thing I want is to understand step from eq. 1. to eq. 2. Thank you for help.



Here's what I'm talking about


Answer



The collection of equations



ai=nj=1pijbji=1,,n




is equivalent to the matrix equation a=Pb where aT=(a1,,an), bT=(b1,,bn) and P is the n×n matrix with (i,j)th element pij.



If P is invertible, then we can rewrite this as b=P1a. Denoting the (i,j)th element of P1 by p1ij (not great notation in my opinion), the matrix equation b=P1a is equivalent to the collection of equations



bi=nj=1p1ijaji=1,,n.



Swapping the roles of the indicies i and j, you get the following collection of equations:



bj=ni=1p1jiaij=1,,n.



No comments:

Post a Comment

analysis - Injection, making bijection

I have injection f:AB and I want to get bijection. Can I just resting codomain to f(A)? I know that every function i...