Thursday, March 24, 2016

linear algebra - Assuming $AB=I$ prove $BA=I$





Most introductory linear algebra texts define the inverse of a square matrix $A$ as such:


Inverse of $A$, if it exists, is a matrix $B$ such that $AB=BA=I$.


That definition, in my opinion, is problematic. A few books (in my sample less than 20%) give a different definition:


Inverse of $A$, if it exists, is a matrix $B$ such that $AB=I$. Then they go and prove that $BA=I$.


Do you know of a proof other than defining inverse through determinants or through using rref?



Is there a general setting in algebra under which $ab=e$ leads to $ba=e$ where $e$ is the identity?


Answer



Multiply both sides of $AB-I=0$ on the left by $B$ to get $$ (BA-I)B=0\tag{1} $$ Let $\{e_j\}$ be the standard basis for $\mathbb{R}^n$. Note that $\{Be_j\}$ are linearly independent: suppose that $$ \sum_{j=1}^n a_jBe_j=0\tag{2} $$ then, multiplying $(2)$ on the left by $A$ gives $$ \sum_{j=1}^n a_je_j=0\tag{3} $$ which implies that $a_j=0$ since $\{e_j\}$ is a basis. Thus, $\{Be_j\}$ is also a basis for $\mathbb{R}^n$.


Multiplying $(1)$ on the right by $e_j$ yields $$ (BA-I)Be_j=0\tag{4} $$ for each basis vector $Be_j$. Therefore, $BA=I$.


Failure in an Infinite Dimension


Let $A$ and $B$ be operators on infinite sequences. $B$ shifts the sequence right by one, filling in the first element with $0$. $A$ shifts the sequence left, dropping the first element.


$AB=I$, but $BA$ sets the first element to $0$.


Arguments that assume $A^{-1}$ or $B^{-1}$ exist and make no reference to the finite dimensionality of the vector space, usually fail to this counterexample.


No comments:

Post a Comment

analysis - Injection, making bijection

I have injection $f \colon A \rightarrow B$ and I want to get bijection. Can I just resting codomain to $f(A)$? I know that every function i...