I have two matrices
$$
A=\begin{pmatrix}
a & 0 & 0 \\
0 & b & 0 \\
0 & 0 & c
\end{pmatrix}
\quad
\text{ and }
\quad
B=\begin{pmatrix}
d & e & f \\
d & e & f \\
d & e & f
\end{pmatrix}
$$
In reality mine are more like 1000 x 1000 matrices but the only thing that is important for now is that the left matrix is diagonal and the right one has one row that repeats itself.
Obviously the eigenvalues of the left matrix are its diagonal components. I want to create a new matrix C
$$C = A+B=\begin{pmatrix}
a & 0 & 0 \\0 & b & 0 \\0 & 0 & c \end{pmatrix}+\begin{pmatrix} d & e & f \\d & e & f \\d & e & f \end{pmatrix}=\begin{pmatrix} a+d & e & f \\d & b+e & f \\d & e & c+f \end{pmatrix}$$
I am now wondering how the eigenvalues of this new matrix C are related to the eigenvalues of the diagonal matrix A. Can I use an argument that uses row reduction in order to relate the eigenvalues of both matrices?
The reason why I am asking is that my 1000 x 1000 matrix (implemented in mathematica) that is described as above gives me almost the same eigenvalues as the corresponding diagonal matrix (only a few eigenvalues differ) and I really cannot think of any reason why that should be the case.
EDIT:
I implemented a simple code in mathematica to illustrate what I mean. One can see that every eigenvalue of the diagonal matrix A appears in C:
dim = 50;
A = DiagonalMatrix[Flatten[RandomInteger[{0, 10}, {1, dim}]]];
mat = RandomReal[{0, 100}, {1, dim}];
B = ArrayFlatten[ConstantArray[{mat}, dim]];
c = A + B;
Abs[Eigenvalues[A]]
Round[Abs[Eigenvalues[c]], 0.01]
(*{10, 10, 10, 10, 10, 10, 9, 9, 9, 9, 9, 9, 8, 8, 8, 8, 7, 7, 7, 7, 7,
6, 6, 6, 6, 5, 5, 5, 5, 5, 4, 4, 4, 4, 3, 3, 3, 3, 3, 2, 2, 2, 2, 2,
1, 1, 1, 0, 0, 0}*)
(*{2084.89, 10., 10., 10., 10., 10., 9.71, 9., 9., 9., 9., 9., 8.54,
8., 8., 8., 7.72, 7., 7., 7., 7., 6.61, 6., 6., 6., 5.44, 5., 5., 5.,
5., 4.29, 4., 4., 4., 3.51, 3., 3., 3., 3., 2.28, 2., 2., 2., 2.,
1.21, 1., 1., 0.33, 0., 0.}*)
Answer
The phenomenon you're observing happens because your example is not a generic one, but have many repeated eigenvalues.
First, if all the eigenvalues of $A$ are distinct then the rank-one perturbation $A+bk^T$ may have an arbitrary set of eigenvalues iff the pair $(A,b)$ is controllable, or equivalently the matrix $R=[b\ Ab\ \ldots\ A^{n-1}b]$ is of full rank. The result is known as the pole placement in control theory. In our case, $b=[1\ 1\ \ldots\ 1]^T$ and $R$ becomes the Vandermonde matrix, which is clearly invertible under our assumption on the eigenvalues of $A$. Conclusion: in general, you cannot say anything about the perturbed eigenvalues if you just know the eigenvalues of $A$ and not the perturbation.
What happens if the eigenvalues are repeated as in your example? Define
$A=\operatorname{diag}\{a_i\}$. Introduce the polynomials
$$
p(\lambda)=\det(\lambda I-A)=\prod_{i=1}^n(\lambda-a_i),\qquad p_i(\lambda)=\frac{p(\lambda)}{\lambda-a_i}=\prod_{j\ne i}(\lambda-a_j),
$$
and calculate the characteristic polynomial for $A+bk^T$ using Sylvester's determinant theorem
\begin{align}
\det(\lambda I-A-bk^T)=p(\lambda)(1-k^T(\lambda I-A)^{-1}b)=p(\lambda)-\sum_{i=1}^n k_ib_ip_i(\lambda).
\end{align}
Note that all the polynomials are going to have the common factor $\lambda-a$ if $a$ is an eigenvalue of multiplicity more than $1$, thus, this $a$ is a perturbed eigenvalue as well. It has got the multiplicity one less than that of the corresponding eigenvalue in $A$. This is what you see in your numerical example. Hence the rule is
If you have an eigenvalue $a$ for $A$ of multiplicity $k>1$ then you gonna have the same perturbed eigenvalue $a$ for $A+bk^T$ of multiplicity at least $k-1$.
EDIT: A simple example, take $A=I$, the $n\times n$ identity matrix. Then
$$
\det(\lambda I-I-bk^T)=\det((\lambda-1)I-bk^T)=[\mu=\lambda-1]=\det(\mu I-bk^T)=0.
$$
The eigenvalues of the rank-one matrix $bk^T$ are $n-1$ zeros and one more whatever. Those $n-1$ zeros for $\mu$ are $n-1$ ones for $\lambda$.
No comments:
Post a Comment