A Linear Transformation Preserves Exactly Two Lines If and Only If There are Two Real Non-Zero Eigenvalues

Problems and Solutions of Eigenvalue, Eigenvector in Linear Algebra

Problem 472

Let $T:\R^2 \to \R^2$ be a linear transformation and let $A$ be the matrix representation of $T$ with respect to the standard basis of $\R^2$.

Prove that the following two statements are equivalent.

(a) There are exactly two distinct lines $L_1, L_2$ in $\R^2$ passing through the origin that are mapped onto themselves:
\[T(L_1)=L_1 \text{ and } T(L_2)=L_2.\]

(b) The matrix $A$ has two distinct nonzero real eigenvalues.

 
LoadingAdd to solve later

Sponsored Links


Proof.

(a) $\implies$ (b).

Suppose that the statement (a) holds.
That is, there are lines $L_1, L_2$ in $\R^2$ passing through the origin that are mapped onto themselves, and no other lines passing through the origin are mapped onto themselves.

A line passing through the origin in $\R^2$ is a one-dimensional subspace in $\R^2$. Thus, it is spanned by a single nonzero vector.

We have
\[L_1=\Span(\mathbf{v}_1) \text{ and } L_2=\Span(\mathbf{v}_2)\] for some nonzero vectors $\mathbf{v}_1, \mathbf{v}_2$.


Since $T(L_1)=L_1$, we have $A\mathbf{v}_1\in L_1=\Span(\mathbf{v}_1)$.
Hence there exists $\lambda_1 \in \R$ such that
\[A\mathbf{v}_1=\lambda_1\mathbf{v}_1.\] Since $\mathbf{v}_1$ is a nonzero vector, this means that $\lambda_1$ is an eigenvalue of $A$.


We claim that $\lambda_1\neq 0$. Otherwise, we have $A\mathbf{v}_1=\mathbf{0}$.
Any vector in $L_1$ is of the form $\mathbf{v}=a\mathbf{v}_1$ for some $a\in \R$. So we have
\[A\mathbf{v}=A(a\mathbf{v}_1)=aA\mathbf{v}_1=\mathbf{0},\] and this yields that $T(L_1)=\{\mathbf{0}\}$, a contradiction.
Hence $\lambda_1$ is a nonzero real eigenvalue of $A$.

By the same argument, there is a nonzero real eigenvalue $\lambda_2$ such that $A\mathbf{v}_2=\lambda_2 \mathbf{v}_2$.


It remains to show that $\lambda_1 \neq \lambda_2$.
Assume on the contrary that $\lambda:=\lambda_1=\lambda_2$.

Since $\mathbf{v}_1, \mathbf{v}_2$ are a basis of two distinct lines, they form a basis of $\R^2$.
Hence any vector $\mathbf{v}\in \R^2$ can be written as a linear combination
\[\mathbf{v}=a\mathbf{v}_1+b\mathbf{v}_2\] for some $a, b \in \R$.

Then we have
\begin{align*}
A\mathbf{v}&=A(a\mathbf{v}_1+b\mathbf{v}_2)\\
&=aA\mathbf{v}_1+bA\mathbf{v}_2\\
&=a\lambda \mathbf{v}_1+b\lambda \mathbf{v}_2\\
&=\lambda (a\mathbf{v}_1+b\mathbf{v}_2)\\
&=\lambda \mathbf{v}.
\end{align*}


This implies that any line spanned by a nonzero vector $\mathbf{v}$ is mapped onto itself by the linear transformation $T$.
This contradicts our assumption that $L_1, L_2$ are the only such lines.

Therefore, $\lambda_1 \neq \lambda_2$, and we have proved that $A$ has two distinct nonzero real eigenvalues.

(b) $\implies$ (a).

Now we suppose that the statement (b) holds.
Namely, we suppose that there are two distinct nonzero real eigenvalues of $A$. Let us call them $\lambda_1, \lambda_2$ and let $\mathbf{v}_1, \mathbf{v}_2$ be eigenvectors corresponding to $\lambda_1, \lambda_2$, respectively.


In general eigenvectors corresponding to distinct eigenvalues are linearly independent.
Thus, $\mathbf{v}_1, \mathbf{v}_2$ are linearly independent.
Hence the lines $L_1, L_2$ spanned by $\mathbf{v}_1, \mathbf{v}_2$ are distinct.

Each vector on $L_1$ is of the form $\mathbf{v}=a\mathbf{v}_1$ for some $a \in R$. Hence we have
\begin{align*}
A\mathbf{v}=A(a\mathbf{v}_1)=aA\mathbf{v}_1=a\lambda_1\mathbf{v}_1=\lambda \mathbf{v}.
\end{align*}

It follows that each vector $\mathbf{v}\in L_1$ is mapped onto $L_1$.
(Since $\lambda \neq 0$, the image under $T$ is one-dimensional, hence $T(L_1)=L_1$.)
Similarly, we have $T(L_2)=L_2$.


Finally, we show that no other lines passing through the origin are mapped onto themselves by $T$.

Assume that we have a line $L=\Span(\mathbf{w})$ such that $T(L)=L$.
Then since we have $A\mathbf{w}\in L=\Span(\mathbf{w})$, there is $\mu\in \R$ such that
\[A\mathbf{w}=\mu\mathbf{w}. \tag{*}\]

Since $\{\mathbf{v}_1, \mathbf{v}_2\}$ is a basis of $\R^2$, we have
\[\mathbf{w}=a\mathbf{v}_1+b\mathbf{v}_2\] for some $a, b\in \R$.
Then the left hand side of (*) becomes
\begin{align*}
A\mathbf{w}&=A(a\mathbf{v}_1+b\mathbf{v}_2)\\
&=aA\mathbf{v}_1+bA\mathbf{v}_2\\
&=a\lambda_1 \mathbf{v}_1+b\lambda_2 \mathbf{v}_2.
\end{align*}


Thus the relation (*) yields that
\begin{align*}
a\lambda_1 \mathbf{v}_1+b\lambda_2 \mathbf{v}_2=\mu a\mathbf{v}_1+\mu b\mathbf{v}_2\\
a(\lambda_1-\mu )\mathbf{v}_1+b(\lambda_2-\mu)\mathbf{v}_2=\mathbf{0}.
\end{align*}
Since $\mathbf{v}_1, \mathbf{v}_2$ are linearly independent, we have
\[a(\lambda_1-\mu )=0 \text{ and } b(\lambda_2-\mu)=0.\]

The equality $a(\lambda_1-\mu )=0$ implies that either $a=0$ or $\lambda_1=\mu$.
If $a=0$, then we have $\mathbf{w}=b \mathbf{v}_2$, and hence $L=L_2$.

If $\mu=\lambda_1$, then the second equality $b(\lambda_2-\mu)=0$ implies that $b=0$ since $\lambda_1\neq \lambda_2$.
Then we have $\mathbf{w}=a\mathbf{v}_1$, and hence $L=L_1$.

In either case, $L$ must be $L_1$ or $L_2$.
This completes the proof.


LoadingAdd to solve later

Sponsored Links

More from my site

You may also like...

1 Response

  1. 06/23/2017

    […] For a proof of this problem, see the post “A Linear Transformation Preserves Exactly Two Lines If and Only If There are Two Real Non-Zero Eigen…“. […]

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.

More in Linear Algebra
Kyushu University Linear Algebra Exam Problems and Solutions
Use the Cayley-Hamilton Theorem to Compute the Power $A^{100}$

Let $A$ be a $3\times 3$ real orthogonal matrix with $\det(A)=1$. (a) If $\frac{-1+\sqrt{3}i}{2}$ is one of the eigenvalues of...

Close