Let $U$ and $V$ be vector spaces over a scalar field $\F$.
Define the map $T:U\to V$ by $T(\mathbf{u})=\mathbf{0}_V$ for each vector $\mathbf{u}\in U$.
(a) Prove that $T:U\to V$ is a linear transformation.
(Hence, $T$ is called the zero transformation.)
(b) Determine the null space $\calN(T)$ and the range $\calR(T)$ of $T$.
Let $T:\R^3 \to \R^3$ be the linear transformation defined by the formula
\[T\left(\, \begin{bmatrix}
x_1 \\
x_2 \\
x_3
\end{bmatrix} \,\right)=\begin{bmatrix}
x_1+3x_2-2x_3 \\
2x_1+3x_2 \\
x_2-x_3
\end{bmatrix}.\]
Determine whether $T$ is an isomorphism and if so find the formula for the inverse linear transformation $T^{-1}$.
For each of the following $3\times 3$ matrices $A$, determine whether $A$ is invertible and find the inverse $A^{-1}$ if exists by computing the augmented matrix $[A|I]$, where $I$ is the $3\times 3$ identity matrix.
Let $\mathbf{v}$ be a vector in an inner product space $V$ over $\R$.
Suppose that $\{\mathbf{u}_1, \dots, \mathbf{u}_n\}$ is an orthonormal basis of $V$.
Let $\theta_i$ be the angle between $\mathbf{v}$ and $\mathbf{u}_i$ for $i=1,\dots, n$.
Prove that
\[\cos ^2\theta_1+\cdots+\cos^2 \theta_n=1.\]
Consider the $2\times 2$ matrix
\[A=\begin{bmatrix}
\cos \theta & -\sin \theta\\
\sin \theta& \cos \theta \end{bmatrix},\]
where $\theta$ is a real number $0\leq \theta < 2\pi$.
(a) Find the characteristic polynomial of the matrix $A$.
(b) Find the eigenvalues of the matrix $A$.
(c) Determine the eigenvectors corresponding to each of the eigenvalues of $A$.
By calculating the Wronskian, determine whether the set of exponential functions
\[\{e^x, e^{2x}, e^{3x}\}\]
is linearly independent on the interval $[-1, 1]$.
An $n\times n$ matrix $A$ is said to be invertible if there exists an $n\times n$ matrix $B$ such that
$AB=I$, and
$BA=I$,
where $I$ is the $n\times n$ identity matrix.
If such a matrix $B$ exists, then it is known to be unique and called the inverse matrix of $A$, denoted by $A^{-1}$.
In this problem, we prove that if $B$ satisfies the first condition, then it automatically satisfies the second condition.
So if we know $AB=I$, then we can conclude that $B=A^{-1}$.
Let $A$ and $B$ be $n\times n$ matrices.
Suppose that we have $AB=I$, where $I$ is the $n \times n$ identity matrix.
The $(i, j)$ cofactor $C_{ij}$ of $A$ is defined to be
\[C_{ij}=(-1)^{ij}\det(M_{ij}),\]
where $M_{ij}$ is the $(i,j)$ minor matrix obtained from $A$ removing the $i$-th row and $j$-th column.
Then consider the $n\times n$ matrix $C=(C_{ij})$, and define the $n\times n$ matrix $\Adj(A)=C^{\trans}$.
The matrix $\Adj(A)$ is called the adjoint matrix of $A$.
When $A$ is invertible, then its inverse can be obtained by the formula
\[A^{-1}=\frac{1}{\det(A)}\Adj(A).\]
For each of the following matrices, determine whether it is invertible, and if so, then find the invertible matrix using the above formula.
Consider the $2\times 2$ real matrix
\[A=\begin{bmatrix}
1 & 1\\
1& 3
\end{bmatrix}.\]
(a) Prove that the matrix $A$ is positive definite.
(b) Since $A$ is positive definite by part (a), the formula
\[\langle \mathbf{x}, \mathbf{y}\rangle:=\mathbf{x}^{\trans} A \mathbf{y}\]
for $\mathbf{x}, \mathbf{y} \in \R^2$ defines an inner product on $\R^n$.
Consider $\R^2$ as an inner product space with this inner product.
Prove that the unit vectors
\[\mathbf{e}_1=\begin{bmatrix}
1 \\
0
\end{bmatrix} \text{ and } \mathbf{e}_2=\begin{bmatrix}
0 \\
1
\end{bmatrix}\]
are not orthogonal in the inner product space $\R^2$.
(c) Find an orthogonal basis $\{\mathbf{v}_1, \mathbf{v}_2\}$ of $\R^2$ from the basis $\{\mathbf{e}_1, \mathbf{e}_2\}$ using the Gram-Schmidt orthogonalization process.
(a) Suppose that $A$ is an $n\times n$ real symmetric positive definite matrix.
Prove that
\[\langle \mathbf{x}, \mathbf{y}\rangle:=\mathbf{x}^{\trans}A\mathbf{y}\]
defines an inner product on the vector space $\R^n$.
(b) Let $A$ be an $n\times n$ real matrix. Suppose that
\[\langle \mathbf{x}, \mathbf{y}\rangle:=\mathbf{x}^{\trans}A\mathbf{y}\]
defines an inner product on the vector space $\R^n$.
Prove that $A$ is symmetric and positive definite.
Consider the complex matrix
\[A=\begin{bmatrix}
\sqrt{2}\cos x & i \sin x & 0 \\
i \sin x &0 &-i \sin x \\
0 & -i \sin x & -\sqrt{2} \cos x
\end{bmatrix},\]
where $x$ is a real number between $0$ and $2\pi$.
Determine for which values of $x$ the matrix $A$ is diagonalizable.
When $A$ is diagonalizable, find a diagonal matrix $D$ so that $P^{-1}AP=D$ for some nonsingular matrix $P$.
Let $V$ denote the vector space of all real $2\times 2$ matrices.
Suppose that the linear transformation from $V$ to $V$ is given as below.
\[T(A)=\begin{bmatrix}
2 & 3\\
5 & 7
\end{bmatrix}A-A\begin{bmatrix}
2 & 3\\
5 & 7
\end{bmatrix}.\]
Prove or disprove that the linear transformation $T:V\to V$ is an isomorphism.
A square matrix $A$ is called idempotent if $A^2=A$.
(a) Let $\mathbf{u}$ be a vector in $\R^n$ with length $1$.
Define the matrix $P$ to be $P=\mathbf{u}\mathbf{u}^{\trans}$.
Prove that $P$ is an idempotent matrix.
(b) Suppose that $\mathbf{u}$ and $\mathbf{v}$ be unit vectors in $\R^n$ such that $\mathbf{u}$ and $\mathbf{v}$ are orthogonal.
Let $Q=\mathbf{u}\mathbf{u}^{\trans}+\mathbf{v}\mathbf{v}^{\trans}$.
Prove that $Q$ is an idempotent matrix.
(c) Prove that each nonzero vector of the form $a\mathbf{u}+b\mathbf{v}$ for some $a, b\in \R$ is an eigenvector corresponding to the eigenvalue $1$ for the matrix $Q$ in part (b).