Let $C[-2\pi, 2\pi]$ be the vector space of all real-valued continuous functions defined on the interval $[-2\pi, 2\pi]$.
Consider the subspace $W=\Span\{\sin^2(x), \cos^2(x)\}$ spanned by functions $\sin^2(x)$ and $\cos^2(x)$.

(a) Prove that the set $B=\{\sin^2(x), \cos^2(x)\}$ is a basis for $W$.

(b) Prove that the set $\{\sin^2(x)-\cos^2(x), 1\}$ is a basis for $W$.

Let $C[-2\pi, 2\pi]$ be the vector space of all continuous functions defined on the interval $[-2\pi, 2\pi]$.
Consider the functions \[f(x)=\sin^2(x) \text{ and } g(x)=\cos^2(x)\]
in $C[-2\pi, 2\pi]$.

Prove or disprove that the functions $f(x)$ and $g(x)$ are linearly independent.

(The Ohio State University, Linear Algebra Midterm)

Let $C[-1, 1]$ be the vector space over $\R$ of all continuous functions defined on the interval $[-1, 1]$. Let
\[V:=\{f(x)\in C[-1,1] \mid f(x)=a e^x+b e^{2x}+c e^{3x}, a, b, c\in \R\}\]
be a subset in $C[-1, 1]$.

(a) Prove that $V$ is a subspace of $C[-1, 1]$.

(b) Prove that the set $B=\{e^x, e^{2x}, e^{3x}\}$ is a basis of $V$.

(c) Prove that
\[B’=\{e^x-2e^{3x}, e^x+e^{2x}+2e^{3x}, 3e^{2x}+e^{3x}\}\]
is a basis for $V$.

Let $V$ be a subset of $\R^4$ consisting of vectors that are perpendicular to vectors $\mathbf{a}, \mathbf{b}$ and $\mathbf{c}$, where
\[\mathbf{a}=\begin{bmatrix}
1 \\
0 \\
1 \\
0
\end{bmatrix}, \quad \mathbf{b}=\begin{bmatrix}
1 \\
1 \\
0 \\
0
\end{bmatrix}, \quad \mathbf{c}=\begin{bmatrix}
0 \\
1 \\
-1 \\
0
\end{bmatrix}.\]

Namely,
\[V=\{\mathbf{x}\in \R^4 \mid \mathbf{a}^{\trans}\mathbf{x}=0, \mathbf{b}^{\trans}\mathbf{x}=0, \text{ and } \mathbf{c}^{\trans}\mathbf{x}=0\}.\]

Let $A$ be an $n\times (n-1)$ matrix and let $\mathbf{b}$ be an $(n-1)$-dimensional vector.
Then the product $A\mathbf{b}$ is an $n$-dimensional vector.
Set the $n\times n$ matrix $B=[A_1, A_2, \dots, A_{n-1}, A\mathbf{b}]$, where $A_i$ is the $i$-th column vector of $A$.

Prove that $B$ is a singular matrix for any choice of $\mathbf{b}$.

By calculating the Wronskian, determine whether the set of exponential functions
\[\{e^x, e^{2x}, e^{3x}\}\]
is linearly independent on the interval $[-1, 1]$.

Let $W=C^{\infty}(\R)$ be the vector space of all $C^{\infty}$ real-valued functions (smooth function, differentiable for all degrees of differentiation).
Let $V$ be the vector space of all linear transformations from $W$ to $W$.
The addition and the scalar multiplication of $V$ are given by those of linear transformations.

Let $T_1, T_2, T_3$ be the elements in $V$ defined by
\begin{align*}
T_1\left(\, f(x) \,\right)&=\frac{\mathrm{d}}{\mathrm{d}x}f(x)\\[6pt]
T_2\left(\, f(x) \,\right)&=\frac{\mathrm{d}^2}{\mathrm{d}x^2}f(x)\\[6pt]
T_3\left(\, f(x) \,\right)&=\int_{0}^x \! f(t)\,\mathrm{d}t.
\end{align*}
Then determine whether the set $\{T_1, T_2, T_3\}$ are linearly independent or linearly dependent.

Let $\calF[0, 2\pi]$ be the vector space of all real valued functions defined on the interval $[0, 2\pi]$.
Define the map $f:\R^2 \to \calF[0, 2\pi]$ by
\[\left(\, f\left(\, \begin{bmatrix}
\alpha \\
\beta
\end{bmatrix} \,\right) \,\right)(x):=\alpha \cos x + \beta \sin x.\]
We put
\[V:=\im f=\{\alpha \cos x + \beta \sin x \in \calF[0, 2\pi] \mid \alpha, \beta \in \R\}.\]

(a) Prove that the map $f$ is a linear transformation.

(b) Prove that the set $\{\cos x, \sin x\}$ is a basis of the vector space $V$.

(c) Prove that the kernel is trivial, that is, $\ker f=\{\mathbf{0}\}$.
(This yields an isomorphism of $\R^2$ and $V$.)

(d) Define a map $g:V \to V$ by
\[g(\alpha \cos x + \beta \sin x):=\frac{d}{dx}(\alpha \cos x+ \beta \sin x)=\beta \cos x -\alpha \sin x.\]
Prove that the map $g$ is a linear transformation.

(e) Find the matrix representation of the linear transformation $g$ with respect to the basis $\{\cos x, \sin x\}$.

Let $A$ and $B$ be $n\times n$ matrices.
Suppose that $A$ and $B$ have the same eigenvalues $\lambda_1, \dots, \lambda_n$ with the same corresponding eigenvectors $\mathbf{x}_1, \dots, \mathbf{x}_n$.
Prove that if the eigenvectors $\mathbf{x}_1, \dots, \mathbf{x}_n$ are linearly independent, then $A=B$.

(a) Let $R$ be a commutative ring. If we regard $R$ as a left $R$-module, then prove that any two distinct elements of the module $R$ are linearly dependent.

(b) Let $f: M\to M’$ be a left $R$-module homomorphism. Let $\{x_1, \dots, x_n\}$ be a subset in $M$. Prove that if the set $\{f(x_1), \dots, f(x_n)\}$ is linearly independent, then the set $\{x_1, \dots, x_n\}$ is also linearly independent.

Let $f(x)=\sin^2(x)$, $g(x)=\cos^2(x)$, and $h(x)=1$. These are vectors in $C[-1, 1]$.
Determine whether the set $\{f(x), \, g(x), \, h(x)\}$ is linearly dependent or linearly independent.

(The Ohio State University, Linear Algebra Midterm Exam Problem)

Let $V$ be a vector space over $\R$ and let $B$ be a basis of $V$.
Let $S=\{v_1, v_2, v_3\}$ be a set of vectors in $V$. If the coordinate vectors of these vectors with respect to the basis $B$ is given as follows, then find the dimension of $V$ and the dimension of the span of $S$.
\[[v_1]_B=\begin{bmatrix}
1 \\
0 \\
0 \\
0
\end{bmatrix}, [v_2]_B=\begin{bmatrix}
0 \\
1 \\
0 \\
0
\end{bmatrix}, [v_3]_B=\begin{bmatrix}
1 \\
1 \\
0 \\
0
\end{bmatrix}.\]

Let $A$ be an $n\times n$ complex matrix.
Let $p(x)=\det(xI-A)$ be the characteristic polynomial of $A$ and write it as
\[p(x)=x^n+a_{n-1}x^{n-1}+\cdots+a_1x+a_0,\]
where $a_i$ are real numbers.

Let $C$ be the companion matrix of the polynomial $p(x)$ given by
\[C=\begin{bmatrix}
0 & 0 & \dots & 0 &-a_0 \\
1 & 0 & \dots & 0 & -a_1 \\
0 & 1 & \dots & 0 & -a_2 \\
\vdots & & \ddots & & \vdots \\
0 & 0 & \dots & 1 & -a_{n-1}
\end{bmatrix}=
[\mathbf{e}_2, \mathbf{e}_3, \dots, \mathbf{e}_n, -\mathbf{a}],\]
where $\mathbf{e}_i$ is the unit vector in $\C^n$ whose $i$-th entry is $1$ and zero elsewhere, and the vector $\mathbf{a}$ is defined by
\[\mathbf{a}=\begin{bmatrix}
a_0 \\
a_1 \\
\vdots \\
a_{n-1}
\end{bmatrix}.\]

Then prove that the following two statements are equivalent.

There exists a vector $\mathbf{v}\in \C^n$ such that
\[\mathbf{v}, A\mathbf{v}, A^2\mathbf{v}, \dots, A^{n-1}\mathbf{v}\]
form a basis of $\C^n$.

There exists an invertible matrix $S$ such that $S^{-1}AS=C$.
(Namely, $A$ is similar to the companion matrix of its characteristic polynomial.)