Tagged: linear independent

Determine Conditions on Scalars so that the Set of Vectors is Linearly Dependent

Problem 279

Determine conditions on the scalars $a, b$ so that the following set $S$ of vectors is linearly dependent.
\begin{align*}
S=\{\mathbf{v}_1, \mathbf{v}_2, \mathbf{v}_3\},
\end{align*}
where
\[\mathbf{v}_1=\begin{bmatrix}
1 \\
3 \\
1
\end{bmatrix}, \mathbf{v}_2=\begin{bmatrix}
1 \\
a \\
4
\end{bmatrix}, \mathbf{v}_3=\begin{bmatrix}
0 \\
2 \\
b
\end{bmatrix}.\]  
Read solution

LoadingAdd to solve later

Linear Combination of Eigenvectors is Not an Eigenvector

Problem 258

Suppose that $\lambda$ and $\mu$ are two distinct eigenvalues of a square matrix $A$ and let $\mathbf{x}$ and $\mathbf{y}$ be eigenvectors corresponding to $\lambda$ and $\mu$, respectively.
If $a$ and $b$ are nonzero numbers, then prove that $a \mathbf{x}+b\mathbf{y}$ is not an eigenvector of $A$ (corresponding to any eigenvalue of $A$).

 
Read solution

LoadingAdd to solve later

Compute Determinant of a Matrix Using Linearly Independent Vectors

Problem 193

Let $A$ be a $3 \times 3$ matrix.
Let $\mathbf{x}, \mathbf{y}, \mathbf{z}$ are linearly independent $3$-dimensional vectors. Suppose that we have
\[A\mathbf{x}=\begin{bmatrix}
1 \\
0 \\
1
\end{bmatrix}, A\mathbf{y}=\begin{bmatrix}
0 \\
1 \\
0
\end{bmatrix}, A\mathbf{z}=\begin{bmatrix}
1 \\
1 \\
1
\end{bmatrix}.\]

Then find the value of the determinant of the matrix $A$.

 
Read solution

LoadingAdd to solve later

Find a Basis and the Dimension of the Subspace of the 4-Dimensional Vector Space

Problem 131

Let $V$ be the following subspace of the $4$-dimensional vector space $\R^4$.
\[V:=\left\{ \quad\begin{bmatrix}
x_1 \\
x_2 \\
x_3 \\
x_4
\end{bmatrix} \in \R^4
\quad \middle| \quad
x_1-x_2+x_3-x_4=0 \quad\right\}.\] Find a basis of the subspace $V$ and its dimension.

 
Read solution

LoadingAdd to solve later

If Vectors are Linearly Dependent, then What Happens When We Add One More Vectors?

Problem 120

Suppose that $\mathbf{v}_1, \mathbf{v}_2, \dots, \mathbf{v}_r$ are linearly dependent $n$-dimensional real vectors.

For any vector $\mathbf{v}_{r+1} \in \R^n$, determine whether the vectors $\mathbf{v}_1, \mathbf{v}_2, \dots, \mathbf{v}_r, \mathbf{v}_{r+1}$ are linearly independent or linearly dependent.

 
Read solution

LoadingAdd to solve later

A Linear Transformation from Vector Space over Rational Numbers to itself

Problem 75

Let $\Q$ denote the set of rational numbers (i.e., fractions of integers). Let $V$ denote the set of the form $x+y \sqrt{2}$ where $x,y \in \Q$. You may take for granted that the set $V$ is a vector space over the field $\Q$.

(a) Show that $B=\{1, \sqrt{2}\}$ is a basis for the vector space $V$ over $\Q$.

(b) Let $\alpha=a+b\sqrt{2} \in V$, and let $T_{\alpha}: V \to V$ be the map defined by
\[ T_{\alpha}(x+y\sqrt{2}):=(ax+2by)+(ay+bx)\sqrt{2}\in V\] for any $x+y\sqrt{2} \in V$.
Show that $T_{\alpha}$ is a linear transformation.

(c) Let $\begin{bmatrix}
x \\
y
\end{bmatrix}_B=x+y \sqrt{2}$.
Find the matrix $T_B$ such that
\[ T_{\alpha} (x+y \sqrt{2})=\left( T_B\begin{bmatrix}
x \\
y
\end{bmatrix}\right)_B,\] and compute $\det T_B$.

 

(The Ohio State University, Linear Algebra Exam)

Read solution

LoadingAdd to solve later

Linear Independent Vectors, Invertible Matrix, and Expression of a Vector as a Linear Combinations

Problem 66

Consider the matrix
\[A=\begin{bmatrix}
1 & 2 & 1 \\
2 &5 &4 \\
1 & 1 & 0
\end{bmatrix}.\]


(a) Calculate the inverse matrix $A^{-1}$. If you think the matrix $A$ is not invertible, then explain why.


(b) Are the vectors
\[ \mathbf{A}_1=\begin{bmatrix}
1 \\
2 \\
1
\end{bmatrix}, \mathbf{A}_2=\begin{bmatrix}
2 \\
5 \\
1
\end{bmatrix},
\text{ and } \mathbf{A}_3=\begin{bmatrix}
1 \\
4 \\
0
\end{bmatrix}\] linearly independent?


(c) Write the vector $\mathbf{b}=\begin{bmatrix}
1 \\
1 \\
1
\end{bmatrix}$ as a linear combination of $\mathbf{A}_1$, $\mathbf{A}_2$, and $\mathbf{A}_3$.

(The Ohio State University, Linear Algebra Exam)

Read solution

LoadingAdd to solve later

If the Images of Vectors are Linearly Independent, then They Are Linearly Independent

Problem 62

Let $T: \R^n \to \R^m$ be a linear transformation.
Suppose that $S=\{\mathbf{x}_1, \mathbf{x}_2,\dots, \mathbf{x}_k\}$ is a subset of $\R^n$ such that $\{T(\mathbf{x}_1), T(\mathbf{x}_2), \dots, T(\mathbf{x}_k) \}$ is a linearly independent subset of $\R^m$.

Prove that the set $S$ is linearly independent.

 

Read solution

LoadingAdd to solve later

Two Subspaces Intersecting Trivially, and the Direct Sum of Vector Spaces.

Problem 61

Let $V$ and $W$ be subspaces of $\R^n$ such that $V \cap W =\{\mathbf{0}\}$ and $\dim(V)+\dim(W)=n$.

(a) If $\mathbf{v}+\mathbf{w}=\mathbf{0}$, where $\mathbf{v}\in V$ and $\mathbf{w}\in W$, then show that $\mathbf{v}=\mathbf{0}$ and $\mathbf{w}=\mathbf{0}$.

(b) If $B_1$ is a basis for the subspace $V$ and $B_2$ is a basis for the subspace $W$, then show that the union $B_1\cup B_2$ is a basis for $R^n$.

(c) If $\mathbf{x}$ is in $\R^n$, then show that $\mathbf{x}$ can be written in the form $\mathbf{x}=\mathbf{v}+\mathbf{w}$, where $\mathbf{v}\in V$ and $\mathbf{w} \in W$.

(d) Show that the representation obtained in part (c) is unique.

Read solution

LoadingAdd to solve later

Properties of Nonsingular and Singular Matrices

Problem 25

An $n \times n$ matrix $A$ is called nonsingular if the only solution of the equation $A \mathbf{x}=\mathbf{0}$ is the zero vector $\mathbf{x}=\mathbf{0}$.
Otherwise $A$ is called singular.

(a) Show that if $A$ and $B$ are $n\times n$ nonsingular matrices, then the product $AB$ is also nonsingular.

(b) Show that if $A$ is nonsingular, then the column vectors of $A$ are linearly independent.

(c) Show that an $n \times n$ matrix $A$ is nonsingular if and only if the equation $A\mathbf{x}=\mathbf{b}$ has a unique solution for any vector $\mathbf{b}\in \R^n$.

Restriction
Do not use the fact that a matrix is nonsingular if and only if the matrix is invertible.

Read solution

LoadingAdd to solve later

Linear Dependent/Independent Vectors of Polynomials

Problem 15

Let $p_1(x), p_2(x), p_3(x), p_4(x)$ be (real) polynomials of degree at most $3$. Which (if any) of the following two conditions is sufficient for the conclusion that these polynomials are linearly dependent?

(a) At $1$ each of the polynomials has the value $0$. Namely $p_i(1)=0$ for $i=1,2,3,4$.

(b) At $0$ each of the polynomials has the value $1$. Namely $p_i(0)=1$ for $i=1,2,3,4$.

(University of California, Berkeley)

Read solution

LoadingAdd to solve later