# Is the Derivative Linear Transformation Diagonalizable?

## Problem 690

Let $\mathrm{P}_2$ denote the vector space of polynomials of degree $2$ or less, and let $T : \mathrm{P}_2 \rightarrow \mathrm{P}_2$ be the derivative linear transformation, defined by
$T( ax^2 + bx + c ) = 2ax + b .$

Is $T$ diagonalizable? If so, find a diagonal matrix which represents $T$. If not, explain why not.

## Solution.

The standard basis of the vector space $\mathrm{P}_2$ is the set $B = \{ 1 , x , x^2 \}$. The matrix representing $T$ with respect to this basis is
$[T]_B = \begin{bmatrix} 0 & 1 & 0 \\ 0 & 0 & 2 \\ 0 & 0 & 0 \end{bmatrix} .$

The characteristic polynomial of this matrix is
$\det ( [T]_B – \lambda I ) = \begin{vmatrix} -\lambda & 1 & 0 \\ 0 & -\lambda & 2 \\ 0 & 0 & -\lambda \end{vmatrix} = \, – \lambda^3 .$ We see that the only eigenvalue of $T$ is $0$ with algebraic multiplicity $3$.

On the other hand, a polynomial $f(x)$ satisfies $T(f)(x) = 0$ if and only if $f(x) = c$ is a constant. The null space of $T$ is spanned by the single constant polynomial $\mathbb{1}(x) = 1$, and thus is one-dimensional. This means that the geometric multiplicity of the eigenvalue $0$ is only $1$.

Because the geometric multiplicity of $0$ is less than the algebraic multiplicity, the map $T$ is defective, and thus not diagonalizable.

### More from my site

• Taking the Third Order Taylor Polynomial is a Linear Transformation The space $C^{\infty} (\mathbb{R})$ is the vector space of real functions which are infinitely differentiable. Let $T : C^{\infty} (\mathbb{R}) \rightarrow \mathrm{P}_3$ be the map which takes $f \in C^{\infty}(\mathbb{R})$ to its third order Taylor polynomial, specifically defined […]
• Differentiation is a Linear Transformation Let $P_3$ be the vector space of polynomials of degree $3$ or less with real coefficients. (a) Prove that the differentiation is a linear transformation. That is, prove that the map $T:P_3 \to P_3$ defined by $T\left(\, f(x) \,\right)=\frac{d}{dx} f(x)$ for any $f(x)\in […] • Differentiating Linear Transformation is Nilpotent Let$P_n$be the vector space of all polynomials with real coefficients of degree$n$or less. Consider the differentiation linear transformation$T: P_n\to P_n$defined by $T\left(\, f(x) \,\right)=\frac{d}{dx}f(x).$ (a) Consider the case$n=2$. Let$B=\{1, x, x^2\}$be a […] • Are Linear Transformations of Derivatives and Integrations Linearly Independent? Let$W=C^{\infty}(\R)$be the vector space of all$C^{\infty}$real-valued functions (smooth function, differentiable for all degrees of differentiation). Let$V$be the vector space of all linear transformations from$W$to$W$. The addition and the scalar multiplication of$V$[…] • Matrix Representations for Linear Transformations of the Vector Space of Polynomials Let$P_2(\R)$be the vector space over$\R$consisting of all polynomials with real coefficients of degree$2$or less. Let$B=\{1,x,x^2\}$be a basis of the vector space$P_2(\R)$. For each linear transformation$T:P_2(\R) \to P_2(\R)$defined below, find the matrix representation […] • Subspace Spanned By Cosine and Sine Functions Let$\calF[0, 2\pi]$be the vector space of all real valued functions defined on the interval$[0, 2\pi]$. Define the map$f:\R^2 \to \calF[0, 2\pi]$by $\left(\, f\left(\, \begin{bmatrix} \alpha \\ \beta \end{bmatrix} \,\right) \,\right)(x):=\alpha \cos x + \beta […] • Find a Quadratic Function Satisfying Conditions on Derivatives Find a quadratic function f(x) = ax^2 + bx + c such that f(1) = 3, f'(1) = 3, and f^{\prime\prime}(1) = 2. Here, f'(x) and f^{\prime\prime}(x) denote the first and second derivatives, respectively. Solution. Each condition required on f can be turned […] • Basis with Respect to Which the Matrix for Linear Transformation is Diagonal Let P_1 be the vector space of all real polynomials of degree 1 or less. Consider the linear transformation T: P_1 \to P_1 defined by \[T(ax+b)=(3a+b)x+a+3,$ for any$ax+b\in P_1$. (a) With respect to the basis$B=\{1, x\}\$, find the matrix of the linear transformation […]

#### You may also like...

##### Dot Product, Lengths, and Distances of Complex Vectors

For this problem, use the complex vectors \[ \mathbf{w}_1 = \begin{bmatrix} 1 + i \\ 1 - i \\ 0...

Close