Skip to main content

Section 5.5 Linear Ordinary Differential Equations

Having studied constant coefficient linear ODE’s to a satisfactory degree, we now consider the more general case of linear ODE’s. Again we will begin with the homogeneous case which appears to us as the equation
\begin{equation} \dot{\mb{x}} (t) = A(t) \mb{x} (t). \tag{5.5.1} \end{equation}
Indeed, we could have written this \(\dot{\mb{x}} = A \mb{x}\) as before, so long as we recognize a key difference - the matrix \(A\) is time dependent. It may seem to a student that this is a minor change, but in fact the general equation may now prove nearly impossible to solve. Notwithstanding this potential danger, there are many subclasses of linear ODE’s which do admit solutions and whose solutions are of great importance. In this section we give a survey of some standard techniques available to solve some special cases.

Subsection 5.5.1 The Matrix Exponential

As it turns out, given any \(n \times n\) matrix \(A\) with real (or even complex) entries, there is a convergent power series
\begin{equation*} e^{A} = I + A + \frac{1}{2!} A^2 + \frac{1}{3!} A^3 + \cdots . \end{equation*}
What we mean by this formula is that each entry in the sum forms a convergent series. This can be seen by taking the maximum of the absolute values of the entries and applying a ratio test to the entries. Now one might imagine that if we replaced \(A\) with a path of matrices \(A (t)\) that we could apply the chain rule and get
\begin{equation*} \text{ ??? } \frac{d}{dt} e^{A (t)} = A^\prime (t) e^{A (t)} \text{ ??? } \end{equation*}
Sadly, one would be disappointed with reality here. Indeed, the main problem that occurs comes from the non-commutative nature of multiplication. To see this, you will show in an exercise that
\begin{align} \frac{d}{dt} [ A (t) B (t) ] \amp = \left[ \frac{d}{dt} A (t) \right] B (t) + A(t) \left[ \frac{d}{dt} B (t) \right], \tag{5.5.2}\\ \amp = A^\prime (t) B (t) + A(t) B^\prime (t). \tag{5.5.3} \end{align}
On the other hand, if \(A(t)\) commutes with its derivative \(A^\prime (t)\) then
\begin{equation*} \frac{d}{dt} (A (t))^n = n A^\prime (t) (A (t))^{n - 1}. \end{equation*}
So under this condition, we would have
\begin{equation} \frac{d}{dt} e^{A (t)} = A^\prime (t) e^{A (t)} \text{ if } A^\prime (t) A(t) = A(t) A^\prime (t).\tag{5.5.4} \end{equation}
Commuting matrices also work well with exponential rules as can be seen in the following lemma.
This brings us to a nice, albeit slightly opaque, theorem.
The proof of this theorem is simply equation (5.5.4). In practice, this theorem yields mixed results, but does come up as a solution in certain circumstances.

Example 5.5.3. Using exponential matrices.

Any homogeneous constant coefficient matrix can be seen to fit into this case. In particular, if \(A\) is diagonalizable with \(P^{-1} A P = \text{Diag} (\lambda_1, \ldots, \lambda_n)\) then
\begin{align*} e^{t A} \amp = P P^{-1} e^{t A} P P^{-1}, \\ \amp = P e^{t P^{-1} A P} P^{-1}, \\ \amp = P e^{\text{Diag} (t \lambda_1, \ldots, t \lambda_n)} P^{-1}, \\ \amp = P \,\text{Diag} (e^{\lambda_1 t}, \ldots, e^{\lambda_n t} ) P^{-1}. \end{align*}
If the columns of \(P\) are the eigenbasis \(\{\mb{v}_1, \ldots , \mb{v}_n\}\text{,}\) the initial condition is
\begin{equation*} \mb{x} (0) = \threevec{C_1}{\vdots}{C_n} \end{equation*}
and
\begin{equation*} P \threevec{C_1}{\vdots}{C_n} = \threevec{D_1}{\vdots}{D_n} \end{equation*}
then this reproduces our solution
\begin{equation*} \mb{x} (t) = D_1 e^{\lambda_1 t} \mb{v}_1 + \ldots + D_n e^{\lambda_n t} \mb{v}_n. \end{equation*}

Example 5.5.4. Using exponential matrices.

Consider the linear system
\begin{align*} x^\prime (t) \amp = 2 t x(t) + y(t), \\ y^\prime (t) \amp = x (t) + 2 t y(t). \end{align*}
This can be written as the matrix equation
\begin{equation*} \dot{\mb{x}} (t) = \left[ \begin{matrix} 2t \amp 1 \\ 1 \amp 2t \end{matrix} \right] \mb{x} (t). \end{equation*}
Letting
\begin{align*} A(t) = \left[ \begin{matrix} 2t \amp 1 \\ 1 \amp 2t \end{matrix} \right], \\ B(t) = \left[ \begin{matrix} t^2 \amp t \\ t \amp t^2 \end{matrix} \right], \end{align*}
One can check that \(B(t)\) and \(A(t)\) commute and that \(B^\prime (t) = A(t)\text{,}\) so by our theorem,
\begin{equation*} \mb{x} (t) = e^{B(t)} \mb{v} \end{equation*}
is a solution.
While this is very nice, the expression of this solution is a bit opaque since the exponential \(e^{B(t)}\) may be well defined, but is not readily computable. However, one can observe that \(B(t)\) is diagonalizable as a matrix of functions in the sense that
\begin{equation*} B(t) = \frac{t}{2} \left[ \begin{matrix} 1 \amp - 1 \\ 1 \amp 1 \end{matrix} \right] \left[ \begin{matrix} t - 1 \amp 0 \\ 0 \amp t + 1 \end{matrix} \right] \left[ \begin{matrix} 1 \amp 1 \\ -1 \amp 1 \end{matrix} \right] . \end{equation*}
This can be used to see that
\begin{align*} e^{B (t)} \amp = \left[ \begin{matrix} 1 \amp - 1 \\ 1 \amp 1 \end{matrix} \right] \left[ \begin{matrix} e^{t(t - 1)/2} \amp 0 \\ 0 \amp e^{t(t + 1)/2} \end{matrix} \right] \left[ \begin{matrix} 1 \amp 1 \\ -1 \amp 1 \end{matrix} \right], \\ \amp = \left[ \begin{matrix} 1 \amp - 1 \\ 1 \amp 1 \end{matrix} \right] \left[ \begin{matrix} e^{t(t - 1)/2} \amp e^{t(t - 1)/2} \\ -e^{t(t + 1)/2} \amp e^{t(t + 1)/2} \end{matrix} \right], \\ \amp = \left[ \begin{matrix} e^{t(t - 1)/2} + e^{t(t + 1)/2} \amp e^{t(t - 1)/2} - e^{t(t + 1)/2} \\ e^{t(t - 1)/2} - e^{t(t + 1)/2} \amp e^{t(t - 1)/2} + e^{t(t + 1)/2} \end{matrix} \right], \end{align*}
which in turn will give explicit solutions to our original linear system.

Subsection 5.5.2 Higher order linear ODE’s II

Given a higher order linear ODE
\begin{equation} x^{(n)} (t) + A_{n -1 } (t) x^{(n - 1)} (t) + \cdots + A_{1} (t) x^\prime (t) + A_0 (t) x(t) = f(t)\tag{5.5.5} \end{equation}
we may have limited resources to find a solution. For example, it is known that even when the coefficient functions \(A_i (t)\) are polynomials ( a so called holonomic differential equation), the solutions may have no expression in terms of exponentials, trigonometric, or polynomial functions. However, we do have qualitative results that can be asserted.
Again, when \(f(t) = 0\) we call equation (5.5.5) a homogeneous linear ODE and the solutions form a vector space. The function \(x_p (t)\) is called a particular solution and \(x_h (t)\) a complementary solution. A basis
\begin{equation*} \mathcal{B} = \left\{ x_1 (t), x_2 (t), \ldots, x_n (t) \right\} \end{equation*}
of solutions to the homogeneous equation are called fundamental solutions.
The proof of this theorem follows immediately from recasting it as a first order matrix equation with an \(n \times n\) matrix and applying the existence and uniqueness theorems (which still are to be proven). However, given any technique to find \(n\) solutions, one may wish to see whether the solutions form a basis. For this, we have the following tool.

Definition 5.5.6.

Given scalar functions \(x_1 (t), x_2(t), \ldots, x_n(t)\text{,}\) their Wronskian is
\begin{equation*} W_{x_1 x_2 \cdots x_n} = \det \left( \left[ \begin{matrix} x_1 (t) \amp x_2(t)\amp \cdots \amp x_n(t) \\ x^\prime_1 (t) \amp x^\prime_2(t)\amp \cdots \amp x^\prime_n(t) \\ \vdots \amp \vdots \amp \amp \vdots\\ x^{(n - 1)}_1 (t) \amp x^{(n - 1)}_2(t)\amp \cdots \amp x^{(n - 1)}_n (t) \end{matrix} \right] \right) \end{equation*}
The Wronskian is useful to show that one has obtained fundamental solutions.

Proof.

If \(W_{x_1 \cdots x_n} (t_0) \ne 0\) then the matrix in Definition 5.5.6 is invertible. This means the columns are linearly independent. But then any initial condition
\begin{equation*} \threevec{x (t_0)}{\vdots}{x^{(n - 1)} (t_0)} \end{equation*}
is a linear combination of those given by \(x_1, \ldots, x_n\) implying \(x\) is a linear combination of these functions (by the uniqueness of solutions with a given initial condition).

Example 5.5.8. Using the Wronskian.

Consider the differential equation
\begin{equation*} 2t^2 x^{\prime \prime} (t) + 3t x^\prime (t) - x(t) = 0. \end{equation*}
One checks that \(x_1(t) = \sqrt{t}\) and \(x_2 (t) = 1/t\) are solutions to this equation for \(t > 0\text{.}\) Calculating the Wronskian gives
\begin{equation*} W_{x_1 x_2} = \det \left[ \begin{matrix} \sqrt{t} \amp \frac{1}{t} \\ \frac{1}{2\sqrt{t}} \amp - \frac{1}{t^2} \end{matrix} \right] = - \frac{1}{t^{3/2}} - \frac{1}{2t^{1/2}}. \end{equation*}
Since this is non-zero, we conclude that \(1/t\) and \(\sqrt{t}\) are fundamental solutions.

Subsection 5.5.3 Power Series Methods

Let us return to a second order linear ODE, which is the subject of a great amount of mathematics. We will work through a sufficiently general example to illustrate the method of using power series in differential equations, but we will refrain from a full exposition on this subject as it can involve a large amount of new and unrelated techniques (but take a numerical analysis course if you are interested in this!). We consider the Bessel equation which seems to be very close to Example 5.5.8, but in fact is much more challenging. The zero-th Bessel equation is
\begin{equation*} t^2 x^{\prime \prime} (t) + t x^\prime (t) + t^2 x (t) = 0. \end{equation*}
The key difference between this and the equation given in the example is that the function in front of \(x(t)\) is no longer constant. The power series method (aka Frobenius method) is to assume that
\begin{equation*} x(t) = \sum_{n = 0}^\infty a_n t^n \end{equation*}
is a convergent power series expression for \(x(t)\) with non-zero radius of convergence. We solve for \(a_n\) using the differential equation. Indeed, putting \(x(t)\) into the equation gives
\begin{align*} 0 \amp = t^2 \left( \sum_{n = 2}^\infty n (n - 1)a_n t^{n - 2} \right) + t \left( \sum_{n = 1}^\infty n a_n t^{n - 1} \right) + t^2 \sum_{n = 0}^\infty a_n t^n, \\ \amp = \sum_{n = 2}^\infty n (n - 1)a_n t^{n } + \sum_{n = 1}^\infty n a_n t^{n} + \sum_{n = 0}^\infty a_n t^{n + 2}. \end{align*}
Now we reindex each summand on the right to get a single series.
\begin{align*} 0 \amp = \sum_{n = 2}^\infty n (n - 1)a_n t^{n } + \sum_{n = 1}^\infty n a_n t^{n} + \sum_{n = 0}^\infty a_n t^{n + 2}, \\ \amp = a_1 t + \sum_{n = 2}^\infty n (n - 1)a_n t^{n } + \sum_{n = 2}^\infty n a_n t^{n} + \sum_{n = 2}^\infty a_{n - 2} t^{n}, \\ \amp = a_1 t + \sum_{n = 2}^\infty \left( n (n - 1)a_n + n a_n + a_{n - 2} \right) t^{n}, \\ \amp = a_1 t + \sum_{n = 2}^\infty \left(n^2 a_n + a_{n - 2} \right) t^{n}, \end{align*}
Now, on the right hand side we have a power series with non-zero radius of convergence around \(0\) which is identically zero. This implies each coefficient must be zero (why?). In particular we obtain that \(a_1 = 0\) and the recursion relations
\begin{equation*} n^2 a_n + a_{n - 2} = 0. \end{equation*}
Combining the fact that \(a_1 = 0\) with this relation gives
\begin{equation*} a_1 = 0, \hspace{.2in} a_3 = 0,\hspace{.2in} a_5 = 0, \ldots \end{equation*}
so there are only even coefficients. Furthermore, we have
\begin{equation*} a_2 = \frac{-1}{2^2} a_0, \hspace{.2in}a_4 = \frac{(-1)^2}{4^2 2^2}a_0,\hspace{.2in} a_6 = \frac{(-1)^3}{6^24^2 2^2} , \ldots \end{equation*}
giving the equation
\begin{align*} a_{2n} \amp = \frac{(-1)^n}{(2n)^2(2(n - 1))^2 \cdots 2^2} a_0, \\ \amp = \frac{(-1)^n}{(2n \cdot 2(n - 1) \cdots 2)^2} a_0, \\ \amp = \frac{(-1)^n}{2^{2n}(n \cdot (n - 1) \cdots 1)^2} a_0, \\ \amp = \frac{(-1)^n}{2^{2n} \, (n!)^2} a_0 \end{align*}
Taking \(a_0 = 1\) we obtain the zeroth Bessel function
\begin{equation*} J_0 (t) = \sum_{n = 0}^\infty \frac{(-1)^n \, t^{2n}}{2^{2n} \, (n!)^2}. \end{equation*}
This function is one of several functions that has several applications in wave propagation and heat transfer. It cannot be expressed with a finite composition using our elementary functions \(t^n\text{,}\) \(\sin t\text{,}\) \(\cos t\text{,}\) \(\ln t\) and \(e^t\text{...}\) it’s brand new!

Exercises 5.5.4 Exercises

3.

Find fundamental solutions to the ordinary differential equation
\begin{equation*} 4 t^2 x^{\prime \prime} (t) + 4 t x^\prime (t) - x (t) = 0. \end{equation*}
Hint.
Try a function like \(x(t) = t^\alpha\text{.}\)

4.

The solution for Bessel’s Equation has exactly one constant \(a_0\) in its expression. However, solutions to second order differential equations normally have two parameters corresponding to the two initial conditions. Explain why there is only one in this case.

5.

Use power series methods to solve the differential equation
\begin{equation*} x^{\prime \prime} (t) - t x^\prime (t) - x (t) = 0 \end{equation*}
with initial conditions \(x(0) = 1\text{,}\) \(x^\prime (0) = 0\text{.}\)