###### Theorem2.3.1Superposition

Suppose \(y_1\text{,}\) \(y_2\text{,}\) , \(y_n\) are solutions of the homogeneous equation (4). Then

also solves (4) for arbitrary constants \(C_1, C_2, \ldots, C_n\text{.}\)

\(\require{cancel}\newcommand{\nicefrac}[2]{{{}^{#1}}\!/\!{{}_{#2}}}
\newcommand{\unitfrac}[3][\!\!]{#1 \,\, {{}^{#2}}\!/\!{{}_{#3}}}
\newcommand{\unit}[2][\!\!]{#1 \,\, #2}
\newcommand{\noalign}[1]{}
\newcommand{\qed}{\qquad \Box}
\newcommand{\lt}{<}
\newcommand{\gt}{>}
\newcommand{\amp}{&}
\)

*somewhat more than 1 lecture, §3.2 and §3.3 in [EP], §4.1 and §4.2 in [BD]*

We briefly study higher order equations. Equations appearing in applications tend to be second order. Higher order equations do appear from time to time, but generally the world around us is “second order.”

The basic results about linear ODEs of higher order are essentially the same as for second order equations, with 2 replaced by \(n\text{.}\) The important concept of linear independence is somewhat more complicated when more than two functions are involved. For higher order constant coefficient ODEs, the methods developed are also somewhat harder to apply, but we will not dwell on these complications. It is also possible to use the methods for systems of linear equations from Chapter 3 to solve higher order constant coefficient equations.

Let us start with a general homogeneous linear equation

\begin{equation}
y^{(n)} + p_{n-1}(x)y^{(n-1)} + \cdots + p_1(x) y' + p_0(x) y = 0 .\label{hol_eqlinhom}\tag{4}
\end{equation}

Suppose \(y_1\text{,}\) \(y_2\text{,}\) , \(y_n\) are solutions of the homogeneous equation (4). Then

\begin{equation*}
y(x) = C_1 y_1(x) + C_2 y_2(x) + \cdots + C_n y_n(x)
\end{equation*}

also solves (4) for arbitrary constants \(C_1, C_2, \ldots, C_n\text{.}\)

In other words, a *linear combination* of solutions to (4) is also a solution to (4). We also have the existence and uniqueness theorem for nonhomogeneous linear equations.

Suppose \(p_0\) through \(p_{n-1}\text{,}\) and \(f\) are continuous functions on some interval \(I\text{,}\) \(a\) is a number in \(I\text{,}\) and \(b_0, b_1, \ldots, b_{n-1}\) are constants. The equation

\begin{equation*}
y^{(n)} + p_{n-1}(x)y^{(n-1)} + \cdots + p_1(x) y' + p_0(x) y = f(x)
\end{equation*}

has exactly one solution \(y(x)\) defined on the same interval \(I\) satisfying the initial conditions

\begin{equation*}
y(a) = b_0, \quad y'(a) = b_1, \quad \ldots, \quad y^{(n-1)}(a) = b_{n-1} .
\end{equation*}

When we had two functions \(y_1\) and \(y_2\) we said they were linearly independent if one was not the multiple of the other. Same idea holds for \(n\) functions. In this case it is easier to state as follows. The functions \(y_1\text{,}\) \(y_2\text{,}\) , \(y_n\) are *linearly independent* if the equation

\begin{equation*}
c_1 y_1 + c_2 y_2 + \cdots + c_n y_n = 0
\end{equation*}

has only the trivial solution \(c_1 = c_2 = \cdots = c_n = 0\text{,}\) where the equation must hold for all \(x\text{.}\) If we can solve equation with some constants where for example \(c_1 \not= 0\text{,}\) then we can solve for \(y_1\) as a linear combination of the others. If the functions are not linearly independent, they are *linearly dependent*.

Show that \(e^x, e^{2x}, e^{3x}\) are linearly independent.

Let us give several ways to show this fact. Many textbooks (including [EP] and [F]) introduce Wronskians, but it is difficult to see why they work and they are not really necessary here.

Let us write down

\begin{equation*}
c_1 e^x + c_2 e^{2x} + c_3 e^{3x} = 0.
\end{equation*}

We use rules of exponentials and write \(z = e^x\text{.}\) Hence \(z^2 = e^{2x}\) and \(z^3 = e^{3x}\text{.}\) Then we have

\begin{equation*}
c_1 z + c_2 z^2 + c_3 z^3 = 0.
\end{equation*}

The left hand side is a third degree polynomial in \(z\text{.}\) It is either identically zero, or it has at most 3 zeros. Therefore, it is identically zero, \(c_1 = c_2 = c_3 = 0\text{,}\) and the functions are linearly independent.

Let us try another way. As before we write

\begin{equation*}
c_1 e^x + c_2 e^{2x} + c_3 e^{3x} = 0.
\end{equation*}

This equation has to hold for all \(x\text{.}\) We divide through by \(e^{3x}\) to get

\begin{equation*}
c_1 e^{-2x} + c_2 e^{-x} + c_3 = 0.
\end{equation*}

As the equation is true for all \(x\text{,}\) let \(x \to \infty\text{.}\) After taking the limit we see that \(c_3 = 0\text{.}\) Hence our equation becomes

\begin{equation*}
c_1 e^x + c_2 e^{2x} = 0.
\end{equation*}

Rinse, repeat!

How about yet another way. We again write

\begin{equation*}
c_1 e^x + c_2 e^{2x} + c_3 e^{3x} = 0.
\end{equation*}

We can evaluate the equation and its derivatives at different values of \(x\) to obtain equations for \(c_1\text{,}\) \(c_2\text{,}\) and \(c_3\text{.}\) Let us first divide by \(e^{x}\) for simplicity.

\begin{equation*}
c_1 + c_2 e^{x} + c_3 e^{2x} = 0.
\end{equation*}

We set \(x=0\) to get the equation \(c_1 + c_2 + c_3 = 0\text{.}\) Now differentiate both sides

\begin{equation*}
c_2 e^{x} + 2 c_3 e^{2x} = 0 .
\end{equation*}

We set \(x=0\) to get \(c_2 + 2c_3 = 0\text{.}\) We divide by \(e^x\) again and differentiate to get \(2 c_3 e^{x} = 0\text{.}\) It is clear that \(c_3\) is zero. Then \(c_2\) must be zero as \(c_2 = -2c_3\text{,}\) and \(c_1\) must be zero because \(c_1 + c_2 + c_3 = 0\text{.}\)

There is no one best way to do it. All of these methods are perfectly valid. The important thing is to understand why the functions are linearly independent.

On the other hand, the functions \(e^x\text{,}\) \(e^{-x}\text{,}\) and \(\cosh x\) are linearly dependent. Simply apply definition of the hyperbolic cosine:

\begin{equation*}
\cosh x = \frac{e^x + e^{-x}}{2}
\qquad
\text{or}
\qquad
2 \cosh x - e^x - e^{-x} = 0.
\end{equation*}

When we have a higher order constant coefficient homogeneous linear equation, the song and dance is exactly the same as it was for second order. We just need to find more solutions. If the equation is \(n^{\text{th}}\) order we need to find \(n\) linearly independent solutions. It is best seen by example.

Find the general solution to

\begin{equation}
y''' - 3 y'' - y' + 3y = 0 .\label{hol_cceq1}\tag{5}
\end{equation}

Try: \(y = e^{rx}\text{.}\) We plug in and get

\begin{equation*}
\underbrace{r^3 e^{rx}}_{y'''} - 3 \underbrace{r^2 e^{rx}}_{y''} -
\underbrace{r e^{rx}}_{y'} + 3 \underbrace{e^{rx}}_{y} = 0 .
\end{equation*}

We divide through by \(e^{rx}\text{.}\) Then

\begin{equation*}
r^3 - 3 r^2 - r + 3 = 0 .
\end{equation*}

The trick now is to find the roots. There is a formula for the roots of degree 3 and 4 polynomials but it is very complicated. There is no formula for higher degree polynomials. That does not mean that the roots do not exist. There are always \(n\) roots for an \(n^{\text{th}}\) degree polynomial. They may be repeated and they may be complex. Computers are pretty good at finding roots approximately for reasonable size polynomials.

A good place to start is to plot the polynomial and check where it is zero. We can also simply try plugging in. We just start plugging in numbers \(r=-2,-1,0,1,2,\ldots\) and see if we get a hit (we can also try complex numbers). Even if we do not get a hit, we may get an indication of where the root is. For example, we plug \(r=-2\) into our polynomial and get \(-15\text{;}\) we plug in \(r=0\) and get 3. That means there is a root between \(r=-2\) and \(r=0\text{,}\) because the sign changed. If we find one root, say \(r_1\text{,}\) then we know \((r-r_1)\) is a factor of our polynomial. Polynomial long division can then be used.

A good strategy is to begin with \(r=-1\text{,}\) 1, or 0. These are easy to compute. Our polynomial happens to have two such roots, \(r_1 = -1\) and \(r_2 = 1\text{.}\) There should be 3 roots and the last root is reasonably easy to find. The constant term in a monic polynomial such as this is the multiple of the negations of all the roots because \(r^3 - 3 r^2 - r + 3 = (r-r_1)(r-r_2)(r-r_3)\text{.}\) So

\begin{equation*}
3 = (-r_1)(-r_2)(-r_3) = (1)(-1)(-r_3) = r_3 .
\end{equation*}

You should check that \(r_3 = 3\) really is a root. Hence we know that \(e^{-x}\text{,}\) \(e^{x}\) and \(e^{3x}\) are solutions to (5). They are linearly independent as can easily be checked, and there are 3 of them, which happens to be exactly the number we need. Hence the general solution is

\begin{equation*}
y = C_1 e^{-x} + C_2 e^{x} + C_3 e^{3x} .
\end{equation*}

Suppose we were given some initial conditions \(y(0) = 1\text{,}\) \(y'(0) = 2\text{,}\) and \(y''(0) = 3\text{.}\) Then

\begin{equation*}
\begin{aligned}
1 = y(0) & = C_1 + C_2 + C_3 , \\
2 = y'(0) & = -C_1 + C_2 + 3C_3 , \\
3 = y''(0) & = C_1 + C_2 + 9C_3 .
\end{aligned}
\end{equation*}

It is possible to find the solution by high school algebra, but it would be a pain. The sensible way to solve a system of equations such as this is to use matrix algebra, see Section 3.2. For now we note that the solution is \(C_1 = -\nicefrac{1}{4}\text{,}\) \(C_2 = 1\text{,}\) and \(C_3 = \nicefrac{1}{4}\text{.}\) The specific solution to the ODE is

\begin{equation*}
y = \frac{-1}{4}\, e^{-x} + e^x + \frac{1}{4}\, e^{3x} .
\end{equation*}

Next, suppose that we have real roots, but they are repeated. Let us say we have a root \(r\) repeated \(k\) times. In the spirit of the second order solution, and for the same reasons, we have the solutions

\begin{equation*}
e^{rx}, \quad xe^{rx}, \quad x^2 e^{rx}, \quad \ldots, \quad x^{k-1} e^{rx} .
\end{equation*}

We take a linear combination of these solutions to find the general solution.

Solve

\begin{equation*}
y^{(4)} - 3 y''' + 3 y'' - y' = 0 .
\end{equation*}

We note that the characteristic equation is

\begin{equation*}
r^4 - 3r^3 + 3r^2 -r = 0 .
\end{equation*}

By inspection we note that \(r^4 - 3r^3 + 3r^2 -r = r{(r-1)}^3\text{.}\) Hence the roots given with multiplicity are \(r = 0, 1, 1, 1\text{.}\) Thus the general solution is

\begin{equation*}
y = \underbrace{(C_1 + C_2 x + C_3 x^2)\, e^x}_{\text{terms coming from }
r=1} + \underbrace{C_4}_{\text{from } r=0} .
\end{equation*}

The case of complex roots is similar to second order equations. Complex roots always come in pairs \(r = \alpha \pm i \beta\text{.}\) Suppose we have two such complex roots, each repeated \(k\) times. The corresponding solution is

\begin{equation*}
( C_0 + C_1 x + \cdots + C_{k-1} x^{k-1} ) \, e^{\alpha x} \cos (\beta x)
+
( D_0 + D_1 x + \cdots + D_{k-1} x^{k-1} ) \, e^{\alpha x} \sin (\beta x) .
\end{equation*}

where \(C_0\text{,}\) , \(C_{k-1}\text{,}\) \(D_0\text{,}\) , \(D_{k-1}\) are arbitrary constants.

Solve

\begin{equation*}
y^{(4)} - 4 y''' + 8 y'' - 8 y' + 4y = 0 .
\end{equation*}

The characteristic equation is

\begin{equation*}
\begin{aligned}
r^4 - 4 r^3 + 8 r^2 - 8 r + 4 & = 0 , \\
{(r^2-2r+2)}^2 & = 0 , \\
{\bigl({(r-1)}^2+1\bigr)}^2 & = 0 .
\end{aligned}
\end{equation*}

Hence the roots are \(1 \pm i\text{,}\) both with multiplicity 2. Hence the general solution to the ODE is

\begin{equation*}
y =
( C_1 + C_2 x ) \, e^{x} \cos x
+
( C_3 + C_4 x ) \, e^{x} \sin x .
\end{equation*}

The way we solved the characteristic equation above is really by guessing or by inspection. It is not so easy in general. We could also have asked a computer or an advanced calculator for the roots.

Find the general solution for \(y''' - y'' + y' - y = 0\text{.}\)

Find the general solution for \(y^{(4)} - 5 y''' + 6 y'' = 0\text{.}\)

Find the general solution for \(y''' + 2 y'' + 2 y' = 0\text{.}\)

Suppose the characteristic equation for a differential equation is \({(r-1)}^2{(r-2)}^2 = 0\text{.}\) a) Find such a differential equation. b) Find its general solution.

Suppose that a fourth order equation has a solution \(y = 2 e^{4x} x \cos x\text{.}\) a) Find such an equation. b) Find the initial conditions that the given solution satisfies.

Find the general solution for the equation of Exercise 2.3.5.

Let \(f(x) = e^x - \cos x\text{,}\) \(g(x) = e^x + \cos x\text{,}\) and \(h(x) = \cos x\text{.}\) Are \(f(x)\text{,}\) \(g(x)\text{,}\) and \(h(x)\) linearly independent? If so, show it, if not, find a linear combination that works.

Let \(f(x) = 0\text{,}\) \(g(x) = \cos x\text{,}\) and \(h(x) = \sin x\text{.}\) Are \(f(x)\text{,}\) \(g(x)\text{,}\) and \(h(x)\) linearly independent? If so, show it, if not, find a linear combination that works.

Are \(x\text{,}\) \(x^2\text{,}\) and \(x^4\) linearly independent? If so, show it, if not, find a linear combination that works.

Are \(e^x\text{,}\) \(xe^x\text{,}\) and \(x^2e^x\) linearly independent? If so, show it, if not, find a linear combination that works.

Find an equation such that \(y=xe^{-2x}\sin(3x)\) is a solution.

Find the general solution of \(y^{(5)}-y^{(4)}=0\)

Answer

\(y=C_1 e^x +C_2 x^3 + C_3 x^2 +C_4 x + C_5\)

Suppose that the characteristic equation of a third order differential equation has roots \(\pm 2i\) and 3. a) What is the characteristic equation? b) Find the corresponding differential equation. c) Find the general solution.

Answer

a) \(r^3-3r^2+4r-12 = 0\text{,}\) b) \(y'''-3y''+4y'-12y = 0\text{,}\) c) \(y = C_1 e^{3x} + C_2 \sin(2x) + C_3 \cos(2x)\)

Solve \(1001y'''+3.2y''+\pi y'-\sqrt{4} y = 0\text{,}\) \(y(0)=0\text{,}\) \(y'(0) = 0\text{,}\) \(y''(0) = 0\text{.}\)

Answer

\(y=0\)

Are \(e^{x}\text{,}\) \(e^{x+1}\text{,}\) \(e^{2x}\text{,}\) \(\sin(x)\) linearly independent? If so, show it, if not find a linear combination that works.

Answer

No. \(e^1 e^x - e^{x+1} = 0\text{.}\)

Are \(\sin(x)\text{,}\) \(x\text{,}\) \(x\sin(x)\) linearly independent? If so, show it, if not find a linear combination that works.

Answer

Yes. (Hint: First note that \(\sin(x)\) is bounded. Then note that \(x\) and \(x\sin(x)\) cannot be multiples of each other.)

Find an equation such that \(y=\cos(x)\text{,}\) \(y=\sin(x)\text{,}\) \(y=e^x\) are solutions.

Answer

\(y'''-y''+y'-y=0\)