Linear Differential Operators
Linear Differential Operators.
This section is divided into three parts. In the first, we shall systematically develop and extend the differential operators [math]D^2 + aD + b[/math] which were introduced in Section 1. In the second part we shall use these operators to obtain directly the general solutions of certain linear differential equations with constant coefficients. Finally, we shall show how these methods can be used to solve any linear differential equation with constant coefficients (whether homogeneous or not) provided we extend our range of functions to include those whose values may be complex numbers. By a \textbf{linear operator} we shall mean any function [math]L[/math] whose domain and range are sets of numerical-valued functions and which satisfies the equations \setcounter{equation}{0}
for every real number [math]k[/math] and every [math]y, y_1[/math], and [math]y_2[/math] in the domain of [math]L[/math]. [The function [math]L(y)[/math] is frequently written simply [math]L_y[/math].] An important example is the function [math]D[/math], which, to every differentiable function [math]y[/math], assigns its derivative [math]Dy = \frac{dy}{dx}[/math]. Another example is the operation of multiplication by a real number. That is, for any real number [math]a[/math], the function [math]L[/math] defined by
obviously satisfies (1) and (2) and hence is a linear operator. If [math]L_1[/math] and [math]L_2[/math] are linear operators, then their sum is the function [math]L_1 + L_2[/math] defined by
for every [math]y[/math] which is in the domains of both [math]L_1[/math] and [math]L_2[/math]. lt is easy to show that
If [math]L_1[/math] and [math]L_2[/math] nre linear operato's, then the sum [math]L_1 + L_2[/math] is also a linear operator.
We shall show that [math]L_1 + L_2[/math], satisfies equation (1) by using successively the definition of [math]L_1 + L_2[/math], the linearity of [math]L_1[/math] and [math]L_2[/math], separately, the commutative law of addition for functions, and finally the definition again. Thus
The proof that [math]L_1 + L_2[/math] satisfies (2) is similar:
If [math]L_1[/math] and [math]L_2[/math] are linear operators, then the composition of [math]L_2[/math], followed by [math]L_1[/math] is the function denoted by [math]L_1L_2[/math] and defined by
for every [math]y[/math] for which the right side is defined. The proof of the following proposition is entirely analogous to that of (3.1) and is left to the reader as an exercise.
If [math]L_1[/math] and [math]L_2[/math] are linear operators, then the composition [math]L_1L_2[/math] is also a /inear operator.
The composition [math]L_1L_2[/math] is also called the product of [math]L_1[/math] and [math]L_2[/math]. There is no reason to suppose from the definition that the commutative law of multiplication holds, and, for linear operators in general, [math]L_1L_2 \neq L_2L_1[/math]. However, the distributive laws hold:
The first of these is proved as follows:
The proof of the second is similar and is left as an exercise.
An important example of the product of linear operators is the composition of a linear operator [math]L[/math] followed by the operation of multiplication by a real number [math]a[/math]. This product, denoted [math]aL[/math], assigns to every [math]y[/math] in the domain of [math]L[/math] the value [math](aL)y[/math] which is equal to the product of a with the function [math]Ly[/math]. That is,
{{{4}}}
The composition in the other order is the product [math]La[/math]. Here we have [math](La)y = L(ay)[/math], and the latter quantity, by the linearity of [math]L[/math] is equal to [math]a(Ly)[/math]. Combining this with (3.4), we obtain the equation [math](La)y = (aL)y[/math]. Thus the operators [math]La[/math] and [math]aL[/math] are equal, and we have proved the following special case of the commutative law:
{{{4}}}
Another example of the product, already encountered, is the operator [math]D0[/math], which is the composition [math]D^2 = D D[/math] of [math]D[/math] with itself. More generally, for every integer [math]n \gt 1[/math], we define the operator [math]D^n[/math] inductively by
The domain of [math]D^n[/math] is the set of all e-times differentiable functions, and, for each such function [math]y[/math], we have
By repeated applications of (3.1) and (3.2), we may conclude that any function formed in a finite number of steps by taking sums and products of linear operators is itself a linear operator. As an example, consider a polynomial [math]p(t)[/math] of degree [math]n[/math]; i.e.,
where [math]a_0, . . ., a_n[/math] are real numbers and an [math]a_n \neq 0[/math]. Then the function
is a linear operator. To every [math]n[/math]-times differentiable function [math]y[/math], it assigns as value the function
We call [math]p(D)[/math] a linear differential operator of order $n$. It is the natural generalization of the differential operators of order 2, of the form [math]D^2 + aD + b[/math], which were discussed in Section 1. [Linear differential operators of types more general than [math]p(D)[/math] certainly exist; e.g., see Problem 9. They are of importance in more advanced treatments of differential equations, but we shall not study them here.] The polynomial differential operators [math]p(D)[/math] can be added and multiplied just like ordinary polynomials. In particular, the following theorem follows from the distributive laws (3.3) and the commutative law (3.5):
If [math]p(t)[/math] and [math]q(t)[/math] are polynomials and if [math]p(t)q(t) = r(t)[/math], then
As an illustration, observe how (3.3) and (3.5) are used to prove the special case of this theorem in which [math]p(t) = at + b[/math] and [math]q(t) = ct + d[/math]. First of all, we have
Then
The proof is the same in principle for arbitrary polynomials [math]p(t)[/math] and [math]q(t)[/math]. It is a corollary of (3.6) that polynomial differential operators satisfy the commutative law of multiplication. Thus
{{{4}}}
For, since [math]p(t)q(t) = q(t)p(t) = r(t)[/math], both sides of (3.7) are equal to [math]r(D)[/math]. \medskip We begin the second part of the section by considering the differential equation
which, with the notation of differential operators, can be written
We have thus far defined the characteristic equation only for homogeneous, second-order, linear differential equations with constant coefficients. The generalization to nonhomogeneous and higher-order equations is: For any polynomial [math]p(t)[/math] and function [math]F(x)[/math], the characteristic equation of the differential equation
is the equation [math]p(t) = 0[/math], and the polynomial [math]p(t)[/math] is its characteristic polynomial. Returning to (5), we see that the characteristic polynomial, which is [math]t^2 - 2t - 3[/math], factors into the product [math](t - 3)(t + 1)[/math]. It follows from (3.6) that [math]D^2 - 2D - 3 = (D - 3)(D + 1)[/math], and (5) can therefore be written
Let us define the function [math]u[/math] by setting [math](D + 1)y = u[/math]. Then (5) becomes equivalent to the pair of first-order linear equations
To solve (6), we use the technique developed in Section 2. For this equation, [math]P(x) = -3[/math] and [math]Q(x) = e^{-x}[/math]. Hence an integrating factor is [math]e^{\int P(x)dx} = e^{-3x} [/math], and therefore
Integrating, we obtain
whence
We now substitute this value for [math]u[/math] in equation (7) to obtain the first-order linear equation
Here, [math]P(x) = 1[/math] and the integrating factor is [math]e^{x}[/math]. Accordingly, we have
integration yields
Replacing [math]\frac{c_1}{4}[/math] by [math]c_1[/math], and multiplying both sides by [math]e^{-x}[/math], we get finally
This, where [math]c_1[/math] and [math]c_2[/math] are arbitrary real constants, is the general solution to the differential equation
This example illustrates the fact that we can in principle solve any secondorder, linear differential equation with constant coefficients provided the characteristic polynomial is the product of linear factors. Thus, if we are given
and if [math]t^2+ at + b = (t - r_1)(t - r_2)[/math], then the differential equation can be written
If [math]u[/math] is defined by setting [math](D - r_2)y= u[/math], then the original second-order equation is equivalent to the two first-order linear differential equations
and these can be solved successively to find first [math]u[/math] and then [math]y[/math]. The same technique can be applied to higher-order equations. Consider an arbitrary polynomial
where [math]n \gt 1[/math] and [math]a_0, . . ., a_{n-1}[/math], are real constants. In addition, we assume that [math]p(t)[/math] is the product of linear factors; i.e.,
Let F(x) be given and consider the differential equation \setcounter{equation}{7}
which is the same as
Since the factorization of [math]p(t)[/math] is assumed, the differential equation can also be written
The functions [math]u_1, . . ., u_{n-1}[/math] are defined by
Then (8) is equivalent to the following set of first-order linear differential equations
which can be solved successively to finally obtain [math]y[/math]. \medskip ln Section 4 of Chapter 7 use was made of the fact that any polynomial with real coefficients and degree at least 1 can be written as the product of linear and irreducible quadratic factors (see page 386). Suppose [math]ct^2 + dt + e[/math] is irreducible. This is equivalent to the assertion that the discriminant [math]d^2 - 4ce[/math] is negative. According to the quadratic formula, the two roots of the equation [math]ct^2 + dt + e = 0[/math] are equal to [math]r_1 = \alpha + i\beta[/math] and [math]r_2 = \alpha - i\beta[/math], where [math]\alpha = - \frac{d}{2c}[/math] and [math]\beta = \frac{\sqrt{4ae - d^2}}{2c}[/math]. By multiplying and substituting these values, one can then easily verify the equation
Thus any irreducible quadratic polynomial with real coefficients is the product of two linear factors with complex coefficients. It follows that, for any polynomial
is a subset of the real numbers and whose range is a subset of the complex numbers. Then two real-valued functions [math]f_1[/math] and [math]f_2[/math] with domain [math]Q[/math] are defined by
That is, we have [math]f(x) = f_1(x) + if_2(x)[/math], for every [math]x[/math] in [math]Q[/math]. The derivative [math]f'[/math] is defined simply by the equation
for every [math]x[/math] for which both [math]f'_1(x)[/math] and [math]f'_2(x)[/math] exist. Alternatively, if we write [math]y = f(x), u = f_1(x)[/math], and [math]v = f_2(x)[/math], then [math]y = u + iv[/math], and we also use the notations
Logically, we must now go back and check that all the formal rules for differentiation and finding antiderivatives are still true for complex-valued functions, and the same applies to several theorems (see, for example, Problems 10 and 11). Much of this work is purely routine, and, to avoid an interruption of our study of differential equations, we shall omit it. It now follows, by factoring the operator [math]p(D)[/math] into linear factors, that any linear differential equation
with constant coefficients can be solved. That is, it can first be replaced by an equivalent set of first-order linear differential equations. For each of these an explicit integrating factor [math]e^{\int P(x)dx}[/math] exists, and by solving them successively, we can eventually obtain the general solution [math]y[/math].
Example Solve the differential equation [math](D^2 + 1)y = 2x[/math]. Since [math]t^2 + 1 = (t + i)(t - i)[/math], we have
Let [math](D - i)y = u[/math], and consider the first-order equation
Since [math]P(x) = i[/math], an integrating factor is [math]e^{ix}[/math], and we obtain
from which it follows by integrating that
By integration by parts it can be verified that
In this case, [math]a = i[/math] and we know that [math]\frac{1}{i} = -i[/math] and that [math]i^2 = - 1[/math]. Hence
and so
It therefore remains to solve the differential equation
This time, an integrating factor is [math]e^{-ix}[/math]. Hence
Integration [with a second application of (9)] yields
Replacing the constant [math]-\frac{c_1}{2i}[/math] by simply [math]c_1[/math], and multiplying both sides by [math]e^{ix}[/math], we obtain
If the function [math]y[/math] is real-valued, then it is easy to prove that [math]c_1[/math] and [math]c_2[/math] are complex conjugates [see (4.3), page 644]. In this case [math]c_1e^{-ix} + c_2e^{ix}[/math] may be replaced by [math]c_1 \cos x + c_2 \sin x[/math], where now the constants [math]c_1[/math] and [math]c_2[/math] denote arbitrary real numbers. We conclude that
is the general real-valued solution to the original differential equation
The computations in this section were long and involved. The important fact we have shown is that the equations can be solved by an iteration of routine steps. As a practical matter, however, it is clear that some general computationally simple techniques are badly needed. These will be developed in the next two sections by breaking the problem into a homogeneous part and a nonhomogeneous part and attacking each one separately.
\end{exercise}
General references
Doyle, Peter G. (2008). "Crowell and Slesnick's Calculus with Analytic Geometry" (PDF). Retrieved Oct 29, 2024.