Difference between revisions of "Linear Differential Equations"
(4 intermediate revisions by 2 users not shown) | |||
Line 1: | Line 1: | ||
− | In | + | In mathematics, a '''linear differential equation''' is a differential equation that is defined by a linear polynomial in the unknown function and its derivatives, that is an equation of the form |
:<math>a_0(x)y + a_1(x)y' + a_2(x)y'' \cdots + a_n(x)y^{(n)} = b(x)</math> | :<math>a_0(x)y + a_1(x)y' + a_2(x)y'' \cdots + a_n(x)y^{(n)} = b(x)</math> | ||
− | where | + | where ''a''<sub>0</sub>(''x''), …, ''a''<sub>''n''</sub>(''x'') and {{math|''b''(''x'')}} are arbitrary differentiable functions that do not need to be linear, and {{math|''y''′, …, ''y''<sup>(''n'')</sup> }} are the successive derivatives of an unknown function {{mvar|y}} of the variable {{mvar|x}}. |
− | Such an equation is an | + | Such an equation is an ordinary differential equation (ODE). A ''linear differential equation'' may also be a linear partial differential equation (PDE), if the unknown function depends on several variables, and the derivatives that appear in the equation are partial derivatives. |
− | A linear differential equation or a system of linear equations such that the associated homogeneous equations have constant coefficients may be solved by | + | A linear differential equation or a system of linear equations such that the associated homogeneous equations have constant coefficients may be solved by quadrature, which means that the solutions may be expressed in terms of integrals. This is also true for a linear equation of order one, with non-constant coefficients. An equation of order two or higher with non-constant coefficients cannot, in general, be solved by quadrature. For order two, Kovacic's algorithm allows deciding whether there are solutions in terms of integrals, and computing them if any. |
− | The solutions of linear differential equations with | + | The solutions of linear differential equations with polynomial coefficients are called holonomic functions. This class of functions is stable under sums, products, differentiation, integration, and contains many usual functions and special functions such as exponential function, logarithm, sine, cosine, inverse trigonometric functions, error function, Bessel functions and hypergeometric functions. Their representation by the defining differential equation and initial conditions allows making algorithmic (on these functions) most operations of calculus, such as computation of antiderivatives, limits, asymptotic expansion, and numerical evaluation to any precision, with a certified error bound. |
==Basic terminology== | ==Basic terminology== | ||
− | The highest | + | The highest order of derivation that appears in a (linear) differential equation is the ''order'' of the equation. The term {{math|''b''(''x'')}}, which does not depend on the unknown function and its derivatives, is sometimes called the ''constant term'' of the equation (by analogy with algebraic equations), even when this term is a non-constant function. If the constant term is the zero function, then the differential equation is said to be ''homogeneous'', as it is a homogeneous polynomial in the unknown function and its derivatives. The equation obtained by replacing, in a linear differential equation, the constant term by the zero function is the ''associated homogeneous equation''. A differential equation has ''constant coefficients'' if only constant functions appear as coefficients in the associated homogeneous equation. |
A ''solution'' of a differential equation is a function that satisfies the equation. | A ''solution'' of a differential equation is a function that satisfies the equation. | ||
− | The solutions of a homogeneous linear differential equation form a | + | The solutions of a homogeneous linear differential equation form a vector space. In the ordinary case, this vector space has a finite dimension, equal to the order of the equation. All solutions of a linear differential equation are found by adding to a particular solution any solution of the associated homogeneous equation. |
==Linear differential operator== | ==Linear differential operator== | ||
− | + | ||
− | A ''basic differential operator'' of order {{mvar|i}} is a mapping that maps any | + | A ''basic differential operator'' of order {{mvar|i}} is a mapping that maps any differentiable function to its {{mvar|i}}th derivative, or, in the case of several variables, to one of its partial derivatives of order {{mvar|i}}. It is commonly denoted |
:<math>\frac{d^i}{dx^i}</math> | :<math>\frac{d^i}{dx^i}</math> | ||
− | in the case of | + | in the case of univariate functions, and |
:<math>\frac{\partial^{i_1+\cdots +i_n}}{\partial x_1^{i_1}\cdots \partial x_n^{i_n}}</math> | :<math>\frac{\partial^{i_1+\cdots +i_n}}{\partial x_1^{i_1}\cdots \partial x_n^{i_n}}</math> | ||
in the case of functions of {{mvar|n}} variables. The basic differential operators include the derivative of order 0, which is the identity mapping. | in the case of functions of {{mvar|n}} variables. The basic differential operators include the derivative of order 0, which is the identity mapping. | ||
− | A '''linear differential operator''' (abbreviated, in this article, as ''linear operator'' or, simply, ''operator'') is a | + | A '''linear differential operator''' (abbreviated, in this article, as ''linear operator'' or, simply, ''operator'') is a linear combination of basic differential operators, with differentiable functions as coefficients. In the univariate case, a linear operator has thus the form |
:<math>a_0(x)+a_1(x)\frac{d}{dx} + \cdots +a_n(x)\frac{d^n}{dx^n},</math> | :<math>a_0(x)+a_1(x)\frac{d}{dx} + \cdots +a_n(x)\frac{d^n}{dx^n},</math> | ||
− | where {{math|''a''<sub>0</sub>(''x''), …, ''a''<sub>''n''</sub>(''x'')}} are differentiable functions, and the nonnegative integer {{mvar|n}} is the ''order'' of the operator (if {{math|''a''<sub>''n''</sub>(''x'')}} is not the | + | where {{math|''a''<sub>0</sub>(''x''), …, ''a''<sub>''n''</sub>(''x'')}} are differentiable functions, and the nonnegative integer {{mvar|n}} is the ''order'' of the operator (if {{math|''a''<sub>''n''</sub>(''x'')}} is not the zero function). |
− | Let {{mvar|L}} be a linear differential operator. The application of {{mvar|L}} to a function {{mvar|f}} is usually denoted {{math|''Lf''}} or {{math|''Lf''(''X'')}}, if one needs to specify the variable (this must not be confused with a multiplication). A linear differential operator is a | + | Let {{mvar|L}} be a linear differential operator. The application of {{mvar|L}} to a function {{mvar|f}} is usually denoted {{math|''Lf''}} or {{math|''Lf''(''X'')}}, if one needs to specify the variable (this must not be confused with a multiplication). A linear differential operator is a linear operator, since it maps sums to sums and the product by a scalar to the product by the same scalar. |
− | As the sum of two linear operators is a linear operator, as well as the product (on the left) of a linear operator by a differentiable function, the linear differential operators form a | + | As the sum of two linear operators is a linear operator, as well as the product (on the left) of a linear operator by a differentiable function, the linear differential operators form a vector space over the real numbers or the complex numbers (depending on the nature of the functions that are considered). They form also a free module over the ring of differentiable functions. |
The language of operators allows a compact writing for differentiable equations: if | The language of operators allows a compact writing for differentiable equations: if | ||
Line 40: | Line 40: | ||
There may be several variants to this notation; in particular the variable of differentiation may appear explicitly or not in {{mvar|y}} and the right-hand and of the equation, such as {{math|1=''Ly''(''x'') = ''b''(''x'')}} or {{math|1=''Ly'' = ''b''}}. | There may be several variants to this notation; in particular the variable of differentiation may appear explicitly or not in {{mvar|y}} and the right-hand and of the equation, such as {{math|1=''Ly''(''x'') = ''b''(''x'')}} or {{math|1=''Ly'' = ''b''}}. | ||
− | The ''kernel'' of a linear differential operator is its | + | The ''kernel'' of a linear differential operator is its kernel as a linear mapping, that is the vector space of the solutions of the (homogeneous) differential equation {{math|1=''Ly'' = 0}}. |
− | In the case of an ordinary differential operator of order {{mvar|n}}, | + | In the case of an ordinary differential operator of order {{mvar|n}}, Carathéodory's existence theorem implies that, under very mild conditions, the kernel of {{mvar|L}} is a vector space of dimension {{mvar|n}}, and that the solutions of the equation {{math|1=''Ly''(''x'') = ''b''(''x'')}} have the form |
:<math>S_0(x) + c_1S_1(x) + \cdots +c_nS_n(x),</math> | :<math>S_0(x) + c_1S_1(x) + \cdots +c_nS_n(x),</math> | ||
− | where {{math|''c''<sub>1</sub>, …, ''c''<sub>''n''</sub>}} are arbitrary numbers. Typically, the hypotheses of Carathéodory's theorem are satisfied in an interval {{mvar|I}}, if the functions {{math|''b'', ''a''<sub>0</sub>, …, ''a''<sub>''n''</sub>}} are continuous in {{mvar|I}}, and there is a positive real number {{mvar|k}} such that | + | where {{math|''c''<sub>1</sub>, …, ''c''<sub>''n''</sub>}} are arbitrary numbers. Typically, the hypotheses of Carathéodory's theorem are satisfied in an interval {{mvar|I}}, if the functions {{math|''b'', ''a''<sub>0</sub>, …, ''a''<sub>''n''</sub>}} are continuous in {{mvar|I}}, and there is a positive real number {{mvar|k}} such that |''a''<sub>''n''</sub>(''x'')| > ''k'' for every {{mvar|x}} in {{mvar|I}}. |
==Homogeneous equation with constant coefficients== | ==Homogeneous equation with constant coefficients== | ||
Line 51: | Line 51: | ||
where {{math|''a''<sub>1</sub>, …, ''a''<sub>''n''</sub>}} are (real or complex) numbers. In other words, it has constant coefficients if it is defined by a linear operator with constant coefficients. | where {{math|''a''<sub>1</sub>, …, ''a''<sub>''n''</sub>}} are (real or complex) numbers. In other words, it has constant coefficients if it is defined by a linear operator with constant coefficients. | ||
− | The study of these differential equations with constant coefficients dates back to | + | The study of these differential equations with constant coefficients dates back to Leonhard Euler, who introduced the exponential function {{math|''e''<sup>''x''</sup>}}, which is the unique solution of the equation {{math|1=''f''′ = ''f''}} such that {{math|1=''f''(0) = 1}}. It follows that the {{mvar|n}}th derivative of {{math|''e''<sup>''cx''</sup> }} is {{math|''c''<sup>''n''</sup>''e''<sup>''cx''</sup>}}, and this allows solving homogeneous linear differential equations rather easily. |
Let | Let | ||
Line 61: | Line 61: | ||
Factoring out {{math|''e''<sup>''αx''</sup>}} (which is never zero), shows that {{mvar|α}} must be a root of the ''characteristic polynomial'' | Factoring out {{math|''e''<sup>''αx''</sup>}} (which is never zero), shows that {{mvar|α}} must be a root of the ''characteristic polynomial'' | ||
:<math>a_0 + a_1t + a_2t^2 + \cdots + a_nt^n</math> | :<math>a_0 + a_1t + a_2t^2 + \cdots + a_nt^n</math> | ||
− | of the differential equation, which is the left-hand side of the | + | of the differential equation, which is the left-hand side of the characteristic equation |
:<math>a_0 + a_1t + a_2t^2 + \cdots + a_nt^n = 0.</math> | :<math>a_0 + a_1t + a_2t^2 + \cdots + a_nt^n = 0.</math> | ||
− | When these roots are all | + | When these roots are all distinct, one has {{mvar|n}} distinct solutions that are not necessarily real, even if the coefficients of the equation are real. These solutions can be shown to be linearly independent, by considering the Vandermonde determinant of the values of these solutions at {{math|1=''x'' = 0, …, ''n'' – 1}}. Together they form a basis of the vector space of solutions of the differential equation (that is, the kernel of the differential operator). |
{| class="toccolours floatright" style="width:35%; margin: 0.5em 0 0.5em 1em;" | {| class="toccolours floatright" style="width:35%; margin: 0.5em 0 0.5em 1em;" | ||
! style="background:#ffffaa; padding: 3px 5px 3px 5px; font-size:larger;" | Example | ! style="background:#ffffaa; padding: 3px 5px 3px 5px; font-size:larger;" | Example | ||
Line 78: | Line 78: | ||
|} | |} | ||
− | In the case where the characteristic polynomial has only | + | In the case where the characteristic polynomial has only simple roots, the preceding provides a complete basis of the solutions vector space. In the case of multiple roots, more linearly independent solutions are needed for having a basis. These have the form |
:<math>x^ke^{\alpha x},</math> | :<math>x^ke^{\alpha x},</math> | ||
− | where {{mvar|k}} is a nonnegative integer, {{mvar|α}} is a root of the characteristic polynomial of multiplicity {{mvar|m}}, and {{math|''k'' < ''m''}}. For proving that these functions are solutions, one may remark that if {{mvar|α}} is a root of the characteristic polynomial of multiplicity {{mvar|m}}, the characteristic polynomial may be factored as {{math|''P''(''t'') (''t'' − ''α'')<sup>''m''</sup>}}. Thus, applying the differential operator of the equation is equivalent with applying first {{mvar|m}} times the operator | + | where {{mvar|k}} is a nonnegative integer, {{mvar|α}} is a root of the characteristic polynomial of multiplicity {{mvar|m}}, and {{math|''k'' < ''m''}}. For proving that these functions are solutions, one may remark that if {{mvar|α}} is a root of the characteristic polynomial of multiplicity {{mvar|m}}, the characteristic polynomial may be factored as {{math|''P''(''t'') (''t'' − ''α'')<sup>''m''</sup>}}. Thus, applying the differential operator of the equation is equivalent with applying first {{mvar|m}} times the operator <math display="inline"> \frac{d}{dx} - \alpha </math>, and then the operator that has {{mvar|P}} as characteristic polynomial. By the exponential shift theorem, |
:<math>\left(\frac{d}{dx}-\alpha\right)\left(x^ke^{\alpha x}\right)= kx^{k-1}e^{\alpha x},</math> | :<math>\left(\frac{d}{dx}-\alpha\right)\left(x^ke^{\alpha x}\right)= kx^{k-1}e^{\alpha x},</math> | ||
− | and thus one gets zero after {{math|''k'' + 1}} application of | + | and thus one gets zero after {{math|''k'' + 1}} application of <math display="inline"> \frac{d}{dx} - \alpha </math>. |
− | As, by the | + | As, by the fundamental theorem of algebra, the sum of the multiplicities of the roots of a polynomial equals the degree of the polynomial, the number of above solutions equals the order of the differential equation, and these solutions form a base of the vector space of the solutions. |
− | In the common case where the coefficients of the equation are real, it is generally more convenient to have a basis of the solutions consisting of | + | In the common case where the coefficients of the equation are real, it is generally more convenient to have a basis of the solutions consisting of real-valued functions. Such a basis may be obtained from the preceding basis by remarking that, if {{math|''a'' + ''ib''}} is a root of the characteristic polynomial, then {{math|''a'' – ''ib''}} is also a root, of the same multiplicity. Thus a real basis is obtained by using Euler's formula, and replacing {{math|''x''<sup>''k''</sup>''e''<sup>(''a''+''ib'')''x''</sup>}} and {{math|''x''<sup>''k''</sup>''e''<sup>(''a''−''ib'')''x''</sup>}} by {{math|''x''<sup>''k''</sup>''e''<sup>''ax''</sup> cos(''bx'')}} and {{math|''x''<sup>''k''</sup>''e''<sup>''ax''</sup> sin(''bx'')}}. |
===Second-order case=== | ===Second-order case=== | ||
Line 95: | Line 95: | ||
:<math>r^2 + ar + b.</math> | :<math>r^2 + ar + b.</math> | ||
− | If {{mvar|a}} and {{mvar|b}} are | + | If {{mvar|a}} and {{mvar|b}} are real, there are three cases for the solutions, depending on the discriminant {{math|1=''D'' = ''a''<sup>2</sup> − 4''b''}}. In all three cases, the general solution depends on two arbitrary constants {{math|''c''<sub>1</sub>}} and {{math|''c''<sub>2</sub>}}. |
* If {{math|''D'' > 0}}, the characteristic polynomial has two distinct real roots {{mvar|α}}, and {{mvar|β}}. In this case, the general solution is | * If {{math|''D'' > 0}}, the characteristic polynomial has two distinct real roots {{mvar|α}}, and {{mvar|β}}. In this case, the general solution is | ||
Line 101: | Line 101: | ||
* If {{math|1=''D'' = 0}}, the characteristic polynomial has a double root {{math|−''a''/2}}, and the general solution is | * If {{math|1=''D'' = 0}}, the characteristic polynomial has a double root {{math|−''a''/2}}, and the general solution is | ||
::<math>(c_1 + c_2 x) e^{-ax/2}.</math> | ::<math>(c_1 + c_2 x) e^{-ax/2}.</math> | ||
− | * If {{math|''D'' < 0}}, the characteristic polynomial has two | + | * If {{math|''D'' < 0}}, the characteristic polynomial has two complex conjugate roots {{math|''α'' ± ''βi''}}, and the general solution is |
::<math>c_1 e^{(\alpha + \beta i)x} + c_2 e^{(\alpha - \beta i)x},</math> | ::<math>c_1 e^{(\alpha + \beta i)x} + c_2 e^{(\alpha - \beta i)x},</math> | ||
− | :which may be rewritten in real terms, using | + | :which may be rewritten in real terms, using Euler's formula as |
::<math> e^{\alpha x} (c_1\cos(\beta x) + c_2 \sin(\beta x)).</math> | ::<math> e^{\alpha x} (c_1\cos(\beta x) + c_2 \sin(\beta x)).</math> | ||
− | Finding the solution {{math|''y''(''x'')}} satisfying {{math|1=''y''(0) = ''d''<sub>1</sub>}} and {{math|1=''y''′(0) = ''d''<sub>2</sub>}}, one equates the values of the above general solution at {{math|0}} and its derivative there to {{math|''d''<sub>1</sub>}} and {{math|''d''<sub>2</sub>}}, respectively. This results in a linear system of two linear equations in the two unknowns {{math|''c''<sub>1</sub>}} and {{math|''c''<sub>2</sub>}}. Solving this system gives the solution for a so-called | + | Finding the solution {{math|''y''(''x'')}} satisfying {{math|1=''y''(0) = ''d''<sub>1</sub>}} and {{math|1=''y''′(0) = ''d''<sub>2</sub>}}, one equates the values of the above general solution at {{math|0}} and its derivative there to {{math|''d''<sub>1</sub>}} and {{math|''d''<sub>2</sub>}}, respectively. This results in a linear system of two linear equations in the two unknowns {{math|''c''<sub>1</sub>}} and {{math|''c''<sub>2</sub>}}. Solving this system gives the solution for a so-called Cauchy problem, in which the values at {{math|0}} for the solution of the DEQ and its derivative are specified. |
==Non-homogeneous equation with constant coefficients== | ==Non-homogeneous equation with constant coefficients== | ||
Line 113: | Line 113: | ||
where {{math|''a''<sub>1</sub>, …, ''a''<sub>''n''</sub>}} are real or complex numbers, {{mvar|f}} is a given function of {{mvar|x}}, and {{mvar|y}} is the unknown function (for sake of simplicity, "{{math|(''x'')}}" will be omitted in the following). | where {{math|''a''<sub>1</sub>, …, ''a''<sub>''n''</sub>}} are real or complex numbers, {{mvar|f}} is a given function of {{mvar|x}}, and {{mvar|y}} is the unknown function (for sake of simplicity, "{{math|(''x'')}}" will be omitted in the following). | ||
− | There are several methods for solving such an equation. The best method depends on the nature of the function {{mvar|f}} that makes the equation non-homogeneous. If {{mvar|f}} is a linear combination of exponential and sinusoidal functions, then the | + | There are several methods for solving such an equation. The best method depends on the nature of the function {{mvar|f}} that makes the equation non-homogeneous. If {{mvar|f}} is a linear combination of exponential and sinusoidal functions, then the exponential response formula may be used. If, more generally, {{mvar|f}} is a linear combination of functions of the form {{math|''x''<sup>''n''</sup>''e''<sup>''ax''</sup>}}, {{math|''x''<sup>''n''</sup> cos(''ax'')}}, and {{math|''x''<sup>''n''</sup> sin(''ax'')}}, where {{mvar|n}} is a nonnegative integer, and {{mvar|a}} a constant (which need not be the same in each term), then the method of undetermined coefficients may be used. Still more general, the annihilator method applies when {{mvar|f}} satisfies a homogeneous linear differential equation, typically, a holonomic function. |
− | The most general method is the | + | The most general method is the variation of constants, which is presented here. |
The general solution of the associated homogeneous equation | The general solution of the associated homogeneous equation | ||
Line 128: | Line 128: | ||
0 &= u'_1y^{(n-2)}_1+u'_2y^{(n-2)}_2 + \cdots + u'_n y^{(n-2)}_n, | 0 &= u'_1y^{(n-2)}_1+u'_2y^{(n-2)}_2 + \cdots + u'_n y^{(n-2)}_n, | ||
\end{align}</math> | \end{align}</math> | ||
− | which imply (by | + | which imply (by product rule and induction) |
:<math>y^{(i)} = u_1 y_1^{(i)} + \cdots + u_n y_n^{(i)}</math> | :<math>y^{(i)} = u_1 y_1^{(i)} + \cdots + u_n y_n^{(i)}</math> | ||
for {{math|1=''i'' = 1, …, ''n'' – 1}}, and | for {{math|1=''i'' = 1, …, ''n'' – 1}}, and | ||
Line 136: | Line 136: | ||
:<math>f=u'_1y_1^{(n-1)} + \cdots + u'_ny_n^{(n-1)}.</math> | :<math>f=u'_1y_1^{(n-1)} + \cdots + u'_ny_n^{(n-1)}.</math> | ||
− | This equation and the above ones with {{math|0}} as left-hand side form a system of {{mvar|n}} linear equations in {{math|''u''′<sub>1</sub>, …, ''u''′<sub>''n''</sub>}} whose coefficients are known functions ({{mvar|f}}, the | + | This equation and the above ones with {{math|0}} as left-hand side form a system of {{mvar|n}} linear equations in {{math|''u''′<sub>1</sub>, …, ''u''′<sub>''n''</sub>}} whose coefficients are known functions ({{mvar|f}}, the <math>y_i</math>, and their derivatives). This system can be solved by any method of linear algebra. The computation of antiderivatives gives {{math|''u''<sub>1</sub>, …, ''u''<sub>''n''</sub>}}, and then {{math|1=''y'' = ''u''<sub>1</sub>''y''<sub>1</sub> + ⋯ + ''u''<sub>''n''</sub>''y''<sub>''n''</sub>}}. |
As antiderivatives are defined up to the addition of a constant, one finds again that the general solution of the non-homogeneous equation is the sum of an arbitrary solution and the general solution of the associated homogeneous equation. | As antiderivatives are defined up to the addition of a constant, one finds again that the general solution of the non-homogeneous equation is the sum of an arbitrary solution and the general solution of the associated homogeneous equation. | ||
Line 146: | Line 146: | ||
If the equation is homogeneous, i.e. {{math|1=''g''(''x'') = 0}}, one may rewrite and integrate: | If the equation is homogeneous, i.e. {{math|1=''g''(''x'') = 0}}, one may rewrite and integrate: | ||
:<math>\frac{y'}{y}= f, \qquad \log y = k +F, </math> | :<math>\frac{y'}{y}= f, \qquad \log y = k +F, </math> | ||
− | where {{mvar|k}} is an arbitrary | + | where {{mvar|k}} is an arbitrary constant of integration and <math>F=\textstyle\int f\,dx</math> is any antiderivative of {{mvar|f}}. Thus, the general solution of the homogeneous equation is |
:<math>y=ce^F,</math> | :<math>y=ce^F,</math> | ||
where {{math|1=''c'' = ''e''<sup>''k''</sup>}} is an arbitrary constant. | where {{math|1=''c'' = ''e''<sup>''k''</sup>}} is an arbitrary constant. | ||
− | For the general non-homogeneous equation, one may multiply it by the | + | For the general non-homogeneous equation, one may multiply it by the reciprocal {{math|''e''<sup>−''F''</sup>}} of a solution of the homogeneous equation. This gives |
:<math>y'e^{-F}-yfe^{-F}= ge^{-F}.</math> | :<math>y'e^{-F}-yfe^{-F}= ge^{-F}.</math> | ||
− | As | + | As <math>-fe^{-F} = \frac{d}{dx} \left(e^{-F}\right),</math> the product rule allows rewriting the equation as |
:<math>\frac{d}{dx}\left(ye^{-F}\right)= ge^{-F}.</math> | :<math>\frac{d}{dx}\left(ye^{-F}\right)= ge^{-F}.</math> | ||
Thus, the general solution is | Thus, the general solution is | ||
Line 234: | Line 234: | ||
:::<math>y^2=5+\frac{C}{e^{x^2}} \,</math> | :::<math>y^2=5+\frac{C}{e^{x^2}} \,</math> | ||
− | == | + | ==Higher order linear equations== |
− | + | A linear ordinary equation of order one with variable coefficients may be solved by quadrature, which means that the solutions may be expressed in terms of integrals. This is not the case for order at least two. This is the main result of Picard–Vessiot theory which was initiated by Émile Picard and Ernest Vessiot, and whose recent developments are called differential Galois theory. | |
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | A linear ordinary equation of order one with variable coefficients may be solved by | ||
− | The impossibility of solving by quadrature can be compared with the | + | The impossibility of solving by quadrature can be compared with the Abel–Ruffini theorem, which states that an algebraic equation of degree at least five cannot, in general, be solved by radicals. This analogy extends to the proof methods and motivates the denomination of differential Galois theory. |
Similarly to the algebraic case, the theory allows deciding which equations may be solved by quadrature, and if possible solving them. However, for both theories, the necessary computations are extremely difficult, even with the most powerful computers. | Similarly to the algebraic case, the theory allows deciding which equations may be solved by quadrature, and if possible solving them. However, for both theories, the necessary computations are extremely difficult, even with the most powerful computers. | ||
− | Nevertheless, the case of order two with rational coefficients has been completely solved by | + | Nevertheless, the case of order two with rational coefficients has been completely solved by Kovacic's algorithm. |
===Cauchy–Euler equation=== | ===Cauchy–Euler equation=== | ||
− | + | Cauchy–Euler equations are examples of equations of any order, with variable coefficients, that can be solved explicitly. These are the equations of the form | |
:<math>x^n y^{(n)}(x) + a_{n-1} x^{n-1} y^{(n-1)}(x) + \cdots + a_0 y(x) = 0,</math> | :<math>x^n y^{(n)}(x) + a_{n-1} x^{n-1} y^{(n-1)}(x) + \cdots + a_0 y(x) = 0,</math> | ||
− | where | + | where <math>a_0, \ldots, a_{n-1} </math> are constant coefficients. |
== Licensing == | == Licensing == |
Latest revision as of 20:01, 5 November 2021
In mathematics, a linear differential equation is a differential equation that is defined by a linear polynomial in the unknown function and its derivatives, that is an equation of the form
where a0(x), …, an(x) and b(x) are arbitrary differentiable functions that do not need to be linear, and y′, …, y(n) are the successive derivatives of an unknown function y of the variable x.
Such an equation is an ordinary differential equation (ODE). A linear differential equation may also be a linear partial differential equation (PDE), if the unknown function depends on several variables, and the derivatives that appear in the equation are partial derivatives.
A linear differential equation or a system of linear equations such that the associated homogeneous equations have constant coefficients may be solved by quadrature, which means that the solutions may be expressed in terms of integrals. This is also true for a linear equation of order one, with non-constant coefficients. An equation of order two or higher with non-constant coefficients cannot, in general, be solved by quadrature. For order two, Kovacic's algorithm allows deciding whether there are solutions in terms of integrals, and computing them if any.
The solutions of linear differential equations with polynomial coefficients are called holonomic functions. This class of functions is stable under sums, products, differentiation, integration, and contains many usual functions and special functions such as exponential function, logarithm, sine, cosine, inverse trigonometric functions, error function, Bessel functions and hypergeometric functions. Their representation by the defining differential equation and initial conditions allows making algorithmic (on these functions) most operations of calculus, such as computation of antiderivatives, limits, asymptotic expansion, and numerical evaluation to any precision, with a certified error bound.
Contents
Basic terminology
The highest order of derivation that appears in a (linear) differential equation is the order of the equation. The term b(x), which does not depend on the unknown function and its derivatives, is sometimes called the constant term of the equation (by analogy with algebraic equations), even when this term is a non-constant function. If the constant term is the zero function, then the differential equation is said to be homogeneous, as it is a homogeneous polynomial in the unknown function and its derivatives. The equation obtained by replacing, in a linear differential equation, the constant term by the zero function is the associated homogeneous equation. A differential equation has constant coefficients if only constant functions appear as coefficients in the associated homogeneous equation.
A solution of a differential equation is a function that satisfies the equation. The solutions of a homogeneous linear differential equation form a vector space. In the ordinary case, this vector space has a finite dimension, equal to the order of the equation. All solutions of a linear differential equation are found by adding to a particular solution any solution of the associated homogeneous equation.
Linear differential operator
A basic differential operator of order i is a mapping that maps any differentiable function to its ith derivative, or, in the case of several variables, to one of its partial derivatives of order i. It is commonly denoted
in the case of univariate functions, and
in the case of functions of n variables. The basic differential operators include the derivative of order 0, which is the identity mapping.
A linear differential operator (abbreviated, in this article, as linear operator or, simply, operator) is a linear combination of basic differential operators, with differentiable functions as coefficients. In the univariate case, a linear operator has thus the form
where a0(x), …, an(x) are differentiable functions, and the nonnegative integer n is the order of the operator (if an(x) is not the zero function).
Let L be a linear differential operator. The application of L to a function f is usually denoted Lf or Lf(X), if one needs to specify the variable (this must not be confused with a multiplication). A linear differential operator is a linear operator, since it maps sums to sums and the product by a scalar to the product by the same scalar.
As the sum of two linear operators is a linear operator, as well as the product (on the left) of a linear operator by a differentiable function, the linear differential operators form a vector space over the real numbers or the complex numbers (depending on the nature of the functions that are considered). They form also a free module over the ring of differentiable functions.
The language of operators allows a compact writing for differentiable equations: if
is a linear differential operator, then the equation
may be rewritten
There may be several variants to this notation; in particular the variable of differentiation may appear explicitly or not in y and the right-hand and of the equation, such as Ly(x) = b(x) or Ly = b.
The kernel of a linear differential operator is its kernel as a linear mapping, that is the vector space of the solutions of the (homogeneous) differential equation Ly = 0.
In the case of an ordinary differential operator of order n, Carathéodory's existence theorem implies that, under very mild conditions, the kernel of L is a vector space of dimension n, and that the solutions of the equation Ly(x) = b(x) have the form
where c1, …, cn are arbitrary numbers. Typically, the hypotheses of Carathéodory's theorem are satisfied in an interval I, if the functions b, a0, …, an are continuous in I, and there is a positive real number k such that |an(x)| > k for every x in I.
Homogeneous equation with constant coefficients
A homogeneous linear differential equation has constant coefficients if it has the form
where a1, …, an are (real or complex) numbers. In other words, it has constant coefficients if it is defined by a linear operator with constant coefficients.
The study of these differential equations with constant coefficients dates back to Leonhard Euler, who introduced the exponential function ex, which is the unique solution of the equation f′ = f such that f(0) = 1. It follows that the nth derivative of ecx is cnecx, and this allows solving homogeneous linear differential equations rather easily.
Let
be a homogeneous linear differential equation with constant coefficients (that is a0, …, an are real or complex numbers).
Searching solutions of this equation that have the form eαx is equivalent to searching the constants α such that
Factoring out eαx (which is never zero), shows that α must be a root of the characteristic polynomial
of the differential equation, which is the left-hand side of the characteristic equation
When these roots are all distinct, one has n distinct solutions that are not necessarily real, even if the coefficients of the equation are real. These solutions can be shown to be linearly independent, by considering the Vandermonde determinant of the values of these solutions at x = 0, …, n – 1. Together they form a basis of the vector space of solutions of the differential equation (that is, the kernel of the differential operator).
Example |
---|
has the characteristic equation This has zeros, i, −i, and 1 (multiplicity 2). The solution basis is thus A real basis of solution is thus |
In the case where the characteristic polynomial has only simple roots, the preceding provides a complete basis of the solutions vector space. In the case of multiple roots, more linearly independent solutions are needed for having a basis. These have the form
where k is a nonnegative integer, α is a root of the characteristic polynomial of multiplicity m, and k < m. For proving that these functions are solutions, one may remark that if α is a root of the characteristic polynomial of multiplicity m, the characteristic polynomial may be factored as P(t) (t − α)m. Thus, applying the differential operator of the equation is equivalent with applying first m times the operator , and then the operator that has P as characteristic polynomial. By the exponential shift theorem,
and thus one gets zero after k + 1 application of .
As, by the fundamental theorem of algebra, the sum of the multiplicities of the roots of a polynomial equals the degree of the polynomial, the number of above solutions equals the order of the differential equation, and these solutions form a base of the vector space of the solutions.
In the common case where the coefficients of the equation are real, it is generally more convenient to have a basis of the solutions consisting of real-valued functions. Such a basis may be obtained from the preceding basis by remarking that, if a + ib is a root of the characteristic polynomial, then a – ib is also a root, of the same multiplicity. Thus a real basis is obtained by using Euler's formula, and replacing xke(a+ib)x and xke(a−ib)x by xkeax cos(bx) and xkeax sin(bx).
Second-order case
A homogeneous linear differential equation of the second order may be written
and its characteristic polynomial is
If a and b are real, there are three cases for the solutions, depending on the discriminant D = a2 − 4b. In all three cases, the general solution depends on two arbitrary constants c1 and c2.
- If D > 0, the characteristic polynomial has two distinct real roots α, and β. In this case, the general solution is
- If D = 0, the characteristic polynomial has a double root −a/2, and the general solution is
- If D < 0, the characteristic polynomial has two complex conjugate roots α ± βi, and the general solution is
- which may be rewritten in real terms, using Euler's formula as
Finding the solution y(x) satisfying y(0) = d1 and y′(0) = d2, one equates the values of the above general solution at 0 and its derivative there to d1 and d2, respectively. This results in a linear system of two linear equations in the two unknowns c1 and c2. Solving this system gives the solution for a so-called Cauchy problem, in which the values at 0 for the solution of the DEQ and its derivative are specified.
Non-homogeneous equation with constant coefficients
A non-homogeneous equation of order n with constant coefficients may be written
where a1, …, an are real or complex numbers, f is a given function of x, and y is the unknown function (for sake of simplicity, "(x)" will be omitted in the following).
There are several methods for solving such an equation. The best method depends on the nature of the function f that makes the equation non-homogeneous. If f is a linear combination of exponential and sinusoidal functions, then the exponential response formula may be used. If, more generally, f is a linear combination of functions of the form xneax, xn cos(ax), and xn sin(ax), where n is a nonnegative integer, and a a constant (which need not be the same in each term), then the method of undetermined coefficients may be used. Still more general, the annihilator method applies when f satisfies a homogeneous linear differential equation, typically, a holonomic function.
The most general method is the variation of constants, which is presented here.
The general solution of the associated homogeneous equation
is
where (y1, …, yn) is a basis of the vector space of the solutions and u1, …, un are arbitrary constants. The method of variation of constants takes its name from the following idea. Instead of considering u1, …, un as constants, they can considered as unknown functions that have to be determined for making y a solution of the non-homogeneous equation. For this purpose, one adds the constraints
which imply (by product rule and induction)
for i = 1, …, n – 1, and
Replacing in the original equation y and its derivatives by these expressions, and using the fact that y1, …, yn are solutions of the original homogeneous equation, one gets
This equation and the above ones with 0 as left-hand side form a system of n linear equations in u′1, …, u′n whose coefficients are known functions (f, the , and their derivatives). This system can be solved by any method of linear algebra. The computation of antiderivatives gives u1, …, un, and then y = u1y1 + ⋯ + unyn.
As antiderivatives are defined up to the addition of a constant, one finds again that the general solution of the non-homogeneous equation is the sum of an arbitrary solution and the general solution of the associated homogeneous equation.
First-order linear equation
The general form of a linear ordinary differential equation of order 1, after dividing out the coefficient of y′(x), is:
If the equation is homogeneous, i.e. g(x) = 0, one may rewrite and integrate:
where k is an arbitrary constant of integration and is any antiderivative of f. Thus, the general solution of the homogeneous equation is
where c = ek is an arbitrary constant.
For the general non-homogeneous equation, one may multiply it by the reciprocal e−F of a solution of the homogeneous equation. This gives
As the product rule allows rewriting the equation as
Thus, the general solution is
where c is a constant of integration, and F is any antiderivative of f (changing of antiderivative amounts to change the constant of integration).
Example
Solving the equation
The associated homogeneous equation gives
that is
Dividing the original equation by one of these solutions gives
That is
and
For the initial condition
one gets the particular solution
Making Linear Equations from Non-Linear Equations
Sometimes a non-linear equation, which is not solvable like this, can be made linear, and more easily solvable, by applying a substitution.
Example
- Let's make the following substitution:
- Plugging in, we get
- We can then solve as a linear equation in v, using the step-by-step method above:
- Step 1: Find the integrating factor:
- Letting C=1 for convenience, we get as our integrating factor.
- Step 2: Multiply through
- Step 3: Recognize that the left hand is
- Step 4: Integrate both sides w.r.t.x.
- Step 5: Solve for v.
- Now that we have v, solve for y.
Higher order linear equations
A linear ordinary equation of order one with variable coefficients may be solved by quadrature, which means that the solutions may be expressed in terms of integrals. This is not the case for order at least two. This is the main result of Picard–Vessiot theory which was initiated by Émile Picard and Ernest Vessiot, and whose recent developments are called differential Galois theory.
The impossibility of solving by quadrature can be compared with the Abel–Ruffini theorem, which states that an algebraic equation of degree at least five cannot, in general, be solved by radicals. This analogy extends to the proof methods and motivates the denomination of differential Galois theory.
Similarly to the algebraic case, the theory allows deciding which equations may be solved by quadrature, and if possible solving them. However, for both theories, the necessary computations are extremely difficult, even with the most powerful computers.
Nevertheless, the case of order two with rational coefficients has been completely solved by Kovacic's algorithm.
Cauchy–Euler equation
Cauchy–Euler equations are examples of equations of any order, with variable coefficients, that can be solved explicitly. These are the equations of the form
where are constant coefficients.
Licensing
Content obtained and/or adapted from:
- Linear differential equation, Wikipedia under a CC BY-SA license
- First Order Linear 1, Wikibooks: Differential Equations under a CC BY-SA license