Difference between revisions of "Power Series Solutions"

From Department of Mathematics at UTSA
Jump to navigation Jump to search
(Created page with "In mathematics, the '''power series method''' is used to seek a power series solution to certain differential equations. In general, such a solution assumes a po...")
 
 
(2 intermediate revisions by the same user not shown)
Line 1: Line 1:
In [[mathematics]], the '''power series method''' is used to seek a [[power series]] solution to certain [[differential equation]]s. In general, such a solution assumes a [[power series]] with unknown coefficients, then substitutes that solution into the differential equation to find a [[recurrence relation]] for the coefficients.
+
In mathematics, the '''power series method''' is used to seek a power series solution to certain differential equations. In general, such a solution assumes a power series with unknown coefficients, then substitutes that solution into the differential equation to find a recurrence relation for the coefficients.
  
 
== Method ==
 
== Method ==
Consider the second-order [[linear differential equation]]
+
Consider the second-order linear differential equation
 
: <math>a_2(z)f''(z)+a_1(z)f'(z)+a_0(z)f(z)=0.\;\!</math>
 
: <math>a_2(z)f''(z)+a_1(z)f'(z)+a_0(z)f(z)=0.\;\!</math>
 
Suppose ''a''<sub>2</sub> is nonzero for all ''z''. Then we can divide throughout to obtain
 
Suppose ''a''<sub>2</sub> is nonzero for all ''z''. Then we can divide throughout to obtain
 
: <math>f''+{a_1(z)\over a_2(z)}f'+{a_0(z)\over a_2(z)}f=0.</math>
 
: <math>f''+{a_1(z)\over a_2(z)}f'+{a_0(z)\over a_2(z)}f=0.</math>
Suppose further that ''a''<sub>1</sub>/''a''<sub>2</sub> and ''a''<sub>0</sub>/''a''<sub>2</sub> are [[analytic function]]s.
+
Suppose further that ''a''<sub>1</sub>/''a''<sub>2</sub> and ''a''<sub>0</sub>/''a''<sub>2</sub> are analytic functions.
  
 
The power series method calls for the construction of a power series solution
 
The power series method calls for the construction of a power series solution
 
:<math>f=\sum_{k=0}^\infty A_kz^k.</math>
 
:<math>f=\sum_{k=0}^\infty A_kz^k.</math>
  
If ''a''<sub>2</sub> is zero for some ''z'', then the [[Frobenius method]], a variation on this method, is suited to deal with so called "[[Regular singular point|singular points]]". The method works analogously for higher order equations as well as for systems.
+
If ''a''<sub>2</sub> is zero for some ''z'', then the Frobenius method, a variation on this method, is suited to deal with so called "singular points". The method works analogously for higher order equations as well as for systems.
  
 
== Example usage ==
 
== Example usage ==
Let us look at the [[Hermite differential equation]],
+
Let us look at the Hermite differential equation,
 
: <math>f''-2zf'+\lambda f=0;\;\lambda=1</math>
 
: <math>f''-2zf'+\lambda f=0;\;\lambda=1</math>
  
Line 44: Line 44:
 
: <math>(k+2)(k+1)A_{k+2}+(-2k+1)A_k=0\;\!</math>
 
: <math>(k+2)(k+1)A_{k+2}+(-2k+1)A_k=0\;\!</math>
  
We can rearrange this to get a [[recurrence relation]] for ''A''<sub>''k''+2</sub>.
+
We can rearrange this to get a recurrence relation for ''A''<sub>''k''+2</sub>.
  
 
: <math>(k+2)(k+1)A_{k+2}=-(-2k+1)A_k\;\!</math>
 
: <math>(k+2)(k+1)A_{k+2}=-(-2k+1)A_k\;\!</math>
Line 54: Line 54:
 
: <math>A_2 = {-1 \over (2)(1)}A_0={-1\over 2}A_0,\, A_3 = {1 \over (3)(2)} A_1={1\over 6}A_1</math>
 
: <math>A_2 = {-1 \over (2)(1)}A_0={-1\over 2}A_0,\, A_3 = {1 \over (3)(2)} A_1={1\over 6}A_1</math>
  
We can determine ''A''<sub>0</sub> and ''A''<sub>1</sub> if there are initial conditions, i.e. if we have an [[initial value]] problem.
+
We can determine ''A''<sub>0</sub> and ''A''<sub>1</sub> if there are initial conditions, i.e. if we have an initial value problem.
  
 
So we have
 
So we have
Line 81: Line 81:
 
: <math>f=A_0 \left(1+{-1\over 2}z^2+{-1 \over 8}z^4+{-7 \over 240}z^6+\cdots\right) + A_1\left(z+{1\over 6}z^3+{1 \over 24}z^5+{1 \over 112}z^7+\cdots\right)</math>
 
: <math>f=A_0 \left(1+{-1\over 2}z^2+{-1 \over 8}z^4+{-7 \over 240}z^6+\cdots\right) + A_1\left(z+{1\over 6}z^3+{1 \over 24}z^5+{1 \over 112}z^7+\cdots\right)</math>
  
which can be further simplified by the use of [[hypergeometric series]].
+
which can be further simplified by the use of hypergeometric series.
  
 
== A simpler way using Taylor series ==
 
== A simpler way using Taylor series ==
A much simpler way of solving this equation (and power series solution in general) using the [[Taylor series]] form of the expansion.
+
A much simpler way of solving this equation (and power series solution in general) using the Taylor series form of the expansion.
 
Here we assume the answer is of the form
 
Here we assume the answer is of the form
 
: <math>f=\sum_{k=0}^\infty {A_kz^k\over {k!}}</math>
 
: <math>f=\sum_{k=0}^\infty {A_kz^k\over {k!}}</math>
Line 105: Line 105:
  
 
==Nonlinear equations==
 
==Nonlinear equations==
The power series method can be applied to certain [[nonlinear]] differential equations, though with less flexibility. A very large class of nonlinear equations can be solved analytically by using the [[Parker–Sochacki method]]. Since the Parker–Sochacki method involves an expansion of the original system of ordinary differential equations through auxiliary equations, it is not simply referred to as the power series method. The Parker–Sochacki method is done before the power series method to make the power series method possible on many nonlinear problems. An ODE problem can be expanded with the auxiliary variables which make the power series method trivial for an equivalent, larger system. Expanding the ODE problem with auxiliary variables produces the same coefficients (since the power series for a function is unique) at the cost of also calculating the coefficients of auxiliary equations. Many times, without using auxiliary variables, there is no known way to get the power series for the solution to a system, hence the power series method alone is difficult to apply to most nonlinear equations.
+
The power series method can be applied to certain nonlinear differential equations, though with less flexibility. A very large class of nonlinear equations can be solved analytically by using the Parker–Sochacki method. Since the Parker–Sochacki method involves an expansion of the original system of ordinary differential equations through auxiliary equations, it is not simply referred to as the power series method. The Parker–Sochacki method is done before the power series method to make the power series method possible on many nonlinear problems. An ODE problem can be expanded with the auxiliary variables which make the power series method trivial for an equivalent, larger system. Expanding the ODE problem with auxiliary variables produces the same coefficients (since the power series for a function is unique) at the cost of also calculating the coefficients of auxiliary equations. Many times, without using auxiliary variables, there is no known way to get the power series for the solution to a system, hence the power series method alone is difficult to apply to most nonlinear equations.
  
The power series method will give solutions only to [[initial value problem]]s (opposed to [[boundary value problem]]s), this is not an issue when dealing with linear equations since the solution may turn up multiple linearly independent solutions which may be combined (by [[Superposition principle|superposition]]) to solve boundary value problems as well. A further restriction is that the series coefficients will be specified by a nonlinear recurrence (the nonlinearities are inherited from the differential equation).
+
The power series method will give solutions only to initial value problems (opposed to boundary value problems), this is not an issue when dealing with linear equations since the solution may turn up multiple linearly independent solutions which may be combined (by superposition) to solve boundary value problems as well. A further restriction is that the series coefficients will be specified by a nonlinear recurrence (the nonlinearities are inherited from the differential equation).
  
 
In order for the solution method to work, as in linear equations, it is necessary to express every term in the nonlinear equation as a power series so that all of the terms may be combined into one power series.
 
In order for the solution method to work, as in linear equations, it is necessary to express every term in the nonlinear equation as a power series so that all of the terms may be combined into one power series.

Latest revision as of 10:28, 29 October 2021

In mathematics, the power series method is used to seek a power series solution to certain differential equations. In general, such a solution assumes a power series with unknown coefficients, then substitutes that solution into the differential equation to find a recurrence relation for the coefficients.

Method

Consider the second-order linear differential equation

Suppose a2 is nonzero for all z. Then we can divide throughout to obtain

Suppose further that a1/a2 and a0/a2 are analytic functions.

The power series method calls for the construction of a power series solution

If a2 is zero for some z, then the Frobenius method, a variation on this method, is suited to deal with so called "singular points". The method works analogously for higher order equations as well as for systems.

Example usage

Let us look at the Hermite differential equation,

We can try to construct a series solution

Substituting these in the differential equation

Making a shift on the first sum

If this series is a solution, then all these coefficients must be zero, so for both k=0 and k>0:

We can rearrange this to get a recurrence relation for Ak+2.

Now, we have

We can determine A0 and A1 if there are initial conditions, i.e. if we have an initial value problem.

So we have

and the series solution is

which we can break up into the sum of two linearly independent series solutions:

which can be further simplified by the use of hypergeometric series.

A simpler way using Taylor series

A much simpler way of solving this equation (and power series solution in general) using the Taylor series form of the expansion. Here we assume the answer is of the form

If we do this, the general rule for obtaining the recurrence relationship for the coefficients is

and

In this case we can solve the Hermite equation in fewer steps:

becomes

or

in the series

Nonlinear equations

The power series method can be applied to certain nonlinear differential equations, though with less flexibility. A very large class of nonlinear equations can be solved analytically by using the Parker–Sochacki method. Since the Parker–Sochacki method involves an expansion of the original system of ordinary differential equations through auxiliary equations, it is not simply referred to as the power series method. The Parker–Sochacki method is done before the power series method to make the power series method possible on many nonlinear problems. An ODE problem can be expanded with the auxiliary variables which make the power series method trivial for an equivalent, larger system. Expanding the ODE problem with auxiliary variables produces the same coefficients (since the power series for a function is unique) at the cost of also calculating the coefficients of auxiliary equations. Many times, without using auxiliary variables, there is no known way to get the power series for the solution to a system, hence the power series method alone is difficult to apply to most nonlinear equations.

The power series method will give solutions only to initial value problems (opposed to boundary value problems), this is not an issue when dealing with linear equations since the solution may turn up multiple linearly independent solutions which may be combined (by superposition) to solve boundary value problems as well. A further restriction is that the series coefficients will be specified by a nonlinear recurrence (the nonlinearities are inherited from the differential equation).

In order for the solution method to work, as in linear equations, it is necessary to express every term in the nonlinear equation as a power series so that all of the terms may be combined into one power series.

As an example, consider the initial value problem

which describes a solution to capillary-driven flow in a groove. There are two nonlinearities: the first and second terms involve products. The initial values are given at , which hints that the power series must be set up as:

since in this way

which makes the initial values very easy to evaluate. It is necessary to rewrite the equation slightly in light of the definition of the power series,

so that the third term contains the same form that shows in the power series.

The last consideration is what to do with the products; substituting the power series in would result in products of power series when it's necessary that each term be its own power series. This is where the Cauchy product

is useful; substituting the power series into the differential equation and applying this identity leads to an equation where every term is a power series. After much rearrangement, the recurrence

is obtained, specifying exact values of the series coefficients. From the initial values, and , thereafter the above recurrence is used. For example, the next few coefficients:

A limitation of the power series solution shows itself in this example. A numeric solution of the problem shows that the function is smooth and always decreasing to the left of , and zero to the right. At , a slope discontinuity exists, a feature which the power series is incapable of rendering, for this reason the series solution continues decreasing to the right of instead of suddenly becoming zero.

Licensing

Content obtained and/or adapted from: