THE
PERFORMANCE OF THE EPSILON AND THETA CONVERGENCE ACCELERATION ALGORITHMS
A Paper
Submitted to the Graduate Faculty
of the
North Dakota State University
of Agriculture and Applied Science
By
Paul Scott Heidmann
In Partial Fulfillment of the Requirements
for the Degree of
MASTER OF SCIENCE
Major Department:
Mathematics
August 1992
Fargo, North Dakota
THE PERFORMANCE OF THE EPSILON AND THETA CONVERGENCE ACCELERATION ALGORITHMS
By
Paul Scott Heidmann
The Supervisory Committee certifies that this paper complies with North Dakota State University's regulations and meets the accepted standards for the degree of
MASTER OF SCIENCE
SUPERVISORY COMMITTEE:
___________________________________________________________________
Chair
___________________________________________________________________
___________________________________________________________________
___________________________________________________________________
Approved by Department Chair:
___________________ __________________________________
Date Signiture
Heidmann, Paul Scott, M.S., Department of Mathematics, College of Science and Mathematics, North Dakota State University, August 1992. The Performance of the Epsilon and Theta Convergence Acceleration Algorithms. Major Advisor: Associate Professor Davis Cope.
Motivations for transforms are first presented, along with the derivation of both the -algorithm and the -algorithm. Known properties are then presented, with a partial classification of the sets for which the transforms are known to have accelerative properties.
Empirical evidence is given illustrating the performance of the algorithms over a broad set of test series. The -algorithm was found to perform better than the -algorithm on all but a few of the test cases considered. No algorithm considered demonstrated any evidence of producing an increase in the exponential rate of convergence. No evidence of instability was seen among the algorithms considered.
Abstract................................................................................................................................ iii
Section 1: Introduction........................................................................................................... 1
Section 2: Transform Motivations and Properties.................................................................... 4
The Motivation of the e Transform................................................................. 4
Properties of the e Transform......................................................................... 7
The e Transform and the -Algorithm........................................................... 10
The Motivation of the q Transform............................................................... 13
Properties of the q Transform....................................................................... 14
Section 3: Empirical and Related Results.............................................................................. 16
Details of the Test Program.......................................................................... 17
Details of Test Procedures........................................................................... 17
Transform Singularities................................................................................. 18
Empirical Results......................................................................................... 23
Group 1.......................................................................................... 23
Group 2.......................................................................................... 77
Group 3.......................................................................................... 80
Group 4.......................................................................................... 95
Group 5........................................................................................ 101
Group 6........................................................................................ 115
Group 7.....................................................................................117
Section 4: Summary........................................................................................................... 126
Bibliography...................................................................................................................... 127
This paper is concerned with the evaluation of the performance of the -algorithm and the -algorithm when applied to sequences of partial sums. Transform definitions and properties are first stated, followed by empirical examples illustrating the comparative performance of the algorithms on a wide variety of test problems.
The two transforms covered in this paper are recursive and operate on the set of complex valued nonconstant sequences. For broad classes of sequences, these algorithms are known to have accelerative properties.
Both algorithms are actually more accurately described as a hierarchy of transforms. That is, the algorithms provide a succession of transforms, with the higher order transforms typically providing greater rates of acceleration.
The first of these algorithms, the -algorithm, is defined as follows for any nonconstant sequence [6, p. 225]:
.
The second algorithm considered in this paper is the -algorithm, defined as follows for any nonconstant sequence [6, p. 225]:
In most practical applications, a transform must have the properties of regularity and of convergence acceleration. This paper was concerned more with the latter than the former; however, both definitions will be required.
Definition: A transform T is said to be regular for a convergent sequence sn if
Definition: A transform T is said to accelerate the convergence of a sequence sn if
where s is the limit of the sequence sn.
Intuitively, a transform accelerates the convergence of a sequence if the convergence of the transformed sequence is of higher order than that of the original sequence. Equivalently, a transform accelerates the convergence of a sequence if the error term of the transformed sequence asymptotically goes to zero with a higher order than the error term of the original sequence.
Section 2 of this paper summarizes the known properties of the transforms. The transforms are first motivated, and then theorems concerning each are presented.
Section 3 presents two new lemmas and presents the empirical research done in connection with this paper.
Section 4 summarizes the results reached in Section 3.
The history of the e transform is not altogether clear. The general ideas which underlie this transform go at least as far back as Jacobi. A special case of the e transform known as the -process is usually attributed to Aitkin; however, this technique can be traced back to Kummer (1837) [8, p. 120]. Apparently, it was developed independently by several authors (although not always in its most general form): Delaunay, Samuelson, Shanks and Walton, Hartee, and Isakson [5, p. 5].
The e transform has been motivated by two distinct but similar sets of ideas. Both begin with the recognition that many sequences in applied mathematics and physics can be approximated by a limit together with exponential transients:
.
Here, it is assumed that none of the 's are zero or one (if a transient were one, it would become part of the limit s) and that the 's are distinct. The elemental idea is to develop a transform which will eliminate these exponential transients.
The first method of accomplishing this task was given by Shanks [5, p. 7-8] and is summarized below. Observe that any nonsingular sequence (definition follows) can be expressed as a kth order transient locally. Fix any n larger than k and represent as follows:
,
where the 's are the respective limits, called the "kth order local bases." Should one of the , then the sequence will be singular.[1] These bases will vary relatively little as long as is nearly kth order [5, p. 7]. This fact suggests that the sequence of 's (for some fixed k) might converge faster than the sequence itself. This sequence of kth order local bases is then defined as the -transform of the sequence [5, p. 8], [in fact, it is identical with as defined in the following discussion.]
The motivation for the e transform given by Wimp [8, p. 121-123] also begins with the recognition that many sequences from applied mathematics and physics behave as if they were a limit combined with exponential transients. As above, the e transform is then developed to eliminate these transients; the technique by which the transform is developed, however, differs significantly.
To begin (as above), assume the sequence has the form:
.
To simplify notation, let:
.
Since is a linear combination of k exponentials, there is a linear difference operator of order k that annihilates :
( not zero).
Now, rewrite as follows:
.
Next, substitute into the difference equation, and solve for .[2] Redefining constants gives:
.
This is a homogeneous system of equations with a nonzero solution . Such a solution will exist if and only if the following determinental equation is satisfied:
.
For simplicity, define:
.
Then, solving for s gives [8, p. 122]:
.
While the above equation will only be valid for sequences that can be represented as a combination of a limit and exponential transients, it can nevertheless be used to define a transform useful for a much larger class of sequences, which is written .
While these two derivations are quite different, they are indeed equivalent [5, p. 2-3,7-8].
At this point, several questions naturally arise. Questions concerning the regularity of the transform, the accelerative properties, and optimality are especially significant. This section presents known results that address these questions.
The first question that must be answered before the e transform can be used in any application is whether or not the transform will be regular for the sequences involved. For many applications, this question will not be difficult to answer, as the e transform can be shown to be regular on broad classes of sequences.
The following theorem (accredited originally to Lubkin) is stated by Shanks [5, p. 13]:
Theorem: If and both converge, then
.
Further regularity results are available for this transform and are discussed in the -algorithm section.
Closely related to the question of regularity is the question of exactness. That is, for what class of sequences will the e transform be exact (produce the limit of the sequence in a finite number of steps)? This question is answered in part by the following theorem, given by Shanks [5, p. 22]:
Theorem: Let f(z) be a rational function, which when reduced to its lowest terms has a q'th degree numerator and a p'th degree denominator; then, if
,
where is the first n terms of the Taylor expansion for f.
In fact, a closely related but stronger theorem has been given by Wimp [8, p. 125]:
Theorem: is defined for all and if and only if satisfies a homogeneous linear difference equation of order k with constant coefficients (and no equation of lower order) and contains no terms that are purely algebraic..
Finally, questions of optimality and accelerative properties must be addressed. Here, while fewer results are known, it is still possible to partially characterize sequences for which e is accelerative. A definition along with two helpful theorems are stated:
Definition: Let be the terms of a convergent series with partial sums and sum S. Let , and suppose that the sequences converge to , respectively. The convergence of the series is said to be logarithmic if , linear if , and of higher order if [6, p. 224].
The following two theorems are very useful for determining whether or not the condition is satisfied:
Theorem: If is of constant sign, then [3, p. 26].
Theorem: If with and if r is not -1, then [3, p. 26].
Wimp has shown [8, p. 104] that will accelerate all linearly convergent series. More interesting, however, are the results produced by Delahaye. Three basic results concerning the optimality of the transform (Aitkin's process) on the family of linearly convergent series are presented [4, p. 213]:
1. The transform is algebraically optimal; that is, it is not possible to find an algebraically simpler transform to accelerate the family of linearly convergent series.
2. It is impossible to accelerate the linear family with degree[3] greater than 1, the degree of acceleration achieved with the transform.
3. The linear family cannot be enlarged into another accelerable family. That is, there does not exist a family of sequences containing the linear family for which can be shown to have the above properties.
Delahaye then interprets these facts [4, p. 213]:
These results do not mean that the is the best algorithm of acceleration. They only mean that concerning the family Lin, it is impossible to find a simpler or a more efficient algorithm of acceleration. This does not contradict the existence of subfamilies of Lin accelerable with degree 2, nor does it contradict the existence of algorithms less simple but equally efficient on Lin.
When first developed, the e transform was merely a curiosity. It had little practical value, as calculating the large order determinants inherent to the transform was computationally prohibitive. This changed in 1956 when Wynn produced the following theorem [9, p. 91]:
Theorem: If
and
,
then
provided that none of the quantities becomes infinite.
With this algorithm, it is now possible to compute the e transform easily and efficiently. Study of the -algorithm has also facilitated much knowledge of the e transform.
One such result is due to Wimp. However, before stating this result, two definitions are required :
Definition: A sequence s is totally monotone (written ) if
.
Definition: A sequence s is totally oscillatory (written ) if
[8, p. 12].
Theorem: Let the sequence s have remainder terms r; then, if r is either totally monotone or totally oscillatory, then is regular for all even m [8, p. 139 and 141].
While the algorithm is regular for large classes of sequences, this is not the only concern. Computing the transform involves repeated divisions by what can be very small quantities. This seems to suggest that the transforms will be prone to numerical instability. However, the empirical results presented in this paper suggest that the -algorithm is far more stable than appears at first.
Theoretical analysis of the stability of the -algorithm has not yet been fully developed. However, Wynn has presented a paper that begins to deal with this issue [10]. The results presented suggest that the -algorithm can be unstable for certain monotonic sequences [10, p. 110-112] and unconditionally stable for certain oscillating sequences [10, p. 112-115]. These theoretical results can be seen intuitively, simply by examining the transform. Oscillating sequences have adjacent terms opposite in sign. This tends to make the denominator in the defining recurrence relation larger. Monotonic sequences have terms of like sign, which in turn tends to make the denominator smaller.
Finally, the -algorithm can be shown to be closely related to the Padé table. The Padé table for an analytic function f is a two dimensional table of rational approximations of f. The [L,M] entry of the Padé table is typically denoted [L/M] and is defined as follows [1, p. 5]:
,
where is a polynomial of degree at most L and is a polymonial of degree at most M. The coeficients of the polynomials are chosen so that , where A(z) is the Taylor expansion of f(z) [6, p. 5]. If we require that Q and P have no common factors and that Q(0) = 1, then [L/M] is unique [1, p. 8].
The -algorithm produces certain entries in the Padé table. The following theorem then is due to Wynn [10, p. 93]:
Theorem: If the -algorithm is applied to the initial values
,
then
.
The transform was developed by Brezinski as a means of accelerating the convergence of logarithmically converging sequences. Its motivation began with the observation that the algorithm fails to accelerate convergence on large classes of logarithmically converging sequences. However, rather than developing a completely new transform, Brezinski opted to modify the algorithm, adding an optimizing parameter.
Define . Then it is known that a set of sequences with logarithmic convergence having the property
may be accelerated by the algorithm modified with a "parameter of acceleration" [2, p. 727], specifically, with the following transform:
A necessary and sufficient condition for to converge faster than is to take
However, calculating all of the parameters can be prohibitive. It was at this point that Brezinski took his departure from what was then known. His idea was to replace with the following approximation [2, p. 729]:
The transform then results.
A property important in the application of any transform is regularity. Brezinski has characterized this property with the following theorem [2, p. 730]:
Theorem: For real valued sequences, a necessary and sufficient condition for
is that
As to the transform's accelerative properties, two theorems are given here. The first, and less general theorem, results from Germain-Bonne's theorem and is given by Smith and Ford [6, p. 225]:
Theorem: accelerates linear convergence.
The second theorem, due to Brezinski [2, p. 730], is much farther reaching:
Theorem: If is regular and if exists and is finite, then converges faster than .
Finally, the question of exactness is addressed by a theorem due to Cordellier and stated by Smith and Ford [6, p. 229]. It should be noted that this theorem characterizes only those sequences for which is exact in one term.
Theorem: for if and only if the sequence is one of the following three types:
1. where
2. where and neither is a positive
integer.
3. and d is not a positive integer.
This section describes the empirical research done in connection with this paper. Many series[5] were computed, and the transforms were applied to the sequences of their partial sums. The error terms for the sequences of partial sums, as well as the transformed sequences, were plotted. This allowed observations to be made concerning acceleration rates, stability, and unusual behavior (where it occurred).
Special software was constructed as arbitrary precision proved to be necessary. The high amount of precision was required for different reasons. Many of the transforms achieved upwards of 35 places of precision in only 30 terms. Other series were chosen to be very close to singular values for the transforms. Also, the high precision allowed elimination of numerical problems, so the properties of the transforms could be easily studied.
This section describes the program used to perform the calculations involved in constructing the many examples presented in this paper.
As many precision problems can arise during the calculation of these transforms, arbitrary precision was necessary for the observation of the behavior of the transforms considered in this paper.
The storage of arbitrary length numbers was accomplished using linked lists. Specifically, the numbers were stored as linked lists of bytes, each byte containing two digits. Also, since the transforms operate on sequences, it was necessary to store sequences of these arbitrarily long numbers. This was accomplished in a similar manner, using linked lists. Finally, subroutines were added to perform arithmetic functions and transforms on these arbitrarily long sequences of arbitrarily long numbers.
Following are several graphs illustrating the results of the empirical research done. All graphs are plots of the log (base 10) of the absolute value of the error. This configuration allows the approximate number of digits accurately calculated to be viewed at a glance.
The graphs contain two types of information. First, there is the size of the error term for each of the series and their transforms. Secondly, the rate of exponential acceleration can be seen in the slope of the corresponding line.
The actual value that any series under consideration converges to must be known. For those series which proved to have a reasonable rate of convergence, this limit was computed simply by summing the series out to several hundred terms before the transforms were computed. For those series with slower rates of convergence, Kummer acceleration was used to arrive at the limit. Several series had to be eliminated from the test set because their limit could not be computed.
Many of the series contained a term . Where it was found, this modifier was defined as follows:
.
The 's provided a nonrational factor that did not change the value of .[6]
In this section, two lemmas are given that demonstrate that it is possible for the transforms to have singularities. Two illustrations are also be given to demonstrate the behavior of the transforms near a singularity.
Lemma: Apply the -algorithm to the sequence . Then, is undefined if .
Proof: From the defining equations,
.
However, since is a sequence of partial sums,
Combining the above,
.
This will be undefined whenever .
Remarks:
1. If the sequence in question is a power series of the form , then will be undefined at .
2. This result may be generalized to an arbitrary sequence . The nth term of will be undefined when .
The following graph illustrates the above lemma. The numeric value used here was chosen to be very near to the value that would cause to be undefined. As can easily be seen, this sort of behavior cannot be ignored in practical applications of this algorithm.
Analogously, a similiar lemma can be proven for the algorithm:
Lemma: Apply the -algorithm to the sequence . Then, is undefined if .
Proof: From the defining equations,
.
This will be undefined when .
Remarks:
1. If the sequence of partial sums is a power series of the form , then is undefined at .
2. This result may be generalized to an arbitrary sequence . In this case, is undefined if where .
The following chart illustrates the above lemma. The numeric value used here was chosen to be very close to the value that would make undefined. As before, this behavior cannot be ignored if the algorithm is to be used in practice.
Other examples of how these singularities affect the behavior of the transforms have been computed, but they are not included here as the behavior is the same in all examples considered.
This set consists of those series for which the following order of performance can be clearly seen (from best to worst): . This is by far the largest group, suggesting that the -algorithm generally outperforms the -algorithm.
The anomoly in this graph is due to the fact that is nearly undefined (as discussed at the end of Section 3).
Note: Calculations made with 50 places of precision.
Note: Calculations made with 50 places of precision.
Note: Calculations made with 50 places of precision.
Note : The graphs of the transforms of this function have a significantly larger slope than the original series. This indicates a change in the exponential rate of convergence.
Group 2
This group contains series for which is exact. These series can all be shown to be a combination of a limit and two exponential transients. Notice that, for all of these series, the order of performance (from best to worst) is .
This group contains series for which eventually outperforms , but otherwise the results are unclear. All series in this group contain the nonrational modifier .
This group contains the series for which the following order of performance is observed (from best to worst): . For all but the first series in this group, the -algorithm will eventually be exact.
This sequence can be shown to be a combination of its limit and three exponential transients. The epsilon algorithm will then be exact by the sixth iteration. Notice here that the epsilon algorithm outperforms the theta algorithm. This is due to the fact that this is exactly the type of sequence that the e transform was designed to accelerate.
This and the next three series are included to demonstrate the effect of singularities on the transforms. All terms of both the -algorithm and the -algorithm are undefined at 1. The inclusion of the successive terms 0.9, 0.99, and 0.999 illustrates the continuing degradation of the algorithms as the singularities are approached.
This group of series bears the following peculiar attribute: only achieves any acceleration.
Note: For the above function, does not exist, but .
Note: For this series, , but does not exist.
This group consists of a single
series which none of the algorithms considered accelerated.
Note: For this series, , but .
For this group of series, the following order of performance is observed (from best to worst): .
Several conclusions have been reached in this paper; this section summarizes these results. The two lemmas presented in Section 3 are worthy of special mention here. In any practical application of the algorithms addressed in this paper, it is critical to be aware of any anomalies that might occur. The graphs associated with the lemmas illustrated the profound effect a singularity can have on a transform.
In the series considered, outperformed except for a few cases.
There did not appear to be a change in the exponential rate of convergence between and unless otherwise noted.
No evidence of numeric instability was observed in any of the transforms considered (only in the case where a transform was nearly undefined could any instability be seen).
[1] Baker, George A., Jr.
Essentials of Padé Approximants
Academic Press, New York, 1975
[2] Brezinski, Claude.
Acceleration de suites a convergence logarithmique.
C. R. Acad. Sc. Paris, 18 October 1971, p. 727-730.
[3] Clark, W. D., Gray, H. L., Adams, J. E.
A Note on the T-Transformation of Lubkin.
Journal of Research of the National Bureau of Standards, B. Mathematical Sciences, Vol 73B, No. 1, January - March 1969, p. 25-29.
[4] Delahaye, Jean-Paul.
Sequence Transformations.
Springer-Verlag, 1980.
[5] Shanks, Daniel.
Non-Linear Transformations of Divergent and Slowly Convergent Sequences.
Journal of Mathematics and Physics, Vol 34, 1955, p. 1-42.
[6] Smith, David A., Ford, William F.
Acceleration of Linear and Logarithmic Convergence.
Journal of Numerical Analysis, Vol 10, No. 2, April 1979, p. 223-240.
[7] Smith, David A., Ford, William F.
Numerical Comparisons of Nonlinear Convergence Accelerators.
Mathematics of Computation, Vol. 38, No. 158, p. 481-499.
[8] Wimp, Jet.
Sequence Transformations and Their Applications.
Academic Press, 1981, p. 120-147.
[9] Wynn, P.
On a Device for Computing the em(Sn) Transformation.
Mathematical Aids to Computation, Vol. 10, 1956, p. 91-96.
[10] Wynn, P.
On the Convergence and Stability of the Epsilon Algorithm.
Siam Journal of Numerical Analysis, Vol. 3, No. 1, 1966, p. 91-122.
[4]Brezinski calls this transform the transform and calls what we know as the transform the -transform. I have interchanged this notation to avoid confusion.
[6]Where the 's are the coefficients of the series in question.