One of the most classical problems in Numerical Analysis is the approximation of the zeros of a given function
In order to approximate these equations we can use iterative methods. The most used scheme is the second order Newton method,
This paper is devoted to the analysis of Chebyshev’s method that is a third order extension of Newton’s method. We present the geometric interpretation of the method and its global convergence. We introduce its extension to Banach spaces and review some applications where this high order method is a good alternative to Newton’s method.
Chebyshev is one of the famous mathematicians of the nineteenth century, creator of several mathematical schools in Russia: number theory, probability theory, function approximation theory, theory of mechanisms and machines, etc. He received primary education at home. His mother, taught him to read and write, while his cousin taught arithmetic and the French language, which will be very useful in his relation with Europe. He followed also completing his secondary education at home, but having as tutor in mathematics Prof. Pogorelsky, known in his day as the best teacher of elementary mathematics in Moscow. Prof. Brashman was who practically ran the university studies of Chebyshev ended in 1841. The department of physics and mathematics in which Chebyshev studied convened an award in the 1840-41 course. Chebyshev submitted a paper on the calculation of roots of equations, using the series expansion of the inverse function. The work, unpublished at the time, was rewarded with the silver medal. Chebyshev worked as a professor at the University of St. Petersburg for 35 years. He is recognized as the founder of the mathematical school in St. Petersburg whose echo and influence has reached our time in many branches of mathematics. This school was distinguished by the tendency to relate the theoretical problems of mathematics with problems in the art and nature.
The geometric interpretation of Newton’s method is well known, given an iterate
to the graph of
The following well-known theorem, giving enough conditions for the global convergence of Newton’s method, follows easily from its geometric interpretation.
Chebyshev’s method is obtained by quadratic interpolation of the inverse function of
that after the imposition of the super tangency conditions,
By calculating the intersection of this parabola with the
where
We refer to [1] for the geometric interpretation of other third order methods.
By using the geometric interpretation of Chebyshev’s method, we can obtain the following global convergence theorem.
First, we begin from a point on the left of
We would like to show that the intersection
Thus, it will be enough, if for
In this case, we will obtain a monotonic increasing sequence, bounded from above by
Inequality (3) is equivalent to
or
As
because
Thus,
and consequently the relation (4) holds.
Finally, if we begin from a point on the right of the root, we will obtain
We refer to [3] for the global convergence of other third order schemes and some comparisons.
In general, the method has not global convergence. We only are able in these cases to ask for local or semilocal convergence. Around the solutions we looking for regions of convergence.
Consider for instance the problem to find the zeros of a polynomial,
Let
The iterative rational function for Chebyshev’s method is given by
We recall the definition of conjugacy.
Let
An important feature of conjugation of rational maps is given by the following classical result.
Conjugacy plays a central rôle in understanding the behavior of classes of maps from the point of view of dynamical systems in the following sense. Suppose that one wishes to describe the quantitative as well as the qualitative behavior of the map
The result that follows, due to A. Cayley [8,21], has great historical importance. In an attempt to understand the dynamics of Newton’s method in the complex plane, Cayley investigated the dynamics of Newton’s method applied to polynomials of a particularly simple form. He realized that major difficulties would arise when attempting to extend the following result for quadratics to cubics and beyond. It is believed that this circumstance motivated further work of P. Fatou and G. Julia along these lines.
For Chebyshev’s method, also known as super–Newton method, the following result holds.
For cubic polynomials we have the following result.
It is possible to see that the one–parameter family
In the two pictures that follow we show the attraction basins of Newton’s and Chebyshev’s iterations applied to
In [4], we study the dynamics of a family of third-order iterative algorithms which includes Chebyshev’s iteration function, Halley’s iterative method, super-Halley’s iterative method and the c -iterative methods. Using results of conjugation of rational maps and the definition of the universal Julia set we find the conjugacy classes of these iterative methods explicitly.
To approximate a solution of the nonlinear equation
where
Next, we review some applications where Chebyshev’s method can be considered a good alternative to Newton’s method.
In this case,
Using Taylor expansions,
Thus,
equivalently,
Notice that we only need a
Moreover, we can obtain a very simple semilocal convergence:
We refer [12] and its references for a general convergence analysis of this type of methods.
Let us consider the equation
where
and
In table 1 we consider
Error max-norm, Iteration Cheby. 1 8.41 2 1.78 3 2.90 4 0.00
For quadratic equations we can find efficient higher order methods [2].
Approximating inverse operators is a very common task in several areas of interest, such as physics, chemistry, engineering, etc. In a general context, we can formulate the following problem: given
To approximate the inverse operator
Let
so that
If we observe Newton’s method,
it is clear that inverse operators are used, but if we take into account the definition of
then we can avoid the use of inverse operators for approximating
Indeed, to obtain the corresponding algorithm, we only need to compute
we have
In consequence, Newton’s method is now given by the following algorithm:
In addition, Newton’s method does not use inverse operators for approximating an inverse operator and the order of convergence is two.
If we now consider Chebyshev’s method,
where
so that we can also avoid the use of inverse operators for approximating
Let
so that
In consequence, we write Chebyshev’s method as
Observe that Chebyshev’s method does not use inverse operators for approximating an inverse operator and the order of convergence is three.
On the other hand, if we observe Newton’s and Chebyshev’s methods, two well-known iterative methods, we conclude that they can approximate inverse operators without using any inverse operator in their application. From this feature of both methods, we pay attention in [5] to the construction of iterative methods of any prefixed order of convergence. Moreover, the final formulation of the methods uses only potencies of matrices that are close to the identity. This fact is crucial to avoid any stability problem in the implementation of the methods. The best method of the family will depend on the particular problem to solve. In any case, we find iterative methods with better behavior than Newton’s method. Finally, we would like to emphasize that in the applications, where we are interested in the computation of an inverse operator, our methods use matrix-matrix multiplications with a computational cost similar to that of Gauss-type methods. But, if we are interested only in the application of the inverse operator, we will be able to implement our methods using only matrix-vector multiplications, so that we reduce considerably the computational cost.
If we consider the complex equation
If we extend Chebyshev’s method for the computation of the
and approximate
so that
Chebyshev’s method, as Newton’s method, cannot be used directly to approximate the principal
The following iteration, given by Denman and Beavers in [9],
is a stable variant of the Newton method for the matrix square root.
The instability of the simplified Newton iterations
and the iteration becomes stable in this form. However, it is useless since it involves the square root of
Following these ideas, Iannazo [17] proposes the following two stable versions of the Newton method for the matrix
and
Observe that in the second case, the matrix
Similarly, for Chebyshev’s method, we propose
and
Note that (17) is useless, since it involves the matrix
Observe that we need to calculate
As a consequence, the algorithm of Chebyshev’s method for solving
Observe that we only need to calculate
For Halley’s method, a similar strategy in order to avoid the use of inverses is not possible. Remember that the algorithm of Halley’s method,
always involves inverses (the quotient includes
After that, we analyze the convergence of Chebyshev’s method. For this, we suppose
To prove the convergence of algorithm (20), we consider a submultiplicative matrix norm || ⋅ || defined in
Next, taking into account Villareal’s formula,
where
where
Note that we can always obtain
In addition, we observe in table 2 some representative values of
Sequence 2 0.375, 0.1406... 3 0.4814..., 0.2222..., 0.0493..., 0.0109... 5 0.56, 0.296, 0.1059..., 0.0355..., 0.0080..., 0.0017..., 2.0736... × 10−4, 2.4883... × 10−5 7 0.5918..., 0.3294..., 0.1345..., 0.0512..., 0.0151..., 0.0041..., 0.0008..., 0.0001..., 2.6281... × 10−5, 3.6251... × 10−6, 2.9592... × 10−7, 2.4157... × 10−8
From Theorem 9, we also deduce that Chebyshev’s method has R-order of convergence at least three.
In [6], we present a family of high-order iterative methods including both Newton and Chebyshev methods. We find algorithms of the family with better numerical behavior than Newton and Halley methods. These two algorithms are basically the iterative methods proposed in the literature to solve this problem.
In this paper, we have reviewed some important properties of the classical Chebyshev method. We have point out the possibility to prove its global convergence from its geometric interpretation. We have mentioned the richer dynamics of this scheme in comparison with Newton’s method. Finally, we have presented several applications where this high order method is a good alternative to Newton’s method. These applications include the approximation of quadratic equations, the approximation of the inverse of a matrix or the pth root of a matrix.