On a Derivative-Free Variant of King's Family with Memory

The aim of this paper is to construct a method with memory according to King's family of methods without memory for nonlinear equations. It is proved that the proposed method possesses higher R-order of convergence using the same number of functional evaluations as King's family. Numerical experiments are given to illustrate the performance of the constructed scheme.


Introduction
Many problems arising in diverse disciplines of mathematical sciences can be described by a nonlinear equation of the following form (see, e.g., [1]): where : ⊆ R → R is a sufficiently differentiable function in a neighborhood of a simple zero of (1). If we are interested in approximating the root , we can do it by means of an iterative fixed-point method in the following form: provided that the starting point 0 is given. In this work, we are concerned with the fixed-point methods that generate sequences presumably convergent to the true solution of a given single smooth equation. These schemes can be divided into one-point and multipoint schemes. We remark that the one-point methods can possess high order by using higher derivatives of the function, which is expensive from a computational point of view. On the other hand, the multipoint methods are allowing the user not to waste information that had already been used. This approach provides the construction of efficient iterative root-finding methods [2].
In such circumstance, special attention is devoted to multipoint methods with memory that use already computed information to considerably increase convergence rate without additional computational costs. This would be the focus of this paper.
Traub in [2] proposed the following method with memory (TM): with the order of convergence 1 + √ 2.
The iterative methods with memory can improve the order of convergence of the without memory method without any additional functional calculations, and this results in a higher computational efficiency index. We remark that it is assumed that an initial approximation 0 close enough to the sought simple zero and 0 are given for iterative methods of type (3).
Recently, authors in [3] designed an approach to make derivative-free families with low complexity out of optimal methods. In fact, they conjectured that every time that one applies the approximation of the derivative ( ) ≈ [ , ], with = + ( ) , on an optimal order 2 , we will need ≥ for preserving the order of convergence.
and the conjecture of Cordero-Torregrosa, one may propose the following method (DKM): wherein In this work, we propose a two-step method with memory possessing a high efficiency index according to the wellknown family of King's methods (5).
Our inspiration and motivation for constructing a higherorder method are linked in a direct manner with the fundamental concept of numerical analysis that any numerical method should give as accurate as possible output results with minimal computational cost. To state the matter differently, it is necessary to pursue methods of higher computational efficiency.
For more background concerning this topic, one may refer to [5,6].
The paper is organized as follows. In Section 2, the aim of this paper is presented by contributing an iterative method with memory based on (5) for solving nonlinear equations. The proposed scheme is an extension over (4) and has a simple structure with an increased computational efficiency. In Section 3, we compare the theoretical results by applying the definition of efficiency index and further supports are furnished whereas numerical reports are stated. Some concluding remarks will be drawn in Section 4 to end the paper.
Toward this goal, one can increase the -order by considering the following substitution: Since the zero is not known, relation (9) cannot be used in its exact form and we must approximate it recursively. This builds a variant with memory for King's family by using where ≈ . Now if we consider 3 ( ) to be Newton's interpolation polynomial of third degree set through four available approximations , −1 , −1 , −1 at the end of each cycle, we can propose the following new method with memory: Note that, for example, we have the following formulation for the interpolating polynomial: Acceleration in convergence for (11) is based on the use of a variation of one free nonzero parameter in each iterative step. This parameter is calculated using information from the current and previous iteration(s) so that the developed method may be regarded as method with memory according to Traub's classification [2].
We are at the time to write about the theoretical aspects of our proposed solver (11). Theorem 1. Let the function ( ) be sufficiently differentiable in a neighborhood of its simple zero . If an initial approximation 0 is sufficiently close to , then, the -order of convergence of the two-step method (11) with memory is at least 4.23607.
Proof. Let { } be a sequence of approximations generated by an iterative method. The error relations with the selfaccelerating parameter = for (11) are in what follows: Using a symbolic computations, we attain that Substituting the value of − 2 ( ( ) 2 2 + 3 ) from (16) in (15), one may obtain Note that in general we know that the error equation should read +1 ∼ , where and are to be determined. Hence, one has ∼ −1 , and subsequently Thus, it is easy to obtain wherein is a constant. This results in with two solutions {−0.236068, 4.23607}. Clearly the value for = 4.23607 is acceptable and would be the convergence -order of method (11) with memory. The proof is complete.
The increase of -order is attained without any (new) additional function calculations so that the novel method with memory possesses a high computational efficiency index. This technique is an extension over scheme (5) to increase the -order from 4 to 4.23607.
The accelerating method (11) is new, simple, and useful, providing considerable improvement of convergence rate without any additional function evaluations in contrast to the optimal two-step methods without memory.
We also remark that an alternative form of our proposed method with memory could be deduced using backward finite difference formula at the beginning of the first substep and a minor modification in the accelerators; that is to say, we have the following alternative method with memory possessing 4.23607 as its -order (APM) as well: Theorem 2. Let the function ( ) be sufficiently differentiable in a neighborhood of its simple zero . If an initial approximation 0 is sufficiently close to , then, the -order of convergence of the two-step method (21) with memory is at least 4.23607.
Proof. The proof of this theorem is similar to Theorem 1. It is hence omitted.

Numerical Computations
Computational efficiency of different iterative methods with and without memory can be measured in a prosperous manner by applying the definition of efficiency index. For an iterative method with convergence ( -)order that requires functional evaluations, the efficiency index (also named computational efficiency) is calculated by Ostrowski-Traub's formula [2]: According to this, we find where SM is the quadratically convergent method of Steffensen without memory [7]. It should be remarked that Džunić in [8] designed an efficient one-step Steffensen-type method with memory possessing (1/2)(3 + √ 17) -order of convergence as follows: and Cordero et al. in [9] presented a two-step biparametric Steffensen-type iterative method with memory possessing seventh -order of convergence: Note that our main aim was to develop King's family in terms of efficiencies index and was not to achieve the highest possible efficiency index.
Although these methods possess higher computational efficiency indices than our proposed method (11), we exclude them from numerical comparisons since our method is not a Steffensen-type method and it is a Newton-type method with memory. For more refer to [10]. Now, we apply and compare the behavior of different methods for finding the simple zeros of some different nonlinear test functions in the programming package Mathematica [11] using multiple precision arithmetic to clearly reveal the high -order of PM and APM. We compare methods with the same number of functional evaluations per cycle.
We notice that, by applying any root solver with local convergence, a special attention must be paid to the choice of initial approximations. If initial values are sufficiently close to the sought roots, then the expected (theoretical) convergence speed is obtainable in practice; otherwise, the iterative methods show slower convergence, especially at the beginning of the iterative process.
In this section, the computational order of convergence (coc) has been computed by The calculated value coc estimates the theoretical order of convergence well when pathological behavior of the iterative method (i.e., slow convergence at the beginning of the implemented iterative method, oscillating behavior of approximations, etc.) does not exist.
Here the results of comparisons for the test functions are given by applying 1000 fixed floating point arithmetic using the stop termination | ( )| ≤ 10 −100 .
Example 3. We consider the following nonlinear test function in the interval = [1.5, 2.5]: using the initial approximation 0 = 1.7. The results are provided in Table 1.
It is evident from Tables 1 and 2 that approximations to the roots possess great accuracy when the proposed method with memory is applied. Results of the fourth iterate in Tables  1 and 2 are given only for demonstration of convergence speed of the tested methods and in most cases they are not required for practical problems at present. The Scientific World Journal 5 We also incorporated and applied the developed methods with memory (11) and (21) for different test examples and obtained results with the same behavior as above. Hence, we could mention that the theoretical results are upheld by numerical experiments and thus the new method is good with a high computational efficiency index.

Summary
In this paper, we have proposed a new two-step Steffensentype iterative method with memory for solving nonlinear scalar equations. Using one self-correcting parameter calculated by Newton interpolatory polynomial, the -order of convergence of the constructed method was increased from 4 to 4.23607 without any additional calculations.
The new method was compared in performance and computational efficiency with some existing methods by numerical examples. We have observed that the computational efficiency index of the presented method with memory is better than those of other existing two-step King-type methods.