Skip to main content
The Scientific World Journal logoLink to The Scientific World Journal
. 2015 Jan 28;2015:934260. doi: 10.1155/2015/934260

A General Class of Derivative Free Optimal Root Finding Methods Based on Rational Interpolation

Fiza Zafar 1,*, Nusrat Yasmin 1, Saima Akram 1, Moin-ud-Din Junjua 1
PMCID: PMC4333564  PMID: 25722990

Abstract

We construct a new general class of derivative free n-point iterative methods of optimal order of convergence 2n−1 using rational interpolant. The special cases of this class are obtained. These methods do not need Newton's iterate in the first step of their iterative schemes. Numerical computations are presented to show that the new methods are efficient and can be seen as better alternates.

1. Introduction

The problem of root finding has been addressed extensively in the last few decades. In 1685, the first scheme to find the roots of nonlinear equations was published by John Wallis. Its simplified description was published in 1690 by Joseph Raphson and was called Newton-Raphson method. In 1740, Thomas Simpson was the first to introduce Newton's method as an iterative method for solving nonlinear equations. The method is quadratically convergent but it may not converge to real root if the initial guess does not lie in the vicinity of root and f′(x) is zero in the neighborhood of the real root. This method is a without-memory method. Later, many derivative free methods were defined, for example, Secant's and Steffensen's methods. However, most of the derivative free iterative methods are with-memory methods; that is, they require old and new information to calculate the next approximation. Inspite of the drawbacks of Newton's method, many multipoint methods for finding simple root of nonlinear equations have been developed in the recent past using Newton's method as the first step. However, many higher order convergent derivative free iterative methods have also appeared most recently by taking Steffensen type methods at the first step. A large number of optimal higher order convergent iterative methods have been investigated recently up to order sixteen [13]. These methods used different interpolating techniques for approximating the first derivative.

In the era of 1960–1965, many authors used rational function approximation for finding the root of nonlinear equation, for example, Tornheim [4], Jarratt, and Nudds [5].

In 1967, Jarratt [6] effectively used rational interpolation of the form

y=xabx2+cx+d, (1)

to approximate f(x) for constructing a with-memory scheme involving first derivative. The order of the scheme was 2.732 and its efficiency was 0.2182.

In 1987, Cuyt and Wuytack [7] described a with-memory iterative method involving first derivative based on rational interpolation and they also discussed two special cases of their scheme having order 1.84 with efficiency 0.1324 and order 2.41 with efficiency 0.1910.

In 1990, Field [8] used rational function to approximate the root of a nonlinear equation as follows:

xi+1=xi+di, (2)

where d i, the correction in each iterate, is the root of the numerator of the Pade approximant to the Taylor series:

fx=j=0fjxij!xxij. (3)

He proved that x i converges to the root ξ with order m + n + 1, where m and n are degrees of the denominator and numerator of the Pade approximant.

In 1974, Kung and Traub [9] conjectured that a multipoint iterative scheme without memory for finding simple root of nonlinear equations requiring n functional evaluations for one complete cycle can have maximum order of convergence 2n−1 with maximal efficiency index 2(n−1)/n. Multipoint methods with this property are usually called optimal methods. Several researchers [13, 10] developed optimal multipoint iterative methods based on this hypothesis.

In 2011, Soleymani and Sharifi [10] developed a four-step without-memory fourteenth order convergent iterative method involving first derivative having an efficiency index of 1.6952. The first derivative at the fourth step is approximated using rational interpolation as follows:

wnϕ8xn,yn,znxn+1=wn1+b4wnxn2fxn+b3wnxn2+b4wnxnfwn, (4)

where ϕ 8 is an optimal eighth order convergent method and the rational interpolant is given as:

mt=b+b2tx+b3tx21+b4tx. (5)

In 2012, Soleymani et al. [3] developed a three-step derivative free eighth order method using rational interpolation as follows:

zn=ϕ4xn,yn,wn,xn+1=zn1+a3znxn2a1a0a3+2a2znxn+a2a3znxn2fzn, (6)

where ϕ 4 is any two-step fourth order convergent derivative free iterative method. They used the same rational interpolant as given in (5). The constants are determined using the interpolating conditions.

In 2012, Soleymani et al. [2] added his contribution by developing a sixteenth order four-point scheme using Pade approximation. The scheme required four evaluations of functions and one evaluation of first derivative and achieved optimal order of sixteen and and an efficiency index 1.741. The scheme was of the form

wn=ϕ8xn,yn,zn,xn+1=wn1+b5wnxn2fwn×fxn+2b3wnxn+3b4+b3b5wnxn2+2b4b5wnxn31, (7)

where ϕ 8 is eighth order optimal method. They used the rational interpolant of the following form:

pt=b1+b2tx+b3tx2+b4tx31+b5tx. (8)

Recently in 2013, Sharma et al. [1] developed a three-step eighth order method and its extension to four-step sixteenth order method using rational interpolation. In the scheme, the first three steps are any arbitrary eight order convergent method. The fourth step is the root of the numerator of the method, which is given as

tn=ϕ8xn,zn,wn,xn+1=xnP1fzn,wn+Q1fxn,wn+Rftn,wnP1L+Q1M+RNfxn, (9)

where,

L=fwnfxn,znfznfxn,wnwnzn,M=fwnfxnfxnfxn,wnwnxn,N=fwnfxn,tnftnfxn,wnwntn,P1=xntnfxnftn,Q1=tnznftnfzn,R=znxnfznfxn, (10)

and rational polynomial of the following form:

p4x=xxi+λμxxi3+νxxi2+ξxxi+η. (11)

The efficiency index of the above sixteenth order method is 1.741. The method involves one derivative evaluation.

In this paper, we present a general class of derivative free n-point iterative method which satisfies Kung and Tarub's Hypothesis [9]. Proposed schemes require n functional evaluations to acquire the convergence order 2n−1 and efficiency index can have 2(n−1)/n. The contents of the paper are summarized as follows. In Section 2, we present a general class of n-point iterative scheme and its special cases with second, fourth, eighth, and sixteenth order convergence. Section 3 consists of the convergence analysis of the iterative methods discussed in Section 2. In the last section of the paper, we give concluding remarks and some numerical results to show the effectiveness of the proposed methods.

2. Higher Order Derivative Free Optimal Methods

In this section, we give a general class of n-point iterative method involving n functional evaluations having order of convergence 2n−1. Thus, the scheme is optimal in the sense of the conjecture of Kung and Traub [9].

Consider a rational polynomial of degree n − 1 as follows:

rn1t=p1tqn2t, (12)

where,

p1t=a0+a1tx,qn2t=1+b1tx++bn2txn2,n2,q01. (13)

We approximate f(x) by rational function given by (12) to construct a general class of n-point iterative scheme. Then, the root of nonlinear equation f(x) = 0 is the root of the numerator of the rational interpolant of degree n − 1 for the n-point method. The unknowns a 0, a 1, b 1,…, b n−1 are determined by the following interpolating conditions:

rn1x=fx,rn1wk=fwk,k=1,,n1,n2. (14)

Then, the general n-point iterative method is given by

w1=x+βfx,wn=xa0a1,n2. (15)

Now, we are going to derive its special cases. For n = 2 in (14)-(15),

r1t=a0+a1tx. (16)

We find a 0 and a 1 such that

r1x=fx,r1w1=fw1. (17)

So, the two-point iterative scheme becomes

w1=x+βfx,w2=xfxfw1,x. (18)

The iterative scheme (18) is the same as given by Steffensen [11] for β = 1; thus, is a particular case of our scheme given by (14)-(15).

For n = 3, we have

r2t=a0+a1tx1+b1tx. (19)

We find a 0, a 1, and b 1 using the following conditions:

r2x=fx,r2w1=fw1,r2w2=fw2. (20)

By using conditions (20), we have

a0=fx,a1=fw2,x+b1fw2,b1=fw1,xfw2,xfw2fw1. (21)

Now, using (21), we have the following three-point iterative scheme:

w1=x+βfx,w2=xfxfw1,x,w3=xfxfw2fw1fw2fw1,xfw1fw2,x. (22)

For n = 4, we have the following rational interpolant:

r3t=a0+a1tx1+b1tx+b2tx2, (23)

such that

r3x=fx,r3w1=fw1,r3w2=fw2,r3w3=fw3. (24)

The conditions (24) are used to determine the unknowns a 0, a 1, b 1, and b 2. Thus, we attain a four-point iterative method as follows:

w1=x+βfx,w2=xfxfw1,x,w3=xfxfw2fw1fw2fw1,xfw1fw2,x,w4=xfxh1+h2+h3h1fw1,x+h2fw2,x+h3fw3,x, (25)

where,

h1=fw2fw3w3w2,h2=fw1fw3w1w3,h3=fw1fw2w2w1. (26)

For n = 5, we have the following five-point iterative scheme:

w1x+βfx,w2=xfxfw1,x,w3=xfxfw2fw1fw2fw1,xfw1fw2,x,w4=xfxh1+h2+h3h1fw1,x+h2fw2,x+h3fw3,x,w5=xa0a1, (27)

where,

r4t=a0+a1tx1+b1tx+b2tx2+b3tx3, (28)

such that

r4x=fx,r4w1=fw1,r4w2=fw2,r4w3=fw3,r4w4=fw4. (29)

The interpolating conditions (29) yield

a0=fx,a1=m1fw1,x+m2fw2,x+m3fw3,x+m4fw4,xm1+m2+m3+m41, (30)

where,

m1=fw2fw3fw4×w3xw4xw4w3+w2xw4xw4w2w2xw3xh1fw4m2=fw1fw3fw4×w3xw4xw4w3w1xw4xw4w1w1xw3xh2fw4m3=fw1fw2fw4×w2xw4xw4w2+w1xw4xw4w1w1xw2xh3fw4m4=fw1fw2fw3×w2xw3xw3w2w1xw3xw3w1+w1xw2xh3fw3, (31)

and h 1, h 2, h 3 are given as in (26). Hence, we obtain the following iterative method:

w1=x+βfx,w2=xfxfw1,x,w3=xfxfw2fw1fw2fw1,xfw1fw2,x,w4=xfxh1+h2+h3h1fw1,x+h2fw2,x+h3fw3,x,w5=xfxm1+m2+m3+m4×m1fw1,x+m2fw2,x+m3fw3,x+m4fw4,x1. (32)

We, now, give the convergence analysis of the proposed iterative methods (18), (22), (25), and (32).

3. Convergence Analysis

Theorem 1 . —

Let us consider ωI as the simple root of sufficiently differentiable function f : IRR in the neighborhood of the root for interval I. If x is sufficiently close to ω, then, for every βR∖{−1}, the iterative methods defined by (18) and (22) are second and fourth order convergent, respectively, with the error equations given by

en+1=c21+βen2+Oen3,en+1=c23c2c3+2c23β+c23β22c2c3βc2c3β2en4+Oen5, (33)

respectively, where,

ck=1k!fkωfω,k=2,3, (34)

Proof —

Let x = ω + e n, where ω is the root of f and e n is the error at nth step. Now, using Taylor expansion of f(x) about the root ω, we have

fx=fωen+c2en2+c3en3+c4en411111111++c8en8+Oen9, (35)

where c k is defined by (34). Taylor's expansions for w 1 and f(w 1) are

w1=en+ω+βen+c2en2+c3en3+c4en4+c5en5+Oen6, (36)
fw1=fω1+βen+3βc2+c2+c2β2en21111111111+2c22β+2c22β2+c3+4βc31111111111111+3c3β2+c3β3en31111111111+5c2βc3+8c2β2c3+3c3β31111111111111+c4+5βc4+6β2c4+4c4β31111111111111+c4β4+c23β2en41111111111+c5+6βc5+10c5β2+10c5β3+5c5β4111111111111+c5β5+6c2βc4+14c2c4β2+12c4c2β3111111111111+4c4c2β4+5c22c3β2+3c32β+6c32β2111111111111+3c3β3c22+3c32β3en5+Oen6. (37)

Using (35), (36), and (37), we have

w2=ω+c21+βen2+2c222c22β+2c3+3βc3+c3β2c22β2en3+4c23+5c23β7c2c310c2βc37c2β2c3+3c23β22c3β3c2+3c4+6βc4+4c4β2+c4β3+c23β3en4+Oen5, (38)

which shows that the method (18) is quadratically convergent for all βR∖{−1}. Again using Taylor expansion of f(w 2), we have

fw2=fωc21+βen21111111111111+2c222c22β+2c31111111111111111+3βc3+c3β2c22β2en31111111111111+5c23+7c23β7c2c310c2βc311111111111111117c2β2c3+4c23β211111111111111112c3β3c2+3c4+6βc41111111111111111+4c4β2+c4β3+c23β3en4+Oen5. (39)

Now, using (35)–(39), we see that the order of convergence of the method (22) is four and the error equation is given by

en+1=c23+2c23βc2c32c2c3βc2c3β2+c23β2en4+Oen5. (40)

Theorem 2 . —

Let us consider ωI as the simple root of sufficiently differentiable function f : IRR in the neighborhood of the root for interval I. If x is sufficiently close to ω, then for all βR∖{−1}, the iterative methods defined by (25) and (32) are eighth and sixteenth order convergent, respectively, with the error equations given by

en+1=c22c25β43c3c23β4+4c25β3+c4c22β4+2c32c2β412c3c23β3+6c25β2c4c3β4+4c4c22β3+8c32c2β318c3c23β2+4c25β4c4c3β3+6c4c22β2+12c32c2β212c3c23β+c256c4c3β2+4c4c22β+8βc2c323c23c34βc3c4+c22c4+2c2c32c3c4en8+Oen9,en+1=4c25c3c5c32c4c5+2c2c32c42+2c2c33c54c23c3c425c23c32c5c24c4c5+18c24c32c413c26c3c49c22c33c4+2c25c42+c34c42c2c35c27c57c29c3+3c28c4+11c23c3421c25c33+18c27c32+2c22c3c4c5+c211c24β8++4c25c3c5c32c4c5+2c2c32c42+2c2c33c54c23c3c425c23c32c5c24c4c5+18c24c32c413c26c3c49c22c33c4+2c25c42+c34c42c2c35c27c57c29c3+3c28c4+11c23c3421c25c33+18c27c32+2c22c3c4c5+c211c24en16+Oen17, (41)

where,

ck=1k!fkωfω,k=2,3, (42)

Proof —

Let x = ω + e n, where ω is the root of f and e n is the error in the approximation at nth iteration. We will use (35), (37), (39), and (40) up to O(e n 30) in this result and set

w3=ω+c23+2c23βc2c32c2c3βc2c3β2+c23β2en4+4c3β3c222c24β3c32β3c4β3c24c2β2c4+14c22β2c36c24β24c32β2+18c22βc35c2βc48c24β5c32β4c242c2c4+8c22c32c32en5++Oen30. (43)

Again, using Taylor expansion of f(w 3), we have

fw3=fωc23+2c23βc2c32c2c3β111111111111111c2c3β2+c23β2en411111111111111+4c3β3c222c24β3c32β3c4β3c2111111111111111114c2β2c4+14c22β2c36c24β24c32β211111111111111111+18c22βc35c2βc48c24β5c32β111111111111111114c242c2c4+8c22c32c32en511111111111111++Oen30. (44)

Now, using (35)–(39), (43), and (44), we see that the method (25) has eighth order convergence with the error equation

en+1=c22c25β43c3c23β4+4c25β3+c4c22β4+2c32c2β412c3c23β3+6c25β2c4c3β4+4c4c22β3+8c32c2β318c3c23β2+4c25β4c4c3β3+6c4c22β2+12c32c2β212c3c23β+c256c4c3β2+4c4c22β+8βc2c323c23c34βc3c4+c22c4+2c2c32c3c4en8+Oen9. (45)

To find the error equation of (32), we use (45) up to O(e n 30) and set

w4=c22c25β43c3c23β4+4c25β3+c4c22β4+2c32c2β412c3c23β3+6c25β2c4c3β4+4c4c22β3+8c32c2β318c3c23β2+4c25β4c4c3β3+6c4c22β2+12c32c2β212c3c23β+c256c4c3β2+4c4c22β+8βc2c323c23c34βc3c4+c22c4+2c2c32c3c4en8++Oen30. (46)

Taylor expansion of f(w 4) is

fw4=fωc27β4+c24c4+c27+c4β4c24c22c4c3c4c3β4c22+8c23c32β3+2β4c32c23+8c23βc32+12c23β2c3218c25c3β212c25c3β3β4c25c3+6c24c4β2+4c24c4β3+4c24c4β12c25c3β3+6c27β2+4c27β3+4c27β3c25c3+2c23c326c3β2c4c224c22c4βc34c4c22c3β3en8++Oen30. (47)

Hence, using (35)–(39), (43), (44), (46), and (47), we see that iterative method (32) is sixteenth order convergent with the error equation given by

en+1=4c25c3c5c32c4c5+2c2c32c42+2c2c33c5111111114c23c3c425c23c32c5c24c4c511111111+18c24c32c413c26c3c49c22c33c4+2c25c4211111111+c34c42c2c35c27c57c29c311111111+3c28c4+11c23c3421c25c33+18c27c3211111111+2c22c3c4c5+c211c24β81111111++4c25c3c5c32c4c5+2c2c32c42+2c2c33c5111111111111114c23c3c425c23c32c5c24c4c5+18c24c32c41111111111111113c26c3c49c22c33c4+2c25c42+c34c42c2c3511111111111111c27c57c29c3+3c28c4+11c23c341111111111111121c25c33+18c27c32+2c22c3c4c511111111111111+c211c24en16+Oen17. (48)

Remark 3 . —

From Theorems 1 and 2, it can be seen that the iterative schemes (18), (22), (25), and (32) are second, fourth, eighth, and sixteenth order convergent requiring two, three, four, and five functional evaluations, respectively. Hence, the proposed iterative schemes (18), (22), (25), and (32) are optimal in the sense of the hypothesis of Kung and Traub [9] with the efficiency indices 1.414, 1.587, 1.681, 1.741. Also, it is clear that (14)-(15) is a general n-point scheme with optimal order of convergence 2n−1. The efficiency index of this scheme is 2(n−1)/n.

4. Numerical Results

In this section, we present some test functions to demonstrate the performance of the newly developed sixteenth order scheme (32) (FNMS-16). For the sake of comparison, we consider the existing higher order convergent methods based on rational interpolation. We consider the fourteenth order method of Soleymani and Sharifi (4) (SS-14), the sixteenth order method of Soleymani et al.(7) (SSS-16), and the sixteenth order method of Sharma et al.(9) (SGG-16). All the computations for the above-mentioned methods are performed using Maple 16 with 4000 decimal digits precision. The test functions given in Table 1 are taken from [1, 2, 10]. We used almost all types of nonlinear functions, polynomials, and transcendental functions to test the new methods. Table 2 shows that the newly developed sixteenth order methods are comparable with the existing methods of this domain in terms of significant digits and number of function evaluations per iteration. In many examples, the newly developed methods perform better than the existing methods. It can also be seen from the tables that, for the choice of initial guess, near to the exact root or far from the exact root, the performance of the new methods is better.

Table 1.

Test functions and exact roots.

Numerical example Exact roots
f1(x)=x4+8sin(π/(x2+2))+x3/(x4+1)-6+8/17 ω 1 = −2
f 2(x) = sin⁡⁡(x) − x/100 ω 2 = 0
f 3(x) = (1/3)x 4x 2 − (1/3)x + 1 ω 3 = 1
f 4(x) = e sin⁡⁡(x) − 1 − x/5 ω 4 = 0
f 5(x) = xe x2 − (sin⁡⁡(x))2 + 3cos⁡(x) + 5 ω 5 ≈ −1.207647827130919
f 6(x) = e x + cos⁡(x) ω 6 ≈ 1.746139530408013
f 7(x) = 10xe x2 − 1 ω 7 ≈ 1.679630610428450
f 8(x) = x 3 + 4x 2 − 15 ω 8 ≈ 1.631980805566063

Table 2.

Comparison of various iterative methods.

f(x), x 0 (SS-14) (SGG-16) (SSS-16) (FNMS-16)
f 1, x 0 = −1.2
 |f 1(x 1)| .3e − 13 .1e − 15 .1e − 13 .1e − 15
 |f 1(x 2)| .1e − 181 .6e − 248 .1e − 211 .5e − 245
 |f 1(x 3)| .6e − 2538 .6e − 3751 .7e − 3379 .6e − 3916
f 1, x 0 = −3
 |f 1(x 1)| .1e − 5 .3e − 6 .3e − 7 .2e − 8
 |f 1(x 2)| .4e − 75 .1e − 96 .2e − 113 .1e − 132
 |f 1(x 3)| .1e − 1050 .3e − 1545 .2e − 1812 .5e − 2123
f 2, x 0 = 1.5
 |f 2(x 1)| .95 .95 .64 .3e − 4
 |f 2(x 2)| .1e − 10 .1e − 4 .3e − 6 .5e − 100
 |f 2(x 3)| .2e − 88 .5e − 90 .1e − 111 .4e − 1632
f 2, x 0 = 3
 |f 2(x 1)| .2e − 24 .1e − 25 .3e − 27 .2e − 44
 |f 2(x 2)| .8e − 358 .8e − 425 .2e − 452 .9e − 743
 |f 2(x 3)| 0 0 0 0
f 3, x 0 = 0.5
 |f 3(x 1)| .5e − 8 .6e − 9 .3e − 9 .9e − 10
 |f 3(x 2)| .8e − 114 .3e − 144 .4e − 149 .6e − 238
 |f 3(x 3)| .2e − 1595 .1e − 2307 .1e − 2386 0
f 3, x 0 = 1.5
 |f 3(x 1)| .2e − 11 .2e − 13 .2e − 11 .1e − 11
 |f 3(x 2)| .2e − 158 .6e − 214 .1e − 179 .6e − 183
 |f 3(x 3)| .3e − 2218 .8e − 3423 .2e − 2869 .6e − 2935
f 4, x 0 = 5
 |f 4(x 1)| .8e − 1 .25 .8e − 1 .1e − 2
 |f 4(x 2)| .4e − 15 .1e − 6 .3e − 16 .3e − 56
 |f 4(x 3)| .1e − 213 .7e − 109 .9e − 262 .1e − 914
f 4, x 0 = 4
 |f 4(x 1)| .92 .9e − 1 2.51 .1e − 17
 |f 4(x 2)| .4e − 5 .1e − 25 .6e − 3 .3e − 295
 |f 4(x 3)| .1e − 73 .1e − 423 .6e − 49 0
f 5, x 0 = −1
 |f 5(x 1)| .1e − 3 .1e − 7 .8e − 4 .7e − 5
 |f 5(x 2)| .2e − 66 .5e − 144 .9e − 80 .4e − 92
 |f 5(x 3)| .4e − 946 .8e − 2327 .1e − 1294 .6e − 1488
f 5, x 0 = −0.6
 |f 5(x 1)| .2e44770 .4e − 1 .2e44770 .1
 |f 5(x 2)| .4e44769 .4e − 40 .7e44768 .3e − 23
 |f 5(x 3)| .5e44767 .4e − 664 .2e44767 .8e − 386
f 6, x 0 = 0.5
 |f 6(x 1)| .6e − 7 .9e − 9 .3e − 8 .1e − 14
 |f 6(x 2)| .8e − 116 .1e − 152 .7e − 145 .1e − 254
 |f 6(x 3)| .1e − 1748 .7e − 2456 .9e − 2332 .1e − 3999
f 6, x 0 = 3
 |f 6(x 1)| .7 .53 .79 0.7e − 8
 |f 6(x 2)| .9e − 5 .4e − 16 .2e − 5 .2e − 145
 |f 6(x 3)| .7e − 102 .1e − 269 .5e − 125 0.3e − 2345
f 7, x 0 = 0
 |f 7(x 1)| .1e − 6 .2e − 9 .1e − 5 .8e − 12
 |f 7(x 2)| .5e − 99 .3e − 161 .6e − 98 .1e − 199
 |f 7(x 3)| .3e − 1394 .6e − 2562 .5e − 1575 .3e − 3205
f 7, x 0 = 2.2
 |f 7(x 1)| 1 .9e − 2 1 .8e − 8
 |f 7(x 2)| 1 .6e − 40 D .9e − 136
 |f 7(x 3)| D .5e − 651 D .4e − 2183
f 8, x 0 = 0.5
 |f 8(x 1)| .1e7 .97 .1e7 1.0
 |f 8(x 2)| 16856.81 .2e − 25 16856.81 .8e − 15
 |f 8(x 3)| 183.46 .3e − 435 183.46 .1e − 256
f 8, x 0 = 1
 |f 8(x 1)| .2e − 2 .5e − 5 .2e − 2 .6e − 3
 |f 8(x 2)| .2e − 65 .1e − 109 .2e − 65 .7e − 67
 |f 8(x 3)| .6e − 1073 .1e − 1782 .6e − 1073 .1e − 1089

*D stands for divergence.

5. Attraction Basins

Let ω i be the roots of the complex polynomial p n(x), n ≥ 1, xC, where i = 1,2, 3,…, n. We use two different techniques to generate basins of attraction on MATLAB software. We take a square box of [−2,2] × [−2,2] ∈ C in the first technique. For every initial guess x 0, a specific color is assigned according to the exact root, and dark blue is assigned for the divergence of the method. We use |f(x k)| < 10−5 as the stopping criteria for convergence and the maximum number of iterations are 30. “Jet” is chosen as the colormap here. For the second technique, the same scale is taken but each initial guess is assigned a color depending upon the number of iterations for the method to converge to any of the roots of the given function. We use 25 as the maximum number of iterations; the stopping criteria are the same as above and colormap is selected as “hot.” The method is considered divergent for that initial guess if it does not converge in the maximum number of iterations and this case is displayed by black color.

We take three test examples to obtain basins of attraction, which are given as p 3(x) = x 3 − 1,  p 4(x) = x 4 − 10x 2 + 9, and p 5(x) = x 5 − 1. The roots of p 3(x) are 1.0, −0.5000 + 0.86605I, and −0.5000 − 0.86605I, the roots of p 4(x) are −3, 3, −1, 1, and for p 5(x) roots are 1.0, 0.3090 + 0.95105I, −0.8090 + 0.58778I, −0.8090 − 0.58778I, and 0.30902 − 0.95105I.

We compare the results of our newly constructed method (32) with some existing methods (7) and (9), as given in Section 1. Figures 1, 2, 3, 4, 5, 6, 7, 8, and 9 show the dynamics of the methods (7), (9), and (32) for the polynomials x 3 − 1, x 4 − 1, and x 5 − 1. Two types of attraction basins are given in all figures. One can easily see that the appearance of darker region shows that the method consumes a fewer number of iterations. Color maps for both types are given with each figure which shows the root to which an initial guess converges and the number of iterations in which the convergence occurs.

Figure 1.

Figure 1

Basins of attraction of method (7) for p 3(x).

Figure 2.

Figure 2

Basins of attraction of method (9) for p 3(x).

Figure 3.

Figure 3

Basins of attraction of method (32) for p 3(x).

Figure 4.

Figure 4

Basins of attraction of method (7) for p 4(x).

Figure 5.

Figure 5

Basins of attraction of method (9) for p 4(x).

Figure 6.

Figure 6

Basins of attraction of method (32) for p 4(x).

Figure 7.

Figure 7

Basins of attraction of method (7) for p 5(x).

Figure 8.

Figure 8

Basins of attraction of (9) for p 5(x).

Figure 9.

Figure 9

Basins of attraction of (32) for p 5(x).

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

References

  • 1.Sharma J. R., Guha R. K., Gupta P. Improved King's methods with optimal order of convergence based on rational approximations. Applied Mathematics Letters. 2013;26(4):473–480. doi: 10.1016/j.aml.2012.11.011. [DOI] [Google Scholar]
  • 2.Soleymani F., Shateyi S., Salmani H. Computing simple roots by an optimal sixteenth-order class. Journal of Applied Mathematics. 2012;2012:13. doi: 10.1155/2012/958020.958020 [DOI] [Google Scholar]
  • 3.Soleymani F., Karimi Vanani S., Paghaleh M. J. A class of three-step derivative-free root solvers with optimal convergence order. Journal of Applied Mathematics. 2012;2012:15. doi: 10.1155/2012/568740.568740 [DOI] [Google Scholar]
  • 4.Tornheim L. Convergence of multipoint iterative methods. Journal of the Association for Computing Machinery. 1964;11:210–220. doi: 10.1145/321217.321224. [DOI] [Google Scholar]
  • 5.Jarratt P., Nudds D. The use of rational functions in the iterative solution of equations on a digital computer. The Computer Journal. 1965;8:62–65. doi: 10.1093/comjnl/8.1.62. [DOI] [Google Scholar]
  • 6.Jarratt P. A rational iteration function for solving equations. The Computer Journal. 1966;9:304–307. doi: 10.1093/comjnl/9.3.304. [DOI] [Google Scholar]
  • 7.Cuyt A., Wuytack L. Nonlinear Methods in Numerical Analysis. Elsevier Science Publishers; 1987. [Google Scholar]
  • 8.Field D. A. Convergence rates for Padé-based iterative solutions of equations. Journal of Computational and Applied Mathematics. 1990;32(1-2):69–75. doi: 10.1016/0377-0427(90)90417-X. [DOI] [Google Scholar]
  • 9.Kung H. T., Traub J. F. Optimal order of one-point and multipoint iteration. Journal of the Association for Computing Machinery. 1974;21(4):643–651. doi: 10.1145/321850.321860. [DOI] [Google Scholar]
  • 10.Soleymani F., Sharifi M. On a general efficient class of four-step root-finding methods. International Journal of Mathematics and Computers in Simulation. 2011;5:181–189. [Google Scholar]
  • 11.Steffensen J. F. Remarks on iterations. Scandinavian Actuarial Journal. 1933;16(1):64–72. [Google Scholar]

Articles from The Scientific World Journal are provided here courtesy of Wiley

RESOURCES