A fast and robust method for computing real roots of nonlinear equations

A fast and robust method for computing real roots of nonlinear equations

Accepted Manuscript A fast and robust method for computing real roots of nonlinear equations Xiao-Diao Chen, Jiaer Shi, Weiyin Ma PII: DOI: Referenc...

432KB Sizes 2 Downloads 159 Views

Accepted Manuscript A fast and robust method for computing real roots of nonlinear equations

Xiao-Diao Chen, Jiaer Shi, Weiyin Ma

PII: DOI: Reference:

S0893-9659(16)30380-9 http://dx.doi.org/10.1016/j.aml.2016.12.013 AML 5150

To appear in:

Applied Mathematics Letters

Received date : 28 October 2016 Revised date : 22 December 2016 Accepted date : 22 December 2016 Please cite this article as: X. Chen, et al., A fast and robust method for computing real roots of nonlinear equations, Appl. Math. Lett. (2016), http://dx.doi.org/10.1016/j.aml.2016.12.013 This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

*Manuscript Click here to view linked References

A fast and robust method for computing real roots of nonlinear equations Xiao-Diao Chen a,b, Jiaer Shi a, Weiyin Ma b a School

of Computer, Hangzhou Dianzi University, Hangzhou, China.

b Department

of MBE, City University of Hong Kong, Hong Kong, China

Abstract The root-finding problem of a univariate nonlinear equation is a fundamental and long-studied problem, and has wide applications in mathematics and engineering computation. This paper presents a fast and robust method for computing the simple root of a nonlinear equation within an interval. It turns the root-finding problem of a nonlinear equation into the solution of a set of linear equations, and explicit formulae are also provided to obtain the solution in a progressive manner. The method avoids the computation of derivatives, and achieves the convergence order 2n−1 by using n evaluations of the function, which is optimal according to Kung and Traub’s conjecture. Comparing with the prevailing Newton’s methods, it can ensure the convergence to the simple root within the given interval. Numerical examples show that the performance of the derived method is better than those of the prevailing methods. Key words: Root finding; Non-linear equations; Linear equations; Progressive formula; Optimal convergence order.

1 2 3 4 5 6 7 8 9 10 11 12 13 14

1 Introduction Finding the roots of a nonlinear equation f (t) = 0 is a common and important problem in mathematics and engineering computation. Many modified iterative methods have been developed to improve the local order of convergence, including Newton, Halley or Ostrowski’s methods. In an iterative step, the convergence order p can be improved by increasing the number n of functional evaluations (FE). The balance between p and n is measured by using p1/n , which is called an efficiency index [1, 6, 7, 10–13, 16–18]. It is conjectured that the order of convergence of any multi-point method cannot exceed the optimal bound 2n−1 [8]. Some classical methods, e.g., Newton’s method, Newton-Secant method and Ostrowski’s method, have efficiency indices of 21/2 ≈ 1.414, 31/3 ≈ 1.442 and 41/3 ≈ 1.587, respectively, and they usually need evaluations of its derivatives. In [9], Li, Mu, Ma and Wang presented a method of sixteen convergence order, whose efficiency index is 161/6 ≈ 1.587. Email addresses: [email protected] (Xiao-Diao Chen), [email protected] (Weiyin Ma).

Preprint submitted to Elsevier

22 December 2016

35

Some optimal methods with eighth convergence order have also been developed, whose efficiency index is 81/4 ≈ 1.682 [4, 14, 15]. For cases having one root within an interval, the above methods still start from an initial value from one side, and their convergence is sensitive to the selection of the initial value. In some worst cases, the iterative methods cannot converge to the proper simple root, even with a good initial value [3] (see also Example 3 in Section 3 for more details). Chen et. al. provided an efficient method based on progressive interpolation, which starts from an interval bounding the unique simple root, and achieves convergence order 3 · 2n−3 with n FEs and no derivative evaluations [3]. It ensures that the roots of progressive interpolation polynomials are within the given interval, and converge to the simple root. This paper focuses on how to compute the simple root t⋆ of a nonlinear equation within an interval [a, b]. It turns the root-finding problem of the given nonlinear equation f (t) into those of a sequence of linear equations. It ensures that each linear equation li (t) has a root ti ∈ [a, b] with explicit formulae, where ti is well-approximated to t⋆ in a progressive way, i = 3, 4, · · · . It can ensure the convergence to the simple root with convergence order 2n−1 by using n FEs without derivative evaluations, which is higher than that of the method in [3] and is optimal in the conjecture of [8] (see also Table 1). Numerical examples show that the performance of the proposed method is better than those of prevailing methods.

36

2

15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34

37 38 39 40 41 42 43 44 45 46 47 48

The main result

Note that a root t⋆ of f (t) should be a simple root of f (t)/f ′ (t). Without loss of generality, we suppose that the simple root t⋆ is the unique root of f (t) within [a, b], and h = b − a. In cases with only one given initial value x0 , one can obtain x1 such that |x1 −t⋆ | < |x0 −x⋆ |/2 by using some iterative methods, such as the Newton’s method, and can thus obtain an interval bounding t⋆ . The details are as follows. If f (x0 ) · f (x1 ) < 0, x0 and x1 already bound the simple root t⋆ ; otherwise, we have that 2x1 − x0 and x1 bound t⋆ . In this paper, we discuss how to compute the simple root t⋆ ∈ [a, b] of f (t), where f (a) · f (b) < 0. We want to find a sequence of linear equations which have roots well-approximating to t⋆ . Let t1 = a and t2 = b. Let the first linear equation l3 (t) be the one interpolating f (t) at two points t1 and t2 , which is trivial to compute its root t3 . Let t − ti li (t) = i−1 , i = 3, 4, . . . , n, (1) P αi,j tj−2 j=2

49 50 51 52 53 54 55

where the i − 1 unknowns, i.e., ti and αi,j , j = 2, 3, · · · , i − 1, are determined li (tj ) = f (tj ), j = 1, 2, . . . , i − 1. (2) by

Note that the denominator of li (t) should not be zero within [a, b], and li (t) = 0 is equivalent to a linear equation with root ti . On the other hand, combining the assumption that f (a) · f (b) < 0 and the constraints that li (a) = f (a) and li (b) = f (b), li (t) must have the root ti . We introduce Theorem 3.5.1 in Page 67, Chapter 3.5 of [5] as follows. 2

56 57 58

59

60 61

Theorem 1. Let w0 , w1 , · · · , wr be r+1 distinct points in [a, b], and n0 , · · · , nr be r+1 integers ≥ 0. Let N = n0 + · · · + nr + r. Suppose that g(t) is a polynomial of degree N such that g (i) (wj ) = f (i) (wj ), i = 0, · · · , nj , j = 0, · · · , r. Then there exists ξ1 (t) ∈ [a, b] such that r f (N +1) (ξ1 (t)) Y f (t) − g(t) = (t − wi )ni .  (N + 1)! i=0

From Eq. (2), tj is a root of the polynomial Ei (t) = f (t)−li (t), j = 1, 2, · · · , i− 1. Combining with Theorem 1, wei−1 have that Y (3) Ei (t) = Fi (t) · (t − tj ), i = 3, 4, · · · , j=1

62 63

where Fi (t) is a bounded rational polynomial within [a, b]. From Eq. (3), we have that |t⋆ − ti | = O(|li (t⋆ ) − li (ti )|) = O(|li (t⋆ )|) (4) i−1 Q = O(|li (t⋆ ) − f (t⋆ )|) = Ei (t⋆ ) = O(| (t⋆ − tj )|), i = 3, 4, · · · , j=1

64

65 66 67 68

69 70

which leads to

), i = 3, 4, · · · .

(5)

Remark 1. In principle, the selection of t1 and t2 can be optimized by using the methods in [17, 18], which can lead to a better approximation effect to the root t⋆ . Eq. (2) is linear in i − 1 unknowns canbibe Mi and · vi = , turned into a matrix form (6) where Mi is a matrix of size (i − 1) × (i − 1), both vi and bi are vectors of       i−3 size i − 1, which 1 f (t1 ) can f (tbe · · · f (tas αi,2 t1 1 ) · tdenoted 1 1 ) · t1 Mi =

71

i−2

|t⋆ − ti | = O(h2

    

f (t2 ) . ..

f (t2 ) · t2 . ..

···

···

f (t2 ) · ti−3 2 . ..



1  . , vi .. 

=

f (ti−1 ) f (ti−1 ) · ti−1 · · · f (ti−1 ) · ti−3 i−1 1



 .   ..   ,    αi,i−1  ti



Combining Eq. (6) with Cramer’s thatt1 f (t1 ) f (t1rule, ) · t1 we · · · obtain f (t1 ) · ti−3 1 ti =

f (t2 ) · t2 · · · f (t2 ) · ti−3 t2 f (t2 ) 2 . .. .. .. . . . ··· . . f (ti−1 ) f (ti−1 ) · ti−1 · · · f (ti−1 ) · ti−3 ti−1 i−1 . f (t1 ) f (t1 ) · t1 · · · f (t1 ) · ti−3 1 1 f (t2 ) · t2 · · · f (t2 ) · ti−3 1 f (t2 ) 2 . .. .. .. . . . ··· . . f (ti−1 ) f (ti−1 ) · ti−1 · · · f (ti−1 ) · ti−3 1

bi =

    

t2 . ..

ti−1

  .  

(7)

i−1

72 73

To expand the two determinants of Eq. (7) in their last columns, it can be verified that i−1 i−1 i−1 Q P P ( Ai,j tj ) · ( f (tk )) Ai,j tj j=1 j=1 k=1 ti = i−1 (8) = i−1 , i ≥ 3, i−1 Q P P ( Ai,j ) · ( f (tk )) Ai,j j=1

j=1

k=1

3

74





where Di,j is the Vandermonde of the matrix such that 1 t1determinant · · · ti−3 1 ¯ i,j ) = Di,j = det(M

. . .  1 det  1  . . .

.. .

···

tj−1 · · · tj+1 · · · .. . ···

    i−3  tj−1   ti−3 j−1  ..   .  .. .

=

1 ti−1 · · · ti−3 i−1

75

and Ai,j = (−1)j ·

76

result as follows.

77 78 79 80 81 82 83

84

85 86 87 88 89 90 91 92 93 94 95 96 97 98

 1,        

Q

i = 3, (tk − tr ), i > 3,

1 ≤ k < r ≤ i − 1, k, r 6= j,

Di,j . Combining Eq. (5) with Eq. (8), we have the main f (tj )

Theorem 2. There are n functional evaluations for computing tn+1 from Formula (8), and it achieves the convergence rate 2n−1 , which is optimal in the Kung’s conjecture. Proof. From Formula (8), there are n functional evaluations f (tj ), j = 1, 2, · · · n+1−2 , n, for computing tn+1 . From Eq. (5), we have that |t⋆ − tn+1 | = O(h2 )= n−1 O(h2 ), and the corresponding convergence rate is 2n−1 , which is directly from its definition.  3

Numerical results and discussions

Let M1 , M2 , M3 , M4 and M5 denote the methods in [9], [19], [15], [3] and the method in this paper, respectively, where q in [19] is set as 2. We show several examples to illustrate M5 and its comparisons with other four methods. In principle, M1 , M2 and M3 need 6, 6 and 4 FEs in each iterative step, while M4 and M5 can stop at arbitrary n ≥ 3 FEs. We utilize 12 FEs for comparing the above five methods. Firstly, we compare convergence order (CO) and efficiency index (EI) of different methods. Table 1 shows that both M2 and M5 have the best EI among the five methods. Example 1 (Ref. [9]): Given f1 (t) = (t − 1)3 − 1 with exact root t⋆ = 2, and initial value is t1 = 1.9. With the Newton’s method, we obtain that t2 = t1 − f1 (t1 )/f1′ (t1 ) = 2.011, and t⋆ is bounded by t1 and t2 . So the initial interval of both M4 and M5 is set as [t1 , t2 ] = [1.9, 2.011]. The comparison results are shown in Table 2. It shows that M5 achieves a better performance than that of M4 . Table 1. Efficiency indexes and convergence orders between different methods M1 [9]

M2 [19]

M3 [15]

CO(n = 12)

28

211

29



EI

1.587

1.887

1.681

1.843 2

99 100 101 102

M4 [3] 29

M5 211 1.887

Example 2 (Ref. [3]): Given f2 (t) = 10150−5x − 1 with exact root t⋆ ≈ 5.477 and initial value t1 = 5.494. Similarly, we obtain t2 = t1 −f2 (t1 )/f2′ (t1 ) = 5.436, and the initial interval [t2 , t1 ] bounds t⋆ . The corresponding results are shown in Table 2. As shown in Table 2, both M1 and M3 diverge (denoted by “/”), 4

103 104

M2 doesn’t converge within 80 iterative processes, while M4 and M5 converge, and M5 achieves a slightly better result than M4 . Table 2. Comparisons of Examples 1 and 2 between difference methods with 12 FEs

105 106

Example

M1 [9]

M2 [19]

M3 [15]

M4 [3]

M5

Exam. 1

4.2e − 18

2.5e − 2695

5.3e − 540

2.5e − 2607

2e − 3305

Exam. 2

/

2.57

/

1e − 371

8e − 476

Example 3 (extracted from [9]): We have tested the following eight smooth functions  2  f3 (t) = e−t +t+3 − t + 2,      f4 (t) = et − 1,     f5 (t) = t3 − 10,     f (t) = et2 +7t−30 − 1,

t⋆ = 2.490, x0 = 2.3, t⋆ = 0,

6

107

 f7 (t) = t5 + t − 104 ,   √    f8 (t) = t − 1/t − 3,     f9 (t) = et + t − 20,     f (t) = ln(t) + √t − 5,

109

= 2.154, x0 = 2,

t⋆

= 3,

x0 = 2.97,

t⋆ = 6.308, x0 = 6, t⋆ = 9.633, x0 = 9.4, t⋆ = 2.842, x0 = 2.5, t⋆ = 8.309, x0 = 7.9.

10

108

x0 = 0.14,

t⋆

The corresponding results are shown in Table 3. It shows that M5 can achieve a better performance than those of other four methods. Table 3. Errors of Example 3 between different methods with 12 FEs

110 111 112 113 114 115 116

Method

f3 (t)

f4 (t)

f5 (t)

f6 (t)

f7 (t)

f8 (t)

f9 (t)

f10 (t)

M1

2e-188

1e-314

2e-322

2e-185

2e-274

4e-520

7e-208

3e-428

M2

1e-1481

6e-2485

5e-1339

1e-1246

3e-506

3e-4101

1e-206

1e-3449

M3

5e-364

3e-641

5e-615

1e-403

2e-544

2e-1094

3e-438

3e-909

M4

2e-1638

3e-2749

9e-2851

2e-2044

6e-2622

7e-4301

1e-2195

2e-3622

M5

4e-1962

1e-3452

2e-3745

2e-2793

1e-3495

2e-4198

3e-2966

3e-3585

1 Example 4. We have also tested the oscillating function f11 (t) = cos( ) + t et − 2.1 (with three roots t⋆0 ≈ 0.1506, t⋆1 ≈ 0.1704 and t⋆2 ≈ 0.6863) and 1 f12 (t) = cos( ) + et − 2.5 (with one root t⋆3 ≈ 0.7890) within [0.01,1.1], see t also Fig. 1. The comparison results are shown in Table 4, where t1 (t⋆i ) denotes the initial value t1 and its closest root t⋆i . It shows that M1 , M2 and M3 may converge to a wrong root, M2 converges very slowly for the f12 case, and M5 work much better and more robust than M1 , M2 , M3 and M4 . Table 4. Errors of Example 4 between difference methods with 12 FEs Exam

f11 (t)

f12 (t)

0.13(t⋆0 )

0.1596(t⋆0 )

0.2 (t⋆1 )

0.22 (t⋆1 )

M1

3e-104

7e-1(t⋆2 )

1e-105

5e-85

3e-295

7e-90

8e-113

M2

6e-789(t⋆2 )

3e-528(t⋆1 )

2e-1138

2e-368(t⋆2 )

8e-2079

1.03

6e-1

M3

4e-181

5e-53(t⋆1 )

1e-274

3e-34(t⋆0 )

4e-617

2e-190

5e-74

M4

5e-1194

4e-89

1e-1084

1e-895

6e-2218

5e-903

1e-913

M5

1e-559

1.4e-502

4e-1261

5e-1026

2e-3690

6e-1413

2e-1361

5

0.78(t⋆2 )

0.1065(t⋆3 )

0.15958(t⋆3 )

(a)

(b)

Fig. 1. Plots of oscillating functions (a) f11 (t); and (b) f12 (t). 117

4

Conclusions

124

This paper presents an improved method of the one in [3]. It achieves the optimal convergence order 2n−1 by using n FEs and no derivative evaluations, which is much higher than 3/4 · 2n−1 of the method in [3]. Different from prevailing iterative methods starting from an initial value, it starts from an interval bounding the simple root and ensures the convergence. Numerical examples show that the new method can achieve better performance and better approximation effect than those of M1 , M2 , M3 and M4 .

125

Acknowledgement

118 119 120 121 122 123

127

This research was partially supported by City University of Hong Kong (SRG 7004452), the National Science Foundation of China (61672009, 61370218).

128

References

129

[1]

126

130 131 132

[2]

133 134

[3]

135 136 137

[4]

138 139 140

[5]

141 142

[6]

143 144

[7]

S. Artidiello, A. Cordero, J.R. Torregrosa, M.P. Vassileva, Two weighted eight-order classes of iterative root-finding methods, Int J Comput Math 92(9)(2015):1790–1805. X. Chen, W. Ma, Y. Ye, A rational cubic clipping method for computing real roots of a polynomial, Comput Aided Geom D 38 (10) (2015) 40–50. X. Chen, Y. Zhang, J. Shi, Y. Wang, An efficient method based on progressive interpolation for solving non-linear equations, Appl Math Lett, 61 (11) (2016) 67–72. A. Cordero, J. Torregrosa, M. Vassileva, A family of modified Ostrowski’s methods with optimal eighth order of convergence, Appl Math Lett 24 (12) (2011) 2082–2086. P. Davis, Interpolation and Approximation, Dover Publications, New York, 1975. J. Dˇzuni´c, On efficient two-parameter methods for solving nonlinear equations, Numer Algorithms 63 (3) (2013) 549–569. Grau-S´anchez, M., Angela Grau, Noguera, M. Ostrowski type methods for

6

145 146 147

[8]

148 149

[9]

150 151

[10]

152 153 154 155

[11] [12]

156 157 158

[13]

159 160

[14]

161 162

[15]

163 164

[16]

165 166

[17]

167 168

[18]

169 170 171 172

[19]

solving systems of nonlinear equations. Applied Mathematics & Computation 218 (6) (2011), 2377–2385. H. Kung, J. Traub, Optimal order of one-point and multi-point iteration, Appl Math Comput 21 (4) (1974) 643–651. X. Li, C. Mu, J. Ma, C. Wang, Sixteenth-order method for nonlinear equations, Appl Math Comput 215 (10) (2010) 3754–3758. M. Noor, W. Khan, New iterative methods for solving nonlinear equation by using homotopy perturbation method, Appl Math Comput 219 (8) (2012) 3565– 3574. A. Ostrowski, Solutions of equations and system of equations, New York, 1960. M. S.Petkovi´c, L.Z. Ran¨ci´c, On the guaranteed convergence of a cubically convergent Weierstrass-like root-finding method, Int J Comput Math 92(6)(2015):1–10. P. Sargolzaei, F. Soleymani, Accurate fourteenth-order methods for solving nonlinear equations, Numer Algorithms 58 (4) (2011) 513–527. J. Sharma, H. Arora, An efficient family of weighted-newton methods with optimal eighth order convergence, Appl Math Lett 29 (1) (2014) 1–6. J. Sharma, H. Arora, A new family of optimal eighth order methods with dynamics for nonlinear equations, Appl Math Comput 273 (1) (2016) 924–933. X. Wang, T. Zhang, A family of Steffensen type methods with seventh-order convergence, Numer Algorithms 62 (3) (2013) 429–444. B. Yun, A non-iterative method for solving non-linear equations. Appl Math Comput 198 (2)(2008)691–699. B. Yun, M.S. Petkovi´c, Iterative methods based on the signum function approach for solving nonlinear equations.Numer Algorithms 2009, 52 (4)(2009) 649–662. B. Yun, A family of optimal multi-point root-finding methods based on the interpolating polynomials, Appl Math Sci 35 (8) (2014) 1723–1730.

7