Non-Linear Equations

Download as doc, pdf, or txt
Download as doc, pdf, or txt
You are on page 1of 9

Newton-Raphson Method of Solving a Nonlinear Equation

Introduction Methods such as the bisection method and the false position method of finding roots of a nonlinear equation f ( x ) = 0 require bracketing of the root by two guesses. Such methods are called bracketing methods. These methods are always convergent since they are based on reducing the interval between the two guesses so as to zero in on the root of the equation. In the Newton-Raphson method, the root is not bracketed. In fact, only one initial guess of the root is needed to get the iterative process started to find the root of an equation. The method hence falls in the category of open methods. Convergence in open methods is not guaranteed but if the method does converge, it does so much faster than the bracketing methods. Derivation The Newton-Raphson method is based on the principle that if the initial guess of the root of f ( x) = 0 is at xi , then if one draws the tangent to the curve at f ( xi ) , the point xi +1 where the tangent crosses the x -axis is an improved estimate of the root (Figure 1). Using the definition of the slope of a function, at x = xi f ( xi ) = tan f ( xi ) 0 = , xi xi +1 which gives f ( xi ) xi +1 = xi (1) f ( xi ) Equation (1) is called the Newton-Raphson formula for solving nonlinear equations of the form f ( x ) = 0 . So starting with an initial guess, xi , one can find the next guess, xi +1 , by using Equation (1). One can repeat this process until one finds the root within a desirable tolerance. Algorithm The steps of the Newton-Raphson method to find the root of an equation f ( x ) = 0 are 1. Evaluate f ( x ) symbolically 2. Use an initial guess of the root, xi , to estimate the new value of the root, xi +1 , as f ( xi ) xi +1 = xi f ( xi ) 3. Find the absolute relative approximate error a as

xi +1 xi 100 xi +1 4. Compare the absolute relative approximate error with the pre-specified relative error tolerance, s . If a >s , then go to Step 2, else stop the algorithm. Also, a = check if the number of iterations has exceeded the maximum number of iterations allowed. If so, one needs to terminate the algorithm and notify the user.

f (x)

f (xi)

[xi, f (xi)]

f (xi+1) xi+2 xi+1 xi x

Figure 1 Geometrical illustration of the Newton-Raphson method. Example 1 You are working for DOWN THE TOILET COMPANY that makes floats for ABC commodes. The floating ball has a specific gravity of 0.6 and has a radius of 5.5 cm. You are asked to find the depth to which the ball is submerged when floating in water.

Figure 2 Floating ball problem.

The equation that gives the depth x in meters to which the ball is submerged under water is given by x 3 0.165 x 2 + 3.993 10 4 = 0 Use the Newton-Raphson method of finding roots of equations to find a) the depth x to which the ball is submerged under water. Conduct three iterations to estimate the root of the above equation. b) the absolute relative approximate error at the end of each iteration, and c) the number of significant digits at least correct at the end of each iteration. Solution f ( x ) = x 3 0.165 x 2 + 3.993 10 4 f ( x ) = 3 x 2 0.33 x

Let us assume the initial guess of the root of f ( x ) = 0 is x0 = 0.05 m. This is a reasonable guess (discuss why x = 0 and x = 0.11 m are not good choices) as the extreme values of the depth x would be 0 and the diameter (0.11 m) of the ball. Iteration 1 The estimate of the root is f ( x0 ) x1 = x0 f ( x0 ) = 0.05 = 0.05

( 0.05) 3 0.165( 0.05) 2 + 3.993 10 4 2 3( 0.05) 0.33( 0.05)

1.118 10 4 9 10 3 = 0.05 ( 0.01242 ) = 0.06242 The absolute relative approximate error a at the end of Iteration 1 is a = = x1 x0 100 x1

0.06242 0.05 100 0.06242 = 19.90% The number of significant digits at least correct is 0, as you need an absolute relative approximate error of 5% or less for at least one significant digit to be correct in your result. Iteration 2 The estimate of the root is f ( x1 ) x 2 = x1 f ( x1 )

( 0.06242) 3 0.165( 0.06242) 2 + 3.993 10 4 = 0.06242 2 3( 0.06242 ) 0.33( 0.06242 )

3.97781 10 7 8.90973 10 3 = 0.06242 4.4646 10 5 = 0.06238 The absolute relative approximate error a at the end of Iteration 2 is = 0.06242

a = =

x2 x1 100 x2

0.06238 0.06242 100 0.06238 = 0.0716% 2m The maximum value of m for which a 0.5 10 is 2.844. Hence, the number of significant digits at least correct in the answer is 2. Iteration 3 The estimate of the root is f ( x2 ) x3 = x 2 f ( x 2 ) = 0.06238 = 0.06238

( 0.06238) 3 0.165( 0.06238) 2 + 3.993 10 4 2 3( 0.06238) 0.33( 0.06238)

4.44 10 11 8.91171 10 3 = 0.06238 4.9822 10 9 = 0.06238 The absolute relative approximate error a at the end of Iteration 3 is

0.06238 0.06238 100 0.06238 =0 The number of significant digits at least correct is 4, as only 4 significant digits are carried through in all the calculations. a = Drawbacks of the Newton-Raphson Method 1. Divergence at inflection points If the selection of the initial guess or an iterated value of the root turns out to be close to the inflection point (see the definition in the appendix of this chapter) of the function f ( x ) in the equation f ( x ) = 0 , Newton-Raphson method may start diverging away from the root. It may then start converging back to the root. For example, to find the root of the equation 3 f ( x ) = ( x 1) + 0.512 = 0 the Newton-Raphson method reduces to 3 ( xi 1)3 + 0.512 xi +1 = xi 3( xi 1) 2

Starting with an initial guess of x0 = 5.0 , Table 1 shows the iterated values of the root of the equation. As you can observe, the root starts to diverge at Iteration 6 because the previous estimate of 0.92589 is close to the inflection point of x = 1 (the value of f ' ( x ) is zero at the inflection point). Eventually, after 12 more iterations the root converges to the exact value of x = 0.2 . Table 1 Divergence near inflection point. Iteration xi Number 0 5.0000 1 3.6560 2 2.7465 3 2.1084 4 1.6000 5 0.92589 6 30.119 7 19.746 8 12.831 9 8.2217 10 5.1498 11 3.1044 12 1.7464 13 0.85356 14 0.28538 15 0.039784 16 0.17475 17 0.19924 18 0.2

Figure 3 Divergence at inflection point for f ( x ) = ( x 1) 3 = 0 .

2. Division by zero For the equation f ( x ) = x 3 0.03x 2 + 2.4 10 6 = 0 the Newton-Raphson method reduces to 3 2 x 0.03xi + 2.4 106 xi +1 = xi i 2 3 xi 0.06 xi For x0 = 0 or x0 = 0.02 , division by zero occurs (Figure 4). For an initial guess close to 0.02 such as x0 = 0.01999 , one may avoid division by zero, but then the denominator in the formula is a small number. For this case, as given in Table 2, even after 9 iterations, the Newton-Raphson method does not converge. Table 2 Division by near zero in Newton-Raphson method. Iteratio xi f ( xi ) a % n Number 0.019990 0 1.60000 10-6 2.6480 100.75 1 18.778 1.7620 50.282 2 5.5638 1.1714 50.422 3 1.6485 0.77765 50.632 4 0.48842 0.51518 50.946 5 0.14470 0.34025 51.413 6 0.042862 0.22369 52.107 7 0.012692 0.14608 53.127 8 0.0037553 54.602 9 0.0011091 0.094490
1 .00E-05 7.50E-06 5.00E-06 2.50E-06 0.00E+ 00 -0.03 -0.02 -0.01 -2.50E-06 -5.00E-06 -7.50E-06 -1 .00E-05 0 0.01 0.02 0.03

f(x)

x
0.04

0.02

Figure 4 Pitfall of division by zero or a near zero number. 3. Oscillations near local maximum and minimum

Results obtained from the Newton-Raphson method may oscillate about the local maximum or minimum without converging on a root but converging on the local maximum or minimum. Eventually, it may lead to division by a number close to zero and may diverge. For example, for f ( x) = x 2 + 2 = 0 the equation has no real roots (Figure 5 and Table 3).
6

f(x)

4
0 -2 -1 0 -1 1 2 3

x
3.142

-1.75

-0.3040

0.5

Figure 5 Oscillations around local minima for f ( x ) = x 2 + 2 .

Table 3 Oscillations near local maxima and minima in Newton-Raphson method. Iteratio xi f ( xi ) a % n Number 1.0000 3.00 0 0.5 2.25 30 1 1.75 5.063 0.00 2 2.092 128.571 3 0.30357 11.87 476.47 4 3.1423 4 109.66 5 1.2529 3.570 150.80 6 2.029 829.88 7 0.17166 34.94 102.99 8 5.7395 2 112.93 9 2.6955 9.266 175.96 0.97678 2.954

4. Root jumping In some case where the function f (x) is oscillating and has a number of roots, one may choose an initial guess close to a root. However, the guesses may jump and converge to some other root. For example for solving the equation sin x = 0 if you choose x0 = 2.4 = ( 7.539822) as an initial guess, it converges to the root of x = 0 as shown in Table 4 and Figure 6. However, one may have chosen this as an initial guess to converge to x = 2 = 6.2831853 . Table 4 Root jumping in Newton-Raphson method. Iteratio xi f ( xi ) a % n Number 7.539822 0.951 0 68.973 4.462 0.969 1 711.44 0.5499 0.5226 2 971.91 0.06307 0.06303 3 4 5 7.54 10 4 8.376 10 8.375 10 4 13 13 4.28 1010 5 1.9586110 1.9586110
f(x)
1 .5

0.5

0 -2

x
0

-0.06307
-0.5

0.5499

4.461

7.539822

1 0

-1

-1 .5

Figure 6 Root jumping from intended location of root for f ( x ) = sin x = 0 . Appendix A. What is an inflection point? For a function f ( x ) , the point where the concavity changes from up-to-down or down-to-up is called its inflection point. For example, for the function f ( x ) = ( x 1) 3 , the concavity changes at x = 1 (see Figure 3), and hence (1,0) is an inflection point. An inflection points MAY exist at a point where f ( x) = 0 and where f ' ' ( x ) does not exist. The reason we say that it MAY exist is because if f ( x) = 0 , it only makes it a

possible inflection point. For example, for f ( x ) = x 4 16 , f (0) = 0 , but the concavity does not change at x = 0 . Hence the point (0, 16) is not an inflection point of f ( x ) = x 4 16 . For f ( x ) = ( x 1) 3 , f (x) changes sign at x = 1 ( f ( x) < 0 for x < 1 , and f ( x) > 0 for x > 1 ), and thus brings up the Inflection Point Theorem for a function f (x) that states the following. If f ' (c) exists and f (c) changes sign at x = c , then the point (c, f (c)) is an inflection point of the graph of f . Appendix B. Derivation of Newton-Raphson method from Taylor series Newton-Raphson method can also be derived from Taylor series. For a general function f ( x ) , the Taylor series is f" ( xi ) f ( xi +1 ) = f ( xi ) + f ( xi )( xi +1 xi ) + ( xi +1 xi ) 2 + 2! As an approximation, taking only the first two terms of the right hand side, f ( xi +1 ) f ( xi ) + f ( xi )( xi +1 xi ) and we are seeking a point where f ( x ) = 0, that is, if we assume f ( xi +1 ) = 0 , 0 f ( xi ) + f ( xi )( xi +1 xi ) which gives f ( xi ) xi +1 = xi f' ( xi ) This is the same Newton-Raphson method formula series as derived previously using the geometric method.

You might also like