Hafner–Sarnak–McCurley constant: Difference between revisions

From formulasearchengine
Jump to navigation Jump to search
en>Helpful Pixie Bot
m ISBNs (Build KE)
 
en>Tassedethe
Line 1: Line 1:
Hi there, I am Alyson Boon even though it is not the title on my birth certificate. I am presently a travel agent. Her family lives in Ohio. To climb is something I really appreciate doing.<br><br>my web blog [http://ltreme.com/index.php?do=/profile-127790/info/ psychic readings]
{{Use dmy dates|date=July 2013}}
In [[numerical analysis]], '''Halley’s method''' is a [[root-finding algorithm]] used for functions of one real variable with a continuous second derivative, i.e., [[Differentiability class|C<sup>2</sup>]] functions. It is named after its inventor [[Edmond Halley]].
 
The algorithm is second in the class of [[Householder's method]]s, right after [[Newton's method]]. Like the latter, it produces iteratively a sequence of approximations to the root; their [[rate of convergence]] to the root is cubic. Multidimensional versions of this method exist.
 
Halley's method can be viewed as exactly finding the roots of a linear-over-linear [[Padé_approximant|Padé approximation]] to the function, in contrast to [[Newton's method]]/[[Secant method]] (approximates/[[Interpolation|interpolates]] the function linearly) or [[Cauchy's method]]/[[Muller's method]] (approximates/interpolates the function quadratically).<ref>{{cite doi|10.1137/110838297}}</ref>
 
==Method==
Halley’s method is a numerical algorithm for solving the nonlinear equation ƒ(''x'') = 0. In this case, the function ƒ has to be a function of one real variable. The method consists of a sequence of iterations:
 
:<math>x_{n+1} = x_n - \frac {2 f(x_n) f'(x_n)} {2 {[f'(x_n)]}^2 - f(x_n) f''(x_n)} </math>
 
beginning with an initial guess ''x''<sub>0</sub>.
 
If ƒ is a three times continuously differentiable function and ''a'' is a zero of &fnof; but not of its derivative, then, in a neighborhood of ''a'', the iterates ''x''<sub>''n''</sub> satisfy:
 
:<math>| x_{n+1} - a | \le K \cdot {| x_n - a |}^3,\text{ for some }K > 0.\!</math>
 
This means that the iterates converge to the zero if the initial guess is sufficiently close, and that the convergence is cubic.
 
The following alternative formulation shows the similarity between Halley’s method and Newton’s method. The expression <math>f(x_n)/f'(x_n)</math> is computed only once, and it is particularly useful when <math>f''(x_n)/f'(x_n)</math> can be simplified.
 
:<math>x_{n+1} = x_n -  \frac {f(x_n)} {f'(x_n)} \left[1 - \frac {f(x_n)}{f'(x_n)} \cdot \frac {f''(x_n)} {2 f'(x_n)} \right]^{-1}</math>
 
A further alternative is as below, in which case the technique is sometimes referred to as '''Bailey’s method'''.<ref>See for example the [[Bond Exchange of South Africa]]’s [http://www.jse.co.za/Libraries/BESA_Bond_pricing/bond_pricing_formula_-_final.sflb.ashx ''Bond Pricing Formula Specifications''], where Bailey’s method is suggested when solving for a [[Bond (finance)|bond’s]] [[Yield to maturity]].</ref>
 
:<math>x_{n+1} = x_n -  f(x_n) / \left[f'(x_n) - \frac {f(x_n) f''(x_n)} {2 f'(x_n)} \right]</math>
 
Using any variation, when the [[second derivative]] is very close to zero, the iteration is almost the same as under Newton’s method.
 
==Derivation==
Consider the function
 
:<math>g(x) = \frac {f(x)} {\sqrt{|f'(x)|}}.</math>
 
Any root of ƒ which is ''not'' a root of its derivative is a root of ''g''; and any root of ''g'' is a root of &fnof;. Applying [[Newton's method]] to ''g'' gives
 
:<math>x_{n+1} = x_n - \frac {g(x_n)} {g'(x_n)}</math>
 
with
 
:<math>g'(x) = \frac {2 {[f'(x)]}^2 - f(x) f''(x)} {2 f'(x) \sqrt{|f'(x)|}}, </math>
 
and the result follows. Notice that if ƒ'(''c'') = 0, then one cannot apply this at ''c'' because ''g''(''c'') would be undefined.
 
==Cubic convergence==
Suppose ''a'' is a root of ''f'' but not of its derivative. And suppose that the third derivative of ''f'' exists and is continuous in a neighborhood of ''a'' and ''x''<sub>''n''</sub> is in that neighborhood. Then [[Taylor's theorem]] implies:
 
:<math>0 = f(a) = f(x_n) + f'(x_n) (a - x_n) + \frac{f''(x_n)}{2} (a - x_n)^2 + \frac{f'''(\xi)}{6} (a - x_n)^3</math>
 
and also
 
:<math>0 = f(a) = f(x_n) + f'(x_n) (a - x_n) + \frac{f''(\eta)}{2} (a - x_n)^2,</math>
 
where ξ and η are numbers lying between ''a'' and ''x''<sub>''n''</sub>. Multiply the first equation by <math>2 f'(x_n)\!</math> and subtract from it the second equation times <math>f''(x_n) (a - x_n)\!</math> to give:
 
:<math>
\begin{align}
0 & {} = 2 f(x_n) f'(x_n) + 2 [f'(x_n)]^2 (a - x_n) \\
& {} + f'(x_n) f''(x_n) (a - x_n)^2 + \frac{f'(x_n) f'''(\xi)} {3} (a - x_n)^3 \\
& {} - f(x_n) f''(x_n) (a - x_n) - f'(x_n) f''(x_n) (a - x_n)^2 \\
& {} - \frac{f''(x_n) f''(\eta)} {2} (a - x_n)^3.
\end{align}
</math>
 
Canceling <math>f'(x_n) f''(x_n) (a - x_n)^2\!</math> and re-organizing terms yields:
 
:<math>
\begin{align}
0 = 2 f(x_n) f'(x_n) &+ \big(2 [f'(x_n)]^2 - f(x_n) f''(x_n) \big) (a - x_n) \\
&+ \left( \frac{f'(x_n) f'''(\xi)} {3} - \frac{f''(x_n) f''(\eta)} {2} \right) (a - x_n)^3.
\end{align}
</math>
 
Put the second term on the left side and divide through by <math> 2 [f'(x_n)]^2 - f(x_n) f''(x_n) </math> to get:
 
:<math>
a - x_{n} = \frac{- 2 f(x_n) f'(x_n)}{2 [f'(x_n)]^2 - f(x_n) f''(x_n)}
- \frac{2 f'(x_n) f'''(\xi) - 3 f''(x_n) f''(\eta)} {6(2 [f'(x_n)]^2 - f(x_n) f''(x_n))} (a - x_n)^3.
</math>
 
Thus:
 
:<math>
a - x_{n+1} =
- \frac{2 f'(x_n) f'''(\xi) - 3 f''(x_n) f''(\eta)} {12 [f'(x_n)]^2 - 6 f(x_n) f''(x_n)} (a - x_n)^3.
</math>
 
The limit of the coefficient on the right side as ''x''<sub>''n''</sub> approaches ''a'' is:
 
:<math>
- \frac{2 f'(a) f'''(a) - 3 f''(a) f''(a)} {12 [f'(a)]^2}.
</math>
 
If we take ''K'' to be a little larger than the absolute value of this, we can take absolute values of both sides of the formula and replace the absolute value of coefficient by its upper bound near ''a'' to get:
 
:<math>|a - x_{n+1}| \leq K |a - x_n|^3</math>
 
which is what was to be proved.
 
To summarize,
<math>
\Delta x_{i+1} =  \frac{3 (f^{\prime\prime})^2 - 2 f^\prime f^{\prime\prime\prime}}{12 (f^\prime)^2} (\Delta x_{i})^3 + O[\Delta x_{i}]^4
</math>,
 
where <math>\Delta x_i \triangleq x_i - a</math>.
 
==References==
'''Notes'''
<references/>
 
'''Sources'''
*T.R. Scavo and J.B. Thoo, On the geometry of Halley’s method. ''American Mathematical Monthly'', '''102''':5 (1995), pp.&nbsp;417–426. [http://www.jstor.org/pss/2975033]
*This article began as a translation from [http://fr.wikipedia.org/w/index.php?title=Itération_de_Halley&oldid=11673690 the equivalent article in French Wikipedia], retrieved 22 January 2007.
 
==External links==
* {{MathWorld|urlname=HalleysMethod|title=Halley’s method}}
*[http://math.fullerton.edu/mathews/n2003/Halley'sMethodMod.html Halley's Method by John H. Mathews]
*[http://home.online.no/~pjacklam/notes/halley/halley.pdf Halley's Method by P. J. Acklam]
* ''[http://numbers.computation.free.fr/Constants/Algorithms/newton.html Newton's method and high order iterations]'', Pascal Sebah and Xavier Gourdon, 2001 (the site has a link to a Postscript version for better formula display)
 
{{DEFAULTSORT:Halley's Method}}
[[Category:Root-finding algorithms]]

Revision as of 14:17, 23 June 2013

30 year-old Entertainer or Range Artist Wesley from Drumheller, really loves vehicle, property developers properties for sale in singapore singapore and horse racing. Finds inspiration by traveling to Works of Antoni Gaudí. In numerical analysis, Halley’s method is a root-finding algorithm used for functions of one real variable with a continuous second derivative, i.e., C2 functions. It is named after its inventor Edmond Halley.

The algorithm is second in the class of Householder's methods, right after Newton's method. Like the latter, it produces iteratively a sequence of approximations to the root; their rate of convergence to the root is cubic. Multidimensional versions of this method exist.

Halley's method can be viewed as exactly finding the roots of a linear-over-linear Padé approximation to the function, in contrast to Newton's method/Secant method (approximates/interpolates the function linearly) or Cauchy's method/Muller's method (approximates/interpolates the function quadratically).[1]

Method

Halley’s method is a numerical algorithm for solving the nonlinear equation ƒ(x) = 0. In this case, the function ƒ has to be a function of one real variable. The method consists of a sequence of iterations:

beginning with an initial guess x0.

If ƒ is a three times continuously differentiable function and a is a zero of ƒ but not of its derivative, then, in a neighborhood of a, the iterates xn satisfy:

This means that the iterates converge to the zero if the initial guess is sufficiently close, and that the convergence is cubic.

The following alternative formulation shows the similarity between Halley’s method and Newton’s method. The expression is computed only once, and it is particularly useful when can be simplified.

A further alternative is as below, in which case the technique is sometimes referred to as Bailey’s method.[2]

Using any variation, when the second derivative is very close to zero, the iteration is almost the same as under Newton’s method.

Derivation

Consider the function

Any root of ƒ which is not a root of its derivative is a root of g; and any root of g is a root of ƒ. Applying Newton's method to g gives

with

and the result follows. Notice that if ƒ'(c) = 0, then one cannot apply this at c because g(c) would be undefined.

Cubic convergence

Suppose a is a root of f but not of its derivative. And suppose that the third derivative of f exists and is continuous in a neighborhood of a and xn is in that neighborhood. Then Taylor's theorem implies:

and also

where ξ and η are numbers lying between a and xn. Multiply the first equation by and subtract from it the second equation times to give:

Canceling and re-organizing terms yields:

Put the second term on the left side and divide through by to get:

Thus:

The limit of the coefficient on the right side as xn approaches a is:

If we take K to be a little larger than the absolute value of this, we can take absolute values of both sides of the formula and replace the absolute value of coefficient by its upper bound near a to get:

which is what was to be proved.

To summarize, ,

where .

References

Notes

  1. Template:Cite doi
  2. See for example the Bond Exchange of South Africa’s Bond Pricing Formula Specifications, where Bailey’s method is suggested when solving for a bond’s Yield to maturity.

Sources

  • T.R. Scavo and J.B. Thoo, On the geometry of Halley’s method. American Mathematical Monthly, 102:5 (1995), pp. 417–426. [1]
  • This article began as a translation from the equivalent article in French Wikipedia, retrieved 22 January 2007.

External links



  • I had like 17 domains hosted on single account, and never had any special troubles. If you are not happy with the service you will get your money back with in 45 days, that's guaranteed. But the Search Engine utility inside the Hostgator account furnished an instant score for my launched website. Fantastico is unable to install WordPress in a directory which already have any file i.e to install WordPress using Fantastico the destination directory must be empty and it should not have any previous installation files. When you share great information, others will take note. Once your hosting is purchased, you will need to setup your domain name to point to your hosting. Money Back: All accounts of Hostgator come with a 45 day money back guarantee. If you have any queries relating to where by and how to use Hostgator Discount Coupon, you can make contact with us at our site. If you are starting up a website or don't have too much website traffic coming your way, a shared plan is more than enough. Condition you want to take advantage of the worldwide web you prerequisite a HostGator web page, -1 of the most trusted and unfailing web suppliers on the world wide web today. Since, single server is shared by 700 to 800 websites, you cannot expect much speed.



    Hostgator tutorials on how to install Wordpress need not be complicated, especially when you will be dealing with a web hosting service that is friendly for novice webmasters and a blogging platform that is as intuitive as riding a bike. After that you can get Hostgator to host your domain and use the wordpress to do the blogging. Once you start site flipping, trust me you will not be able to stop. I cut my webmaster teeth on Control Panel many years ago, but since had left for other hosting companies with more commercial (cough, cough) interfaces. If you don't like it, you can chalk it up to experience and go on. First, find a good starter template design. When I signed up, I did a search for current "HostGator codes" on the web, which enabled me to receive a one-word entry for a discount. Your posts, comments, and pictures will all be imported into your new WordPress blog.
  • Halley's Method by John H. Mathews
  • Halley's Method by P. J. Acklam
  • Newton's method and high order iterations, Pascal Sebah and Xavier Gourdon, 2001 (the site has a link to a Postscript version for better formula display)