Taylor series

In mathematics, a Taylor series is a representation of a function as an infinite sum of terms that are calculated from the values of the function's derivatives at a single point. The concept of a Taylor series can be traced back to the 14th century Indian mathematician, Madhava of Sangamagrama, who is credited with the series for the sine, cosine, tangent, and arctangent functions by later (14th to 16th century) mathematicians of the Kerala School of Astronomy and Mathematics. In the West, the subject was formulated by the Scottish mathematician James Gregory and formally introduced by the English mathematician Brook Taylor in 1715. If the Taylor series is centered at zero, then that series is also called a Maclaurin series, after the Scottish mathematician Colin Maclaurin, who made extensive use of this special case of Taylor series in the 18th century. A function can be approximated by using a finite number of terms of its Taylor series. Taylor's theorem gives quantitative estimates on the error introduced by the use of such an approximation. The polynomial formed by taking some initial terms of the Taylor series is called a Taylor polynomial. The Taylor series of a function is the limit of that function's Taylor polynomials as the degree increases, provided that the limit exists. A function may not be equal to its Taylor series, even if its Taylor series converges at every point. A function that is equal to its Taylor series in an open interval (or a disc in the complex plane) is known as an analytic function in that interval.


This table shows the example usage of word lists for keywords extraction from the text above.

WordWord FrequencyNumber of ArticlesRelevance

This website uses cookies to ensure you get the best experience on our website. Learn more. Got it.