PolylogarithmIn mathematics, the polylogarithm (also known as Jonquière's function, for Alfred Jonquière) is a special function Li_s(z) of order s and argument z. Only for special values of s does the polylogarithm reduce to an elementary function such as the natural logarithm or a rational function. In quantum statistics, the polylogarithm function appears as the closed form of integrals of the Fermi–Dirac distribution and the Bose–Einstein distribution, and is also known as the Fermi–Dirac integral or the Bose–Einstein integral.
Binomial seriesIn mathematics, the binomial series is a generalization of the polynomial that comes from a binomial formula expression like for a nonnegative integer . Specifically, the binomial series is the Taylor series for the function centered at , where and . Explicitly, where the power series on the right-hand side of () is expressed in terms of the (generalized) binomial coefficients If α is a nonnegative integer n, then the (n + 2)th term and all later terms in the series are 0, since each contains a factor (n − n); thus in this case the series is finite and gives the algebraic binomial formula.
Newton polynomialIn the mathematical field of numerical analysis, a Newton polynomial, named after its inventor Isaac Newton, is an interpolation polynomial for a given set of data points. The Newton polynomial is sometimes called Newton's divided differences interpolation polynomial because the coefficients of the polynomial are calculated using Newton's divided differences method. Given a set of k + 1 data points where no two xj are the same, the Newton interpolation polynomial is a linear combination of Newton basis polynomials with the Newton basis polynomials defined as for j > 0 and .
Madhava seriesIn mathematics, a Madhava series is one of the three Taylor series expansions for the sine, cosine, and arctangent functions discovered in 14th or 15th century Kerala by the mathematician and astronomer Madhava of Sangamagrama (c. 1350 – c. 1425) or his followers in the Kerala school of astronomy and mathematics. Using modern notation, these series are: All three series were later independently discovered in 17th century Europe.
Logarithm of a matrixIn mathematics, a logarithm of a matrix is another matrix such that the matrix exponential of the latter matrix equals the original matrix. It is thus a generalization of the scalar logarithm and in some sense an inverse function of the matrix exponential. Not all matrices have a logarithm and those matrices that do have a logarithm may have more than one logarithm. The study of logarithms of matrices leads to Lie theory since when a matrix has a logarithm then it is in an element of a Lie group and the logarithm is the corresponding element of the vector space of the Lie algebra.
Non-analytic smooth functionIn mathematics, smooth functions (also called infinitely differentiable functions) and analytic functions are two very important types of functions. One can easily prove that any analytic function of a real argument is smooth. The converse is not true, as demonstrated with the counterexample below. One of the most important applications of smooth functions with compact support is the construction of so-called mollifiers, which are important in theories of generalized functions, such as Laurent Schwartz's theory of distributions.
VersineThe versine or versed sine is a trigonometric function found in some of the earliest (Sanskrit Aryabhatia, Section I) trigonometric tables. The versine of an angle is 1 minus its cosine. There are several related functions, most notably the coversine and haversine. The latter, half a versine, is of particular importance in the haversine formula of navigation. The versine or versed sine is a trigonometric function already appearing in some of the earliest trigonometric tables.
Power ruleIn calculus, the power rule is used to differentiate functions of the form , whenever is a real number. Since differentiation is a linear operation on the space of differentiable functions, polynomials can also be differentiated using this rule. The power rule underlies the Taylor series as it relates a power series with a function's derivatives. Let be a function satisfying for all , where . Then, The power rule for integration states that for any real number . It can be derived by inverting the power rule for differentiation.
Rate of convergenceIn numerical analysis, the order of convergence and the rate of convergence of a convergent sequence are quantities that represent how quickly the sequence approaches its limit. A sequence that converges to is said to have order of convergence and rate of convergence if The rate of convergence is also called the asymptotic error constant. Note that this terminology is not standardized and some authors will use rate where this article uses order (e.g., ).
Abel's theoremIn mathematics, Abel's theorem for power series relates a limit of a power series to the sum of its coefficients. It is named after Norwegian mathematician Niels Henrik Abel. Let the Taylor series be a power series with real coefficients with radius of convergence Suppose that the series converges. Then is continuous from the left at that is, The same theorem holds for complex power series provided that entirely within a single Stolz sector, that is, a region of the open unit disk where for some fixed finite .