Laplace transformIn mathematics, the 'Laplace transform, named after its discoverer Pierre-Simon Laplace (ləˈplɑ:s), is an integral transform that converts a function of a real variable (usually , in the time domain) to a function of a complex variable (in the complex frequency domain, also known as s-domain', or s-plane). The transform has many applications in science and engineering because it is a tool for solving differential equations. In particular, it transforms ordinary differential equations into algebraic equations and convolution into multiplication.
Bessel functionBessel functions, first defined by the mathematician Daniel Bernoulli and then generalized by Friedrich Bessel, are canonical solutions y(x) of Bessel's differential equation for an arbitrary complex number , which represents the order of the Bessel function. Although and produce the same differential equation, it is conventional to define different Bessel functions for these two values in such a way that the Bessel functions are mostly smooth functions of . The most important cases are when is an integer or half-integer.
Riemann hypothesisIn mathematics, the Riemann hypothesis is the conjecture that the Riemann zeta function has its zeros only at the negative even integers and complex numbers with real part 1/2. Many consider it to be the most important unsolved problem in pure mathematics. It is of great interest in number theory because it implies results about the distribution of prime numbers. It was proposed by , after whom it is named.
Airy functionIn the physical sciences, the Airy function (or Airy function of the first kind) Ai(x) is a special function named after the British astronomer George Biddell Airy (1801–1892). The function Ai(x) and the related function Bi(x), are linearly independent solutions to the differential equation known as the Airy equation or the Stokes equation. This is the simplest second-order linear differential equation with a turning point (a point where the character of the solutions changes from oscillatory to exponential).
Sinc functionIn mathematics, physics and engineering, the sinc function, denoted by sinc(x), has two forms, normalized and unnormalized. In mathematics, the historical unnormalized sinc function is defined for x ≠ 0 by Alternatively, the unnormalized sinc function is often called the sampling function, indicated as Sa(x). In digital signal processing and information theory, the normalized sinc function is commonly defined for x ≠ 0 by In either case, the value at x = 0 is defined to be the limiting value for all real a ≠ 0 (the limit can be proven using the squeeze theorem).
Prime-counting functionIn mathematics, the prime-counting function is the function counting the number of prime numbers less than or equal to some real number x. It is denoted by pi(x) (unrelated to the number pi). Prime number theorem Of great interest in number theory is the growth rate of the prime-counting function. It was conjectured in the end of the 18th century by Gauss and by Legendre to be approximately where log is the natural logarithm, in the sense that This statement is the prime number theorem.
OscillationOscillation is the repetitive or periodic variation, typically in time, of some measure about a central value (often a point of equilibrium) or between two or more different states. Familiar examples of oscillation include a swinging pendulum and alternating current. Oscillations can be used in physics to approximate complex interactions, such as those between atoms.
SummationIn mathematics, summation is the addition of a sequence of any kind of numbers, called addends or summands; the result is their sum or total. Beside numbers, other types of values can be summed as well: functions, vectors, matrices, polynomials and, in general, elements of any type of mathematical objects on which an operation denoted "+" is defined. Summations of infinite sequences are called series. They involve the concept of limit, and are not considered in this article.
Root mean squareIn mathematics and its applications, the root mean square of a set of numbers (abbreviated as RMS, or rms and denoted in formulas as either or ) is defined as the square root of the mean square (the arithmetic mean of the squares) of the set. The RMS is also known as the quadratic mean (denoted ) and is a particular case of the generalized mean. The RMS of a continuously varying function (denoted ) can be defined in terms of an integral of the squares of the instantaneous values during a cycle.
Exponential sumIn mathematics, an exponential sum may be a finite Fourier series (i.e. a trigonometric polynomial), or other finite sum formed using the exponential function, usually expressed by means of the function Therefore, a typical exponential sum may take the form summed over a finite sequence of real numbers xn. If we allow some real coefficients an, to get the form it is the same as allowing exponents that are complex numbers. Both forms are certainly useful in applications.