Spectrum (functional analysis)In mathematics, particularly in functional analysis, the spectrum of a bounded linear operator (or, more generally, an unbounded linear operator) is a generalisation of the set of eigenvalues of a matrix. Specifically, a complex number is said to be in the spectrum of a bounded linear operator if either has no set-theoretic inverse; or the set-theoretic inverse is either unbounded or defined on a non-dense subset. Here, is the identity operator. By the closed graph theorem, is in the spectrum if and only if the bounded operator is non-bijective on .
Operator theoryIn mathematics, operator theory is the study of linear operators on function spaces, beginning with differential operators and integral operators. The operators may be presented abstractly by their characteristics, such as bounded linear operators or closed operators, and consideration may be given to nonlinear operators. The study, which depends heavily on the topology of function spaces, is a branch of functional analysis. If a collection of operators forms an algebra over a field, then it is an operator algebra.
Fredholm theoryIn mathematics, Fredholm theory is a theory of integral equations. In the narrowest sense, Fredholm theory concerns itself with the solution of the Fredholm integral equation. In a broader sense, the abstract structure of Fredholm's theory is given in terms of the spectral theory of Fredholm operators and Fredholm kernels on Hilbert space. The theory is named in honour of Erik Ivar Fredholm. The following sections provide a casual sketch of the place of Fredholm theory in the broader context of operator theory and functional analysis.
Differential algebraIn mathematics, differential algebra is, broadly speaking, the area of mathematics consisting in the study of differential equations and differential operators as algebraic objects in view of deriving properties of differential equations and operators without computing the solutions, similarly as polynomial algebras are used for the study of algebraic varieties, which are solution sets of systems of polynomial equations. Weyl algebras and Lie algebras may be considered as belonging to differential algebra.
Runge–Kutta methodsIn numerical analysis, the Runge–Kutta methods (ˈrʊŋəˈkʊtɑː ) are a family of implicit and explicit iterative methods, which include the Euler method, used in temporal discretization for the approximate solutions of simultaneous nonlinear equations. These methods were developed around 1900 by the German mathematicians Carl Runge and Wilhelm Kutta. The most widely known member of the Runge–Kutta family is generally referred to as "RK4", the "classic Runge–Kutta method" or simply as "the Runge–Kutta method".
Elliptic operatorIn the theory of partial differential equations, elliptic operators are differential operators that generalize the Laplace operator. They are defined by the condition that the coefficients of the highest-order derivatives be positive, which implies the key property that the principal symbol is invertible, or equivalently that there are no real characteristic directions. Elliptic operators are typical of potential theory, and they appear frequently in electrostatics and continuum mechanics.
Normal matrixIn mathematics, a complex square matrix A is normal if it commutes with its conjugate transpose A^: The concept of normal matrices can be extended to normal operators on infinite dimensional normed spaces and to normal elements in C-algebras. As in the matrix case, normality means commutativity is preserved, to the extent possible, in the noncommutative setting. This makes normal operators, and normal elements of C*-algebras, more amenable to analysis.
Unbounded operatorIn mathematics, more specifically functional analysis and operator theory, the notion of unbounded operator provides an abstract framework for dealing with differential operators, unbounded observables in quantum mechanics, and other cases. The term "unbounded operator" can be misleading, since "unbounded" should sometimes be understood as "not necessarily bounded"; "operator" should be understood as "linear operator" (as in the case of "bounded operator"); the domain of the operator is a linear subspace, not necessarily the whole space; this linear subspace is not necessarily closed; often (but not always) it is assumed to be dense; in the special case of a bounded operator, still, the domain is usually assumed to be the whole space.
Stirling's approximationIn mathematics, Stirling's approximation (or Stirling's formula) is an approximation for factorials. It is a good approximation, leading to accurate results even for small values of . It is named after James Stirling, though a related but less precise result was first stated by Abraham de Moivre. One way of stating the approximation involves the logarithm of the factorial: where the big O notation means that, for all sufficiently large values of , the difference between and will be at most proportional to the logarithm.
Cousin problemsIn mathematics, the Cousin problems are two questions in several complex variables, concerning the existence of meromorphic functions that are specified in terms of local data. They were introduced in special cases by Pierre Cousin in 1895. They are now posed, and solved, for any complex manifold M, in terms of conditions on M. For both problems, an open cover of M by sets Ui is given, along with a meromorphic function fi on each Ui. The first Cousin problem or additive Cousin problem assumes that each difference is a holomorphic function, where it is defined.