Orthogonal basisIn mathematics, particularly linear algebra, an orthogonal basis for an inner product space is a basis for whose vectors are mutually orthogonal. If the vectors of an orthogonal basis are normalized, the resulting basis is an orthonormal basis. Any orthogonal basis can be used to define a system of orthogonal coordinates Orthogonal (not necessarily orthonormal) bases are important due to their appearance from curvilinear orthogonal coordinates in Euclidean spaces, as well as in Riemannian and pseudo-Riemannian manifolds.
System of linear equationsIn mathematics, a system of linear equations (or linear system) is a collection of one or more linear equations involving the same variables. For example, is a system of three equations in the three variables x, y, z. A solution to a linear system is an assignment of values to the variables such that all the equations are simultaneously satisfied. A solution to the system above is given by the ordered triple since it makes all three equations valid. The word "system" indicates that the equations should be considered collectively, rather than individually.
Additive Schwarz methodIn mathematics, the additive Schwarz method, named after Hermann Schwarz, solves a boundary value problem for a partial differential equation approximately by splitting it into boundary value problems on smaller domains and adding the results. Partial differential equations (PDEs) are used in all sciences to model phenomena. For the purpose of exposition, we give an example physical problem and the accompanying boundary value problem (BVP). Even if the reader is unfamiliar with the notation, the purpose is merely to show what a BVP looks like when written down.
Covariance and contravariance of vectorsIn physics, especially in multilinear algebra and tensor analysis, covariance and contravariance describe how the quantitative description of certain geometric or physical entities changes with a change of basis. In modern mathematical notation, the role is sometimes swapped. A simple illustrative case is that of a vector. For a vector, once a set of basis vectors has been defined, then the components of that vector will always vary opposite to that of the basis vectors. A vector is therefore a contravariant tensor.
Symplectic matrixIn mathematics, a symplectic matrix is a matrix with real entries that satisfies the condition where denotes the transpose of and is a fixed nonsingular, skew-symmetric matrix. This definition can be extended to matrices with entries in other fields, such as the complex numbers, finite fields, p-adic numbers, and function fields. Typically is chosen to be the block matrix where is the identity matrix. The matrix has determinant and its inverse is .
Natural deductionIn logic and proof theory, natural deduction is a kind of proof calculus in which logical reasoning is expressed by inference rules closely related to the "natural" way of reasoning. This contrasts with Hilbert-style systems, which instead use axioms as much as possible to express the logical laws of deductive reasoning. Natural deduction grew out of a context of dissatisfaction with the axiomatizations of deductive reasoning common to the systems of Hilbert, Frege, and Russell (see, e.g., Hilbert system).
One-parameter groupIn mathematics, a one-parameter group or one-parameter subgroup usually means a continuous group homomorphism from the real line (as an additive group) to some other topological group . If is injective then , the image, will be a subgroup of that is isomorphic to as an additive group. One-parameter groups were introduced by Sophus Lie in 1893 to define infinitesimal transformations. According to Lie, an infinitesimal transformation is an infinitely small transformation of the one-parameter group that it generates.
Kolmogorov–Arnold–Moser theoremThe Kolmogorov–Arnold–Moser (KAM) theorem is a result in dynamical systems about the persistence of quasiperiodic motions under small perturbations. The theorem partly resolves the small-divisor problem that arises in the perturbation theory of classical mechanics. The problem is whether or not a small perturbation of a conservative dynamical system results in a lasting quasiperiodic orbit. The original breakthrough to this problem was given by Andrey Kolmogorov in 1954.
Stone's theorem on one-parameter unitary groupsIn mathematics, Stone's theorem on one-parameter unitary groups is a basic theorem of functional analysis that establishes a one-to-one correspondence between self-adjoint operators on a Hilbert space and one-parameter families of unitary operators that are strongly continuous, i.e., and are homomorphisms, i.e., Such one-parameter families are ordinarily referred to as strongly continuous one-parameter unitary groups.