Rank (linear algebra)In linear algebra, the rank of a matrix A is the dimension of the vector space generated (or spanned) by its columns. This corresponds to the maximal number of linearly independent columns of A. This, in turn, is identical to the dimension of the vector space spanned by its rows. Rank is thus a measure of the "nondegenerateness" of the system of linear equations and linear transformation encoded by A. There are multiple equivalent definitions of rank. A matrix's rank is one of its most fundamental characteristics.
Rank–nullity theoremThe rank–nullity theorem is a theorem in linear algebra, which asserts: the number of columns of a matrix M is the sum of the rank of M and the nullity of M; and the dimension of the domain of a linear transformation f is the sum of the rank of f (the dimension of the of f) and the nullity of f (the dimension of the kernel of f). It follows that for linear transformations of vector spaces of finite dimension, either injectivity or surjectivity implies bijectivity.
Adjugate matrixIn linear algebra, the adjugate or classical adjoint of a square matrix A is the transpose of its cofactor matrix and is denoted by adj(A). It is also occasionally known as adjunct matrix, or "adjoint", though the latter term today normally refers to a different concept, the adjoint operator which for a matrix is the conjugate transpose. The product of a matrix with its adjugate gives a diagonal matrix (entries not on the main diagonal are zero) whose diagonal entries are the determinant of the original matrix: where I is the identity matrix of the same size as A.
Antilinear mapIn mathematics, a function between two complex vector spaces is said to be antilinear or conjugate-linear if hold for all vectors and every complex number where denotes the complex conjugate of Antilinear maps stand in contrast to linear maps, which are additive maps that are homogeneous rather than conjugate homogeneous. If the vector spaces are real then antilinearity is the same as linearity.
Symmetric bilinear formIn mathematics, a symmetric bilinear form on a vector space is a bilinear map from two copies of the vector space to the field of scalars such that the order of the two vectors does not affect the value of the map. In other words, it is a bilinear function that maps every pair of elements of the vector space to the underlying field such that for every and in . They are also referred to more briefly as just symmetric forms when "bilinear" is understood.
Multilinear formIn abstract algebra and multilinear algebra, a multilinear form on a vector space over a field is a map that is separately -linear in each of its arguments. More generally, one can define multilinear forms on a module over a commutative ring. The rest of this article, however, will only consider multilinear forms on finite-dimensional vector spaces. A multilinear -form on over is called a (covariant) -tensor, and the vector space of such forms is usually denoted or .
Gram matrixIn linear algebra, the Gram matrix (or Gramian matrix, Gramian) of a set of vectors in an inner product space is the Hermitian matrix of inner products, whose entries are given by the inner product . If the vectors are the columns of matrix then the Gram matrix is in the general case that the vector coordinates are complex numbers, which simplifies to for the case that the vector coordinates are real numbers. An important application is to compute linear independence: a set of vectors are linearly independent if and only if the Gram determinant (the determinant of the Gram matrix) is non-zero.