Projective line over a ringIn mathematics, the projective line over a ring is an extension of the concept of projective line over a field. Given a ring A with 1, the projective line P(A) over A consists of points identified by projective coordinates. Let U be the group of units of A; pairs (a, b) and (c, d) from A × A are related when there is a u in U such that ua = c and ub = d. This relation is an equivalence relation. A typical equivalence class is written U[a, b]. P(A) = { U[a, b] : aA + bA = A }, that is, U[a, b] is in the projective line if the ideal generated by a and b is all of A.
Linear fractional transformationIn mathematics, a linear fractional transformation is, roughly speaking, an invertible transformation of the form The precise definition depends on the nature of a, b, c, d, and z. In other words, a linear fractional transformation is a transformation that is represented by a fraction whose numerator and denominator are linear. In the most basic setting, a, b, c, d, and z are complex numbers (in which case the transformation is also called a Möbius transformation), or more generally elements of a field.
HomographyIn projective geometry, a homography is an isomorphism of projective spaces, induced by an isomorphism of the vector spaces from which the projective spaces derive. It is a bijection that maps lines to lines, and thus a collineation. In general, some collineations are not homographies, but the fundamental theorem of projective geometry asserts that is not so in the case of real projective spaces of dimension at least two. Synonyms include projectivity, projective transformation, and projective collineation.
Hilbert spaceIn mathematics, Hilbert spaces (named after David Hilbert) allow the methods of linear algebra and calculus to be generalized from (finite-dimensional) Euclidean vector spaces to spaces that may be infinite-dimensional. Hilbert spaces arise naturally and frequently in mathematics and physics, typically as function spaces. Formally, a Hilbert space is a vector space equipped with an inner product that induces a distance function for which the space is a complete metric space.
Matrix (mathematics)In mathematics, a matrix (plural matrices) is a rectangular array or table of numbers, symbols, or expressions, arranged in rows and columns, which is used to represent a mathematical object or a property of such an object. For example, is a matrix with two rows and three columns. This is often referred to as a "two by three matrix", a " matrix", or a matrix of dimension . Without further specifications, matrices represent linear maps, and allow explicit computations in linear algebra.
Eigenvalues and eigenvectorsIn linear algebra, an eigenvector (ˈaɪgənˌvɛktər) or characteristic vector of a linear transformation is a nonzero vector that changes at most by a constant factor when that linear transformation is applied to it. The corresponding eigenvalue, often represented by , is the multiplying factor. Geometrically, a transformation matrix rotates, stretches, or shears the vectors it acts upon. The eigenvectors for a linear transformation matrix are the set of vectors that are only stretched, with no rotation or shear.
Hyperbolic geometryIn mathematics, hyperbolic geometry (also called Lobachevskian geometry or Bolyai–Lobachevskian geometry) is a non-Euclidean geometry. The parallel postulate of Euclidean geometry is replaced with: For any given line R and point P not on R, in the plane containing both line R and point P there are at least two distinct lines through P that do not intersect R. (Compare the above with Playfair's axiom, the modern version of Euclid's parallel postulate.) The hyperbolic plane is a plane where every point is a saddle point.