PixelIn digital imaging, a pixel (abbreviated px), pel, or picture element is the smallest addressable element in a raster image, or the smallest addressable element in a dot matrix display device. In most digital display devices, pixels are the smallest element that can be manipulated through software. Each pixel is a sample of an original or synthetic image; more samples typically provide more accurate representations of the original. The intensity of each pixel is variable.
Terahertz metamaterialA terahertz metamaterial is a class of composite metamaterials designed to interact at terahertz (THz) frequencies. The terahertz frequency range used in materials research is usually defined as 0.1 to 10 THz. This bandwidth is also known as the terahertz gap because it is noticeably underutilized. This is because terahertz waves are electromagnetic waves with frequencies higher than microwaves but lower than infrared radiation and visible light.
Cardinal point (optics)In Gaussian optics, the cardinal points consist of three pairs of points located on the optical axis of a rotationally symmetric, focal, optical system. These are the focal points, the principal points, and the nodal points. For ideal systems, the basic imaging properties such as image size, location, and orientation are completely determined by the locations of the cardinal points; in fact only four points are necessary: the focal points and either the principal or nodal points.
Metamaterial antennaMetamaterial antennas are a class of antennas which use metamaterials to increase performance of miniaturized (electrically small) antenna systems. Their purpose, as with any electromagnetic antenna, is to launch energy into free space. However, this class of antenna incorporates metamaterials, which are materials engineered with novel, often microscopic, structures to produce unusual physical properties. Antenna designs incorporating metamaterials can step-up the antenna's radiated power.
Spectral imagingSpectral imaging is imaging that uses multiple bands across the electromagnetic spectrum. While an ordinary camera captures light across three wavelength bands in the visible spectrum, red, green, and blue (RGB), spectral imaging encompasses a wide variety of techniques that go beyond RGB. Spectral imaging may use the infrared, the visible spectrum, the ultraviolet, x-rays, or some combination of the above.
Transformation opticsTransformation optics is a branch of optics which applies metamaterials to produce spatial variations, derived from coordinate transformations, which can direct chosen bandwidths of electromagnetic radiation. This can allow for the construction of new composite artificial devices, which probably could not exist without metamaterials and coordinate transformation. Computing power that became available in the late 1990s enables prescribed quantitative values for the permittivity and permeability, the constitutive parameters, which produce localized spatial variations.
Active-pixel sensorAn active-pixel sensor (APS) is an , which was invented by Peter J.W. Noble in 1968, where each pixel sensor unit cell has a photodetector (typically a pinned photodiode) and one or more active transistors. In a metal–oxide–semiconductor (MOS) active-pixel sensor, MOS field-effect transistors (MOSFETs) are used as amplifiers. There are different types of APS, including the early NMOS APS and the now much more common complementary MOS (CMOS) APS, also known as the CMOS sensor.
Multispectral imagingMultispectral imaging captures image data within specific wavelength ranges across the electromagnetic spectrum. The wavelengths may be separated by filters or detected with the use of instruments that are sensitive to particular wavelengths, including light from frequencies beyond the visible light range, i.e. infrared and ultra-violet. It can allow extraction of additional information the human eye fails to capture with its visible receptors for red, green and blue.
Image sensorAn image sensor or imager is a sensor that detects and conveys information used to form an . It does so by converting the variable attenuation of light waves (as they pass through or reflect off objects) into signals, small bursts of current that convey the information. The waves can be light or other electromagnetic radiation. Image sensors are used in electronic imaging devices of both analog and digital types, which include digital cameras, camera modules, camera phones, optical mouse devices, medical imaging equipment, night vision equipment such as thermal imaging devices, radar, sonar, and others.
Ray (optics)In optics, a ray is an idealized geometrical model of light or other electromagnetic radiation, obtained by choosing a curve that is perpendicular to the wavefronts of the actual light, and that points in the direction of energy flow. Rays are used to model the propagation of light through an optical system, by dividing the real light field up into discrete rays that can be computationally propagated through the system by the techniques of ray tracing. This allows even very complex optical systems to be analyzed mathematically or simulated by computer.