Data assimilationData assimilation is a mathematical discipline that seeks to optimally combine theory (usually in the form of a numerical model) with observations. There may be a number of different goals sought – for example, to determine the optimal state estimate of a system, to determine initial conditions for a numerical forecast model, to interpolate sparse observation data using (e.g. physical) knowledge of the system being observed, to set numerical parameters based on training a model from observed data.
Cultural assimilationCultural assimilation is the process in which a minority group or culture comes to resemble a society's majority group or assimilate the values, behaviors, and beliefs of another group whether fully or partially. The different types of cultural assimilation include full assimilation and forced assimilation. Full assimilation is the more prevalent of the two, as it occurs spontaneously. When used as a political ideology, assimilationism refers to governmental policies of deliberately assimilating ethnic groups into the national culture.
Forced assimilationForced assimilation is the involuntary cultural assimilation of religious or ethnic minority groups, during which they are forced to adopt the language, national identity, norms, mores, customs, traditions, values, mentality, perceptions, way of life, and often the religion and ideology of an established and generally larger community belonging to dominant culture by a government. The enforced use of a dominant language in legislation, education, literature, and worship also counts as forced assimilation.
Kalman filterFor statistics and control theory, Kalman filtering, also known as linear quadratic estimation (LQE), is an algorithm that uses a series of measurements observed over time, including statistical noise and other inaccuracies, and produces estimates of unknown variables that tend to be more accurate than those based on a single measurement alone, by estimating a joint probability distribution over the variables for each timeframe. The filter is named after Rudolf E. Kálmán, who was one of the primary developers of its theory.
Nuisance parameterIn statistics, a nuisance parameter is any parameter which is unspecified but which must be accounted for in the hypothesis testing of the parameters which are of interest. The classic example of a nuisance parameter comes from the normal distribution, a member of the location–scale family. For at least one normal distribution, the variance(s), σ2 is often not specified or known, but one desires to hypothesis test on the mean(s).
Estimation theoryEstimation theory is a branch of statistics that deals with estimating the values of parameters based on measured empirical data that has a random component. The parameters describe an underlying physical setting in such a way that their value affects the distribution of the measured data. An estimator attempts to approximate the unknown parameters using the measurements.
Ensemble Kalman filterThe ensemble Kalman filter (EnKF) is a recursive filter suitable for problems with a large number of variables, such as discretizations of partial differential equations in geophysical models. The EnKF originated as a version of the Kalman filter for large problems (essentially, the covariance matrix is replaced by the sample covariance), and it is now an important data assimilation component of ensemble forecasting.
Nuclear weapon yieldThe explosive yield of a nuclear weapon is the amount of energy released such as blast, thermal, and nuclear radiation, when that particular nuclear weapon is detonated, usually expressed as a TNT equivalent (the standardized equivalent mass of trinitrotoluene which, if detonated, would produce the same energy discharge), either in kilotonnes (kt—thousands of tonnes of TNT), in megatonnes (Mt—millions of tonnes of TNT), or sometimes in terajoules (TJ). An explosive yield of one terajoule is equal to .
Likelihood functionIn statistical inference, the likelihood function quantifies the plausibility of parameter values characterizing a statistical model in light of observed data. Its most typical usage is to compare possible parameter values (under a fixed set of observations and a particular model), where higher values of likelihood are preferred because they correspond to more probable parameter values.
Constraint satisfaction problemConstraint satisfaction problems (CSPs) are mathematical questions defined as a set of objects whose state must satisfy a number of constraints or limitations. CSPs represent the entities in a problem as a homogeneous collection of finite constraints over variables, which is solved by constraint satisfaction methods. CSPs are the subject of research in both artificial intelligence and operations research, since the regularity in their formulation provides a common basis to analyze and solve problems of many seemingly unrelated families.