HyperparameterIn Bayesian statistics, a hyperparameter is a parameter of a prior distribution; the term is used to distinguish them from parameters of the model for the underlying system under analysis. For example, if one is using a beta distribution to model the distribution of the parameter p of a Bernoulli distribution, then: p is a parameter of the underlying system (Bernoulli distribution), and α and β are parameters of the prior distribution (beta distribution), hence hyperparameters.
Rational agentA rational agent or rational being is a person or entity that always aims to perform optimal actions based on given premises and information. A rational agent can be anything that makes decisions, typically a person, firm, machine, or software. The concept of rational agents can be found in various disciplines such as artificial intelligence, cognitive science, decision theory, economics, ethics, game theory, and the study of practical reason. In reference to economics, rational agent refers to hypothetical consumers and how they make decisions in a free market.
Variational Bayesian methodsVariational Bayesian methods are a family of techniques for approximating intractable integrals arising in Bayesian inference and machine learning. They are typically used in complex statistical models consisting of observed variables (usually termed "data") as well as unknown parameters and latent variables, with various sorts of relationships among the three types of random variables, as might be described by a graphical model. As typical in Bayesian inference, the parameters and latent variables are grouped together as "unobserved variables".
Misuse of statisticsStatistics, when used in a misleading fashion, can trick the casual observer into believing something other than what the data shows. That is, a misuse of statistics occurs when a statistical argument asserts a falsehood. In some cases, the misuse may be accidental. In others, it is purposeful and for the gain of the perpetrator. When the statistical reason involved is false or misapplied, this constitutes a statistical fallacy. The false statistics trap can be quite damaging for the quest for knowledge.
Moment problemIn mathematics, a moment problem arises as the result of trying to invert the mapping that takes a measure μ to the sequence of moments More generally, one may consider for an arbitrary sequence of functions Mn. In the classical setting, μ is a measure on the real line, and M is the sequence { xn : n = 0, 1, 2, ... }. In this form the question appears in probability theory, asking whether there is a probability measure having specified mean, variance and so on, and whether it is unique.
Geometric meanIn mathematics, the geometric mean is a mean or average which indicates a central tendency of a finite set of real numbers by using the product of their values (as opposed to the arithmetic mean which uses their sum). The geometric mean is defined as the nth root of the product of n numbers, i.e., for a set of numbers a1, a2, ..., an, the geometric mean is defined as or, equivalently, as the arithmetic mean in logscale: Most commonly the numbers are restricted to being non-negative, to avoid complications related to negative numbers not having real roots, and frequently they are restricted to being positive, to enable the use of logarithms.
Data dredgingData dredging (also known as data snooping or p-hacking) is the misuse of data analysis to find patterns in data that can be presented as statistically significant, thus dramatically increasing and understating the risk of false positives. This is done by performing many statistical tests on the data and only reporting those that come back with significant results.
Generalized meanIn mathematics, generalized means (or power mean or Hölder mean from Otto Hölder) are a family of functions for aggregating sets of numbers. These include as special cases the Pythagorean means (arithmetic, geometric, and harmonic means). If p is a non-zero real number, and are positive real numbers, then the generalized mean or power mean with exponent p of these positive real numbers is (See p-norm).
HistogramA histogram is an approximate representation of the distribution of numerical data. The term was first introduced by Karl Pearson. To construct a histogram, the first step is to "bin" (or "bucket") the range of values—that is, divide the entire range of values into a series of intervals—and then count how many values fall into each interval. The bins are usually specified as consecutive, non-overlapping intervals of a variable. The bins (intervals) must be adjacent and are often (but not required to be) of equal size.
Decision modelA decision model in decision theory is the starting point for a decision method within a formal (axiomatic) system. Decision models contain at least one action axiom. An action is in the form "IF is true, THEN do ". An action axiom tests a condition (antecedent) and, if the condition has been met, then (consequent) it suggests (mandates) an action: from knowledge to action. A decision model may also be a network of connected decisions, information and knowledge that represents a decision-making approach that can be used repeatedly (such as one developed using the Decision Model and Notation standard).