Mutual informationIn probability theory and information theory, the mutual information (MI) of two random variables is a measure of the mutual dependence between the two variables. More specifically, it quantifies the "amount of information" (in units such as shannons (bits), nats or hartleys) obtained about one random variable by observing the other random variable. The concept of mutual information is intimately linked to that of entropy of a random variable, a fundamental notion in information theory that quantifies the expected "amount of information" held in a random variable.
Kumada couplingIn organic chemistry, the Kumada coupling is a type of cross coupling reaction, useful for generating carbon–carbon bonds by the reaction of a Grignard reagent and an organic halide. The procedure uses transition metal catalysts, typically nickel or palladium, to couple a combination of two alkyl, aryl or vinyl groups. The groups of Robert Corriu and Makoto Kumada reported the reaction independently in 1972. The reaction is notable for being among the first reported catalytic cross-coupling methods.
Interaction informationThe interaction information is a generalization of the mutual information for more than two variables. There are many names for interaction information, including amount of information, information correlation, co-information, and simply mutual information. Interaction information expresses the amount of information (redundancy or synergy) bound up in a set of variables, beyond that which is present in any subset of those variables. Unlike the mutual information, the interaction information can be either positive or negative.
Conditional mutual informationIn probability theory, particularly information theory, the conditional mutual information is, in its most basic form, the expected value of the mutual information of two random variables given the value of a third. For random variables , , and with support sets , and , we define the conditional mutual information as This may be written in terms of the expectation operator: . Thus is the expected (with respect to ) Kullback–Leibler divergence from the conditional joint distribution to the product of the conditional marginals and .
Variation of informationIn probability theory and information theory, the variation of information or shared information distance is a measure of the distance between two clusterings (partitions of elements). It is closely related to mutual information; indeed, it is a simple linear expression involving the mutual information. Unlike the mutual information, however, the variation of information is a true metric, in that it obeys the triangle inequality. Suppose we have two partitions and of a set into disjoint subsets, namely and .
Information theoryInformation theory is the mathematical study of the quantification, storage, and communication of information. The field was originally established by the works of Harry Nyquist and Ralph Hartley, in the 1920s, and Claude Shannon in the 1940s. The field, in applied mathematics, is at the intersection of probability theory, statistics, computer science, statistical mechanics, information engineering, and electrical engineering. A key measure in information theory is entropy.
Cross-coupling reactionIn organic chemistry, a cross-coupling reaction is a reaction where two different fragments are joined. Cross-couplings are a subset of the more general coupling reactions. Often cross-coupling reactions require metal catalysts. One important reaction type is this: (R, R' = organic fragments, usually aryle; M = main group center such as Li or MgX; X = halide) These reactions are used to form carbon–carbon bonds but also carbon-heteroatom bonds. Cross-coupling reaction are a subset of coupling reactions.
Hiyama couplingThe Hiyama coupling is a palladium-catalyzed cross-coupling reaction of organosilanes with organic halides used in organic chemistry to form carbon–carbon bonds (C-C bonds). This reaction was discovered in 1988 by Tamejiro Hiyama and Yasuo Hatanaka as a method to form carbon-carbon bonds synthetically with chemo- and regioselectivity. The Hiyama coupling has been applied to the synthesis of various natural products.
Human impact on the environmentHuman impact on the environment (or anthropogenic environmental impact) refers to changes to biophysical environments and to ecosystems, biodiversity, and natural resources caused directly or indirectly by humans. Modifying the environment to fit the needs of society (as in the built environment) is causing severe effects including global warming, environmental degradation (such as ocean acidification), mass extinction and biodiversity loss, ecological crisis, and ecological collapse.
Internal auditInternal auditing is an independent, objective assurance and consulting activity designed to add value and improve an organization's operations. It helps an organization accomplish its objectives by bringing a systematic, disciplined approach to evaluate and improve the effectiveness of risk management, control and governance processes. Internal auditing might achieve this goal by providing insight and recommendations based on analyses and assessments of data and business processes.