In probability theory and statistics, the hypergeometric distribution is a discrete probability distribution that describes the probability of successes (random draws for which the object drawn has a specified feature) in draws, without replacement, from a finite population of size that contains exactly objects with that feature, wherein each draw is either a success or a failure. In contrast, the binomial distribution describes the probability of successes in draws with replacement. The following conditions characterize the hypergeometric distribution: The result of each draw (the elements of the population being sampled) can be classified into one of two mutually exclusive categories (e.g. Pass/Fail or Employed/Unemployed). The probability of a success changes on each draw, as each draw decreases the population (sampling without replacement from a finite population). A random variable follows the hypergeometric distribution if its probability mass function (pmf) is given by where is the population size, is the number of success states in the population, is the number of draws (i.e. quantity drawn in each trial), is the number of observed successes, is a binomial coefficient. The () is positive when . A random variable distributed hypergeometrically with parameters , and is written and has probability mass function above. As required, we have which essentially follows from Vandermonde's identity from combinatorics. Also note that This identity can be shown by expressing the binomial coefficients in terms of factorials and rearranging the latter, but it also follows from the symmetry of the problem. Indeed, consider two rounds of drawing without replacement. In the first round, out of neutral marbles are drawn from an urn without replacement and coloured green. Then the colored marbles are put back. In the second round, marbles are drawn without replacement and colored red. Then, the number of marbles with both colors on them (that is, the number of marbles that have been drawn twice) has the hypergeometric distribution.

About this result
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.
Related courses (3)
MATH-232: Probability and statistics
A basic course in probability and statistics
CS-101: Advanced information, computation, communication I
Discrete mathematics is a discipline with applications to almost all areas of study. It provides a set of indispensable tools to computer science in particular. This course reviews (familiar) topics a
COM-406: Foundations of Data Science
We discuss a set of topics that are important for the understanding of modern data science but that are typically not taught in an introductory ML course. In particular we discuss fundamental ideas an
Related lectures (31)
Estimating R: Uniform, Hypergeometric, Poisson
Covers the estimation of R through balanced dice, uniform, hypergeometric, and Poisson distributions.
Counting: Bit Strings and Committees
Explores counting bit strings, committees, balls distribution, poker hands, and coefficients, with examples of the Pigeonhole Principle and card selection.
Probability Theory: Midterm Solutions
Covers the solutions to the midterm exam of a Probability Theory course, including calculations of probabilities and expectations.
Show more
Related publications (15)

Beyond the ultradeep frontier fields and legacy observations (BUFFALO): a high-resolution strong plus weak-lensing view of Abell 370

David Richard Harvey, Mathilde Jauzac

The HST treasury programme BUFFALO provides extended wide-field imaging of the six Hubble Frontier Fields galaxy clusters. Here we present the combined strong and weak-lensing analysis of Abell 370, a massive cluster at z = 0.375. From the reconstructed to ...
OXFORD UNIV PRESS2023

VL4Pose: Active Learning Through Out-Of-Distribution Detection For Pose Estimation

Alexandre Massoud Alahi, Megh Hiren Shukla

Advances in computing have enabled widespread access to pose estimation, creating new sources of data streams. Unlike mock set-ups for data collection, tapping into these data streams through on-device active learning allows us to directly sample from the ...
BMVA Press2022

Intrinsic area near the origin for self-similar growth-fragmentations and related random surfaces

We study the behaviour of a natural measure defined on the leaves of the genealogical tree of some branching processes, namely self-similar growth-fragmentation processes. Each particle, or cell, is attributed a positive mass that evolves in continuous tim ...
2022
Show more
Related concepts (16)
Beta-binomial distribution
In probability theory and statistics, the beta-binomial distribution is a family of discrete probability distributions on a finite support of non-negative integers arising when the probability of success in each of a fixed or known number of Bernoulli trials is either unknown or random. The beta-binomial distribution is the binomial distribution in which the probability of success at each of n trials is not fixed but randomly drawn from a beta distribution.
Conjugate prior
In Bayesian probability theory, if the posterior distribution is in the same probability distribution family as the prior probability distribution , the prior and posterior are then called conjugate distributions, and the prior is called a conjugate prior for the likelihood function . A conjugate prior is an algebraic convenience, giving a closed-form expression for the posterior; otherwise, numerical integration may be necessary. Further, conjugate priors may give intuition by more transparently showing how a likelihood function updates a prior distribution.
Multinomial distribution
In probability theory, the multinomial distribution is a generalization of the binomial distribution. For example, it models the probability of counts for each side of a k-sided dice rolled n times. For n independent trials each of which leads to a success for exactly one of k categories, with each category having a given fixed success probability, the multinomial distribution gives the probability of any particular combination of numbers of successes for the various categories.
Show more

Graph Chatbot

Chat with Graph Search

Ask any question about EPFL courses, lectures, exercises, research, news, etc. or try the example questions below.

DISCLAIMER: The Graph Chatbot is not programmed to provide explicit or categorical answers to your questions. Rather, it transforms your questions into API requests that are distributed across the various IT services officially administered by EPFL. Its purpose is solely to collect and recommend relevant references to content that you can explore to help you answer your questions.