In probability theory, calculation of the sum of normally distributed random variables is an instance of the arithmetic of random variables.
This is not to be confused with the sum of normal distributions which forms a mixture distribution.
Let X and Y be independent random variables that are normally distributed (and therefore also jointly so), then their sum is also normally distributed. i.e., if
then
This means that the sum of two independent normally distributed random variables is normal, with its mean being the sum of the two means, and its variance being the sum of the two variances (i.e., the square of the standard deviation is the sum of the squares of the standard deviations).
In order for this result to hold, the assumption that X and Y are independent cannot be dropped, although it can be weakened to the assumption that X and Y are jointly, rather than separately, normally distributed. (See here for an example.)
The result about the mean holds in all cases, while the result for the variance requires uncorrelatedness, but not independence.
The characteristic function
of the sum of two independent random variables X and Y is just the product of the two separate characteristic functions:
of X and Y.
The characteristic function of the normal distribution with expected value μ and variance σ2 is
So
This is the characteristic function of the normal distribution with expected value and variance
Finally, recall that no two distinct distributions can both have the same characteristic function, so the distribution of X + Y must be just this normal distribution.
For independent random variables X and Y, the distribution fZ of Z = X + Y equals the convolution of fX and fY:
Given that fX and fY are normal densities,
Substituting into the convolution:
Defining , and completing the square:
The expression in the integral is a normal density distribution on x, and so the integral evaluates to 1. The desired result follows:
It can be shown that the Fourier transform of a Gaussian, , is
By the convolution theorem:
First consider the normalized case when X, Y ~ N(0, 1), so that their PDFs are
and
Let Z = X + Y.
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.
A product distribution is a probability distribution constructed as the distribution of the product of random variables having two other known distributions. Given two statistically independent random variables X and Y, the distribution of the random variable Z that is formed as the product is a product distribution. The product distribution is the PDF of the product of sample values. This is not the same as the product of their PDF's yet the concepts are often ambiguously termed as "product of Gaussians".
In probability and statistics, a mixture distribution is the probability distribution of a random variable that is derived from a collection of other random variables as follows: first, a random variable is selected by chance from the collection according to given probabilities of selection, and then the value of the selected random variable is realized. The underlying random variables may be random real numbers, or they may be random vectors (each having the same dimension), in which case the mixture distribution is a multivariate distribution.
Wyner's common information is a measure that quantifies and assesses the commonality between two random variables. Based on this, we introduce a novel two-step procedure to construct features from data, referred to as Common Information Components Analysis ...
Given two jointly distributed random variables (X,Y), a functional representation of X is a random variable Z independent of Y, and a deterministic function g(⋅,⋅) such that X=g(Y,Z). The problem of finding a minimum entropy functional representation is kn ...
Consider a linear elliptic PDE defined over a stochastic stochastic geometry a function of N random variables. In many application, quantify the uncertainty propagated to a quantity of interest (QoI) is an important problem. The random domain is split into ...