Skip to content

Joint Distribution

We define the joint distribution to study the behavior of multiple random variables simultaneously.

Independence can be characterized purely in terms of distribution functions.

This generalizes to nn random variables X1,,XnX_1, \dots, X_n. They are mutually independent if and only if:

P(X1x1,,Xnxn)=i=1nP(Xixi)\mathbb{P}(X_1 \le x_1, \dots, X_n \le x_n) = \prod_{i=1}^n \mathbb{P}(X_i \le x_i)

for all xiRx_i \in \mathbb{R}.

If we define a measure η\eta on R2\mathbb{R}^2 by the distribution of the random vector (X,Y)(X, Y):

η((,x1]×(,x2])=P(Xx1,Yx2)\eta((-\infty, x_1] \times (-\infty, x_2]) = \mathbb{P}(X \le x_1, Y \le x_2)

If XYX \perp Y, this factorizes:

η((,x1]×(,x2])=μ((,x1])ν((,x2])\eta((-\infty, x_1] \times (-\infty, x_2]) = \mu((-\infty, x_1]) \cdot \nu((-\infty, x_2])

where μ\mu and ν\nu are the distributions (laws) of XX and YY respectively.

This implies that η=μ×ν\eta = \mu \times \nu is a product measure, satisfying η(A×B)=μ(A)ν(B)\eta(A \times B) = \mu(A)\nu(B) for Borel sets A,BA, B.

The product structure allows us to compute expectations of functions of independent variables as iterated integrals.