We’ve seen independence of events before. Now we formalize independence for random variables.
Definition
Two random variables X X X and Y Y Y are independent if the σ \sigma σ -fields they generate are independent , denoted σ ( X ) ⊥ σ ( Y ) \sigma(X) \perp \sigma(Y) σ ( X ) ⊥ σ ( Y ) .
This means that for any Borel sets A , B ∈ B A, B \in \mathcal{B} A , B ∈ B :
{ X ∈ A } ⊥ { Y ∈ B } \{X \in A\} \perp \{Y \in B\} { X ∈ A } ⊥ { Y ∈ B } Or equivalently:
P ( X ∈ A , Y ∈ B ) = P ( X ∈ A ) P ( Y ∈ B ) \mathbb{P}(X \in A, Y \in B) = \mathbb{P}(X \in A)\mathbb{P}(Y \in B) P ( X ∈ A , Y ∈ B ) = P ( X ∈ A ) P ( Y ∈ B )
More generally, X 1 , … , X n X_1, \dots, X_n X 1 , … , X n are mutually independent if σ ( X 1 ) , … , σ ( X n ) \sigma(X_1), \dots, \sigma(X_n) σ ( X 1 ) , … , σ ( X n ) are independent. This holds if and only if the following is true for all Borel sets A 1 , … , A n ∈ B A_1, \dots, A_n \in \mathcal{B} A 1 , … , A n ∈ B :
P ( X 1 ∈ A 1 , … , X n ∈ A n ) = ∏ i = 1 n P ( X i ∈ A i ) \mathbb{P}(X_1 \in A_1, \dots, X_n \in A_n) = \prod_{i=1}^n \mathbb{P}(X_i \in A_i) P ( X 1 ∈ A 1 , … , X n ∈ A n ) = i = 1 ∏ n P ( X i ∈ A i )
Remark
Recall that for events E 1 , … , E n E_1, \dots, E_n E 1 , … , E n , the condition P ( ∩ E i ) = ∏ P ( E i ) \mathbb{P}(\cap E_i) = \prod \mathbb{P}(E_i) P ( ∩ E i ) = ∏ P ( E i ) is in general not enough for mutual independence; we need it to hold for all subsets of indices (or equivalently for E i E_i E i replaced by E i c E_i^c E i c ).
However, for random variables, the definition above is sufficient. Why?
Because the sets A i A_i A i are arbitrary Borel sets. We can choose A i = R A_i = \mathbb{R} A i = R for indices we want to “ignore”, effectively removing those variables from the intersection. Since the definition requires the equality for all A i A_i A i , it covers all sub-combinations automatically.
Independence is preserved under measurable transformations.
Theorem
If X ⊥ Y X \perp Y X ⊥ Y are independent random variables, and f , g f, g f , g are measurable functions, then f ( X ) f(X) f ( X ) and g ( Y ) g(Y) g ( Y ) are independent.
We need to show that for any A , B ∈ B A, B \in \mathcal{B} A , B ∈ B :
P ( f ( X ) ∈ A , g ( Y ) ∈ B ) = P ( f ( X ) ∈ A ) P ( g ( Y ) ∈ B ) \mathbb{P}(f(X) \in A, g(Y) \in B) = \mathbb{P}(f(X) \in A)\mathbb{P}(g(Y) \in B) P ( f ( X ) ∈ A , g ( Y ) ∈ B ) = P ( f ( X ) ∈ A ) P ( g ( Y ) ∈ B )
Rewrite the events using preimages:
{ f ( X ) ∈ A } = { X ∈ f − 1 ( A ) } , { g ( Y ) ∈ B } = { Y ∈ g − 1 ( B ) } \{f(X) \in A\} = \{X \in f^{-1}(A)\}, \quad \{g(Y) \in B\} = \{Y \in g^{-1}(B)\} { f ( X ) ∈ A } = { X ∈ f − 1 ( A )} , { g ( Y ) ∈ B } = { Y ∈ g − 1 ( B )}
Since f , g f, g f , g are measurable, A ′ = f − 1 ( A ) ∈ B A' = f^{-1}(A) \in \mathcal{B} A ′ = f − 1 ( A ) ∈ B and B ′ = g − 1 ( B ) ∈ B B' = g^{-1}(B) \in \mathcal{B} B ′ = g − 1 ( B ) ∈ B .
Since X ⊥ Y X \perp Y X ⊥ Y , we have:
P ( X ∈ A ′ , Y ∈ B ′ ) = P ( X ∈ A ′ ) P ( Y ∈ B ′ ) \mathbb{P}(X \in A', Y \in B') = \mathbb{P}(X \in A')\mathbb{P}(Y \in B') P ( X ∈ A ′ , Y ∈ B ′ ) = P ( X ∈ A ′ ) P ( Y ∈ B ′ )
Substituting back gives the result.
It is often tedious to check independence for all sets in a σ \sigma σ -field. It suffices to check it on a generating π \pi π -system (a family of sets closed under finite intersection).
Definition: Independent Classes
Two classes of sets A 1 \mathcal{A}_1 A 1 and A 2 \mathcal{A}_2 A 2 are independent if:
P ( A 1 ∩ A 2 ) = P ( A 1 ) P ( A 2 ) ∀ A 1 ∈ A 1 , A 2 ∈ A 2 \mathbb{P}(A_1 \cap A_2) = \mathbb{P}(A_1)\mathbb{P}(A_2) \quad \forall A_1 \in \mathcal{A}_1, A_2 \in \mathcal{A}_2 P ( A 1 ∩ A 2 ) = P ( A 1 ) P ( A 2 ) ∀ A 1 ∈ A 1 , A 2 ∈ A 2