Skip to content

Moments, Variance

We define higher-order statistics based on the expectation.

  1. Bernoulli Distribution: XBern(p)X \sim \text{Bern}(p).

    P(X=1)=p,P(X=0)=1p\mathbb{P}(X=1) = p, \quad \mathbb{P}(X=0) = 1-p
    • Mean: E[X]=1p+0(1p)=p\mathbb{E}[X] = 1 \cdot p + 0 \cdot (1-p) = p.
    • Second Moment: Since XX takes values {0,1}\{0, 1\}, we have X2=XX^2 = X. Thus E[X2]=E[X]=p\mathbb{E}[X^2] = \mathbb{E}[X] = p
    • Variance: Var(X)=E[X2](E[X])2=pp2=p(1p)\text{Var}(X) = \mathbb{E}[X^2] - (\mathbb{E}[X])^2 = p - p^2 = p(1-p)
  2. Poisson Distribution: XPoi(λ), λ>0X \sim \text{Poi}(\lambda), \ \lambda > 0.

    P(X=k)=eλλkk!,k=0,1,\mathbb{P}(X=k) = e^{-\lambda} \frac{\lambda^k}{k!}, \quad k = 0, 1, \dots

    To calculate moments, it is easier to use factorial moments. Consider E[X(X1)(Xk+1)]\mathbb{E}[X(X-1)\dots(X-k+1)].

    E[X(X1)(Xk+1)]=j=0j(j1)(jk+1)eλλjj!=j=kj!(jk)!eλλjj!=j=keλλj(jk)!\begin{aligned} \mathbb{E}[X(X-1)\dots(X-k+1)] &= \sum_{j=0}^{\infty} j(j-1)\dots(j-k+1) e^{-\lambda} \frac{\lambda^j}{j!} \\ &= \sum_{j=k}^{\infty} \frac{j!}{(j-k)!} e^{-\lambda} \frac{\lambda^j}{j!} = \sum_{j=k}^{\infty} e^{-\lambda} \frac{\lambda^j}{(j-k)!} \end{aligned}

    Let i=jki = j-k. Then:

    =λki=0eλλii!=λk(i=0eλλii!)1 (sum of PMF)=λk= \lambda^k \sum_{i=0}^{\infty} e^{-\lambda} \frac{\lambda^i}{i!} = \lambda^k \underbrace{\left( \sum_{i=0}^{\infty} e^{-\lambda} \frac{\lambda^i}{i!} \right)}_{1 \text{ (sum of PMF)}} = \lambda^k
    • Mean: Put k=1k=1. E[X]=λ1=λ\mathbb{E}[X] = \lambda^1 = \lambda.
    • Second Moment: Using k=2k=2, E[X(X1)]=λ2\mathbb{E}[X(X-1)] = \lambda^2. E[X2]=E[X(X1)]+E[X]=λ2+λ\mathbb{E}[X^2] = \mathbb{E}[X(X-1)] + \mathbb{E}[X] = \lambda^2 + \lambda
    • Variance: Var(X)=(λ2+λ)λ2=λ\text{Var}(X) = (\lambda^2 + \lambda) - \lambda^2 = \lambda
  3. Exponential Distribution: XExp(λ)X \sim \text{Exp}(\lambda). Density f(x)=λeλx, x0f(x) = \lambda e^{-\lambda x}, \ x \ge 0.

    E[Xk]=0xkλeλxdx\mathbb{E}[X^k] = \int_0^{\infty} x^k \lambda e^{-\lambda x} \, dx

    Substitute y=λx    dx=dy/λy = \lambda x \implies dx = dy/\lambda.

    =1λk0ykeydy=Γ(k+1)λk=k!λk= \frac{1}{\lambda^k} \int_0^{\infty} y^k e^{-y} \, dy = \frac{\Gamma(k+1)}{\lambda^k} = \frac{k!}{\lambda^k}
    • Mean: E[X]=1!/λ1=1/λ\mathbb{E}[X] = 1! / \lambda^1 = 1/\lambda.
    • Variance: Var(X)=2!λ2(1λ)2=2λ21λ2=1λ2\text{Var}(X) = \frac{2!}{\lambda^2} - \left( \frac{1}{\lambda} \right)^2 = \frac{2}{\lambda^2} - \frac{1}{\lambda^2} = \frac{1}{\lambda^2}

If the MGF exists in a neighborhood of t=0t=0, we can expand etXe^{tX} as a Taylor series:

etX=1+tX+(tX)22!+(tX)33!+e^{tX} = 1 + tX + \frac{(tX)^2}{2!} + \frac{(tX)^3}{3!} + \dots

Taking expectations (assuming we can swap sum and expectation):

MX(t)=1+tE[X]+t22!E[X2]+t33!E[X3]+M_X(t) = 1 + t\mathbb{E}[X] + \frac{t^2}{2!}\mathbb{E}[X^2] + \frac{t^3}{3!}\mathbb{E}[X^3] + \dots

Thus, the kk-th derivative at t=0t=0 generates the kk-th moment:

MX(k)(0)=E[Xk]M_X^{(k)}(0) = \mathbb{E}[X^k]

The MGF does not always exist! The term etXe^{tX} grows very fast. For E[etX]\mathbb{E}[e^{tX}] to be finite, the probability density of XX must decay fast enough (faster than etxe^{-tx}) to counteract this growth.

  • Light Tails: If tails decay exponentially (like Poisson, Exponential, Normal), the MGF usually exists in a neighborhood of 0.
  • Heavy Tails: If tails decay slower (polyomially), the integral might diverge for t>0t > 0.