Skip to main content

Module 5: Multivariate Normal Distribution

A variable X follows a discrete probability distribution if the possible values of X are either:

  • A finite set
  • A countable infinite sequence

px(xi) = P(X=xi) is called the probability mass function (PMF)

  • px(xi)  >= 0 as it is a probability
  • The sum of PMF for all values of X = 1

Recall that in a Discrete Probability Distribution :

image-1660918683481.png

In a Continuous Probability Distribution:

image-1660918718518.png

Moment Generating Function

Moments, such as E(X), V(X), can also be calculated using the Moment Generating Function (MGF):

image-1660918098650.png

The rth moment of X, E(Xr) can be obtained by differentiating Mx(t) r times with respect to t and setting t=0

  • Mx(0) = 1
  • MIx(0) = E(X)
  • MIIx(0) = E(X2) -> V(X) = MIIx(0) - (MIx(0))2
  • In general, Mx(r)(0) = E(Xr)

Uniqueness: if X and Y are two random variables and Mx(t) = My(t) when |t| < h for some positive number h, then X and Y have the same distribution

Note: MGF does not exist for all distributions (E(etx) may be infinity)

Binomial Distribution

X ~ Binomial(n, p)      𝑝 ∈ [0, 1]

X = the number of successes in n trials when the probability of success in each trail is p.

We can think of X as the sum of n independent Bernoulli(p) random variables, with the same p for every Xi

image-1660920861915.png

  • PMF = P(X = x) = image-1660918544947.png
  • Expected value = E(X) = np
  • Variance = V(X) = np(1-p)
  • MGF = Mx(t) = (pet + (1-p))n
  • Two discrete random variables are independent if: P(X = x & Y = y) = P(X = x)*P(Y=y)

Poisson Distribution

X ~ Poisson(λ)     λ > 0

X = The number of occurrences of an event of interest.

  • PMF = image-1660920679945.png
  • Expected Values = E(X) = λ
  • Variance = V(X) = λ
  • MGF = Mx(t) = eλ(e^t - 1)

Poisson as an approximation  of the Binomial Distribution

  • If X ~ Binomial(n, p) and n -> infinity, p-> 0 such that np is a constant  =>  X ~ Poisson(np)
  • Often used analyzing rare diseases

Geometric Distribution

X ~ Geometric(p)      𝑝 ∈ (0, 1]

If Y1, Y2, Y3 ... are a sequence of independent Bernoulli(p) random variables then the number of failures before the first success, X, follows a Geometric distribution.

  • PMF = P(X = x) = p(1-p)x
  • Expected value = E(X) = (1-p)/p
  • Variance = V(X) = (1-p)/p2
  • MGF = Mx(t) = p / (1 - (1 - p)et)

Hyper-Geometric Distribution

X ~ Hypergeometric(N, K, n)

Suppose a finite population of size N contains two mutually exclusive events: K success events and N-K failure events. If n events are randomly chosen without replacement X is the number of success events chosen.

  • PMF = P(X = x) = image-1660921389776.png
  • Expected value = E(X) = nk / N
  • Variance = V(X) = ((nK) / N) * ((N-K) / N) * ((N - n) / (N - 1))