Download - CHAPTER 1: DISTRIBUTION THEORY

Transcript
Page 1: CHAPTER 1: DISTRIBUTION THEORY

CHAPTER 1: DISTRIBUTION THEORY

Page 2: CHAPTER 1: DISTRIBUTION THEORY

Basic Concepts

Page 3: CHAPTER 1: DISTRIBUTION THEORY

• Random Variable (R.V.)

� discrete random variable: probability mass function� continuous random variable: probability density function

• Random Vector� multivariate version of random variables� joint probability function, joint density function

Page 4: CHAPTER 1: DISTRIBUTION THEORY

• Formal definitions based on measure theory

� R.V. is a measurable function on a probability measure space,which consists of sets in a σ-field equipped with aprobability measure;� probability mass:

the Radon-Nykodym derivative of a random variable-inducedmeasure w.r.t. to a counting measure;� density function:

the Radon-Nykodym derivative of a random variable-inducedmeasure w.r.t. the Lebesgue measure.

Page 5: CHAPTER 1: DISTRIBUTION THEORY

• Descriptive quantities of a univariate distribution

� cumulative distribution function

FX(x) ≡ P(X ≤ x)

� moments (mean): E[Xk], k = 1, 2, . . . (mean if k = 1) ;centralized moments: E[(X − µ)k] (variance if k = 2)� quantile function

F−1X (p) ≡ inf

x{FX(x) > p}

� the skewness: E[(X − µ)3]/Var(X)3/2

� the kurtosis: E[(X − µ)4]/Var(X)2

Page 6: CHAPTER 1: DISTRIBUTION THEORY

• Characteristic function (c.f.)� definition

φX(t) = E[exp{itX}]

� c.f. uniquely determines the distribution

FX(b)− FX(a) = limT→∞

12π

∫ T

−T

e−ita − e−itb

itφX(t)dt

fX(x) =1

∫ ∞−∞

e−itxφX(t)dt for continuous r.v.

� moment generating function (if it exists)

mX(t) = E[exp{tX}]

Page 7: CHAPTER 1: DISTRIBUTION THEORY

• Descriptive quantities of two R.V.s (X,Y)

� covariance and correlation: Cov(X,Y), corr(X,Y)

� conditional density or probability (f ’s become probabilitymass functions if R.V.s are discrete):

fX|Y(x|y) = fX,Y(x, y)/fY(y)

E[X|Y] =

∫xfX|Y(x|y)dx

� independence

P(X ≤ x,Y ≤ y) = P(X ≤ x)P(Y ≤ y)

fX,Y(x, y) = fX(x)fY(y)

Page 8: CHAPTER 1: DISTRIBUTION THEORY

• Double expectation formulae

E[X] = E[E[X|Y]]

Var(X) = E[Var(X|Y)] + Var(E[X|Y])

• This is useful for computing the moments of X when X givenY has a simple distribution!In-class exercise Compute E[X] and Var(X) if Y ∼ Bernoulli(p) and X given Y = y follows N(µy, σ

2y).

Page 9: CHAPTER 1: DISTRIBUTION THEORY

Discrete Random Variables

Page 10: CHAPTER 1: DISTRIBUTION THEORY

• Binomial distribution� Binomial r.v. is the total number of successes in n

independent Bernoulli trials.� Binomial(n, p):

P(Sn = k) =

(nk

)pk(1− p)n−k, k = 0, ...,n

E[Sn] = np, Var(Sn) = np(1− p)

φX(t) = (1− p + peit)n

� Summation of two independent Binomial r.v.’s

Binomial(n1, p) + Binomial(n2, p) ∼ Binomial(n1 + n2, p)

In-class exercise Prove the last statement.

Page 11: CHAPTER 1: DISTRIBUTION THEORY

• Negative Binomial distribution� It describes the distribution of the number of independent

Bernoulli trials until m successes.� NB(m, p):

P(Wm = k) =

(k− 1m− 1

)pm(1− p)k−m, k = m,m + 1, ...

E[Wm] = m/p, Var(Wm) = m/p2 −m/p

� Summation of two independent r.v.’s:

NB(m1, p) + NB(m2, p) ∼ NB(m1 + m2, p)

In-class exercise Prove the last statement.

Page 12: CHAPTER 1: DISTRIBUTION THEORY

• Hypergeometric distribution� It describes the distribution of the number of black balls if

we randomly draw n balls from an urn with M blacks and(N −M) whites (urn model).� HG(N,M,n):

P(Sn = k) =

(Mk

)(N−Mn−k

)(Nn

) , k = 0, 1, ..,n

E[Sn] = Mn/(M + N), Var(Sn) =nMN(M + N − n)

(M + N)2(M + N − 1)

Page 13: CHAPTER 1: DISTRIBUTION THEORY

• Poisson distribution� It is used to describe the distribution of the number of

incidences in an infinite population.� Poisson(λ):

P(X = k) = λke−λ/k!, k = 0, 1, 2, ...

E[X] = Var(X) = λ, φX(t) = exp{−λ(1− eit)}

� Summation of two independent r.v.’s:

Poisson(λ1) + Poisson(λ2) ∼ Poisson(λ1 + λ2)

Page 14: CHAPTER 1: DISTRIBUTION THEORY

• Poisson distribution related to Binomial distribution

� Assume

X1 ∼ Poisson(λ1), X2 ∼ Poisson(λ2), X1 ⊥⊥ X2.

ThenX1|X1 + X2 = n ∼ Binomial(n,

λ1

λ1 + λ2).

� If Xn1, ...,Xnn are i.i.d Bernoulli(pn) and npn → λ, then

P(Sn = k) =n!

k!(n− k)!pk

n(1− pn)n−k → λk

k!e−λ.

In-class exercise Prove these statements.

Page 15: CHAPTER 1: DISTRIBUTION THEORY

•Multinomial Distribution� Assume Y1, ...,Yn are independent k-category random

varaibles such that P(Yi = category l) = pl,∑k

l=1 pl = 1.� Nl, 1 ≤ l ≤ k counts the number of times that {Y1, ...,Yn}

taking category value l.� Then P(N1 = n1, ...,Nk = nk) =

( nn1,...,nk

)pn1

1 · · · pnkk

n1 + ...+ nk = n

� the covariance matrix for (N1, ...,Nk)

n

p1(1− p1) . . . −p1pk...

. . ....

−p1pk . . . pk(1− pk)

= n

diag(p1, ..., pk)−

p1...

pk

(p1 . . . pk)

Page 16: CHAPTER 1: DISTRIBUTION THEORY

Continuous Random Variables

Page 17: CHAPTER 1: DISTRIBUTION THEORY

• Uniform distribution� Uniform(a, b):

fX(x) = I[a,b](x)/(b− a),

where IA(x) is an indicator function taking value 1 if x ∈ Aand 0, otherwise.�

E[X] = (a + b)/2, Var(X) = (b− a)2/12

� The uniform distribution in [0, 1] is the most fundamentaldistribution each pseudo random number generator triesto generate numbers from.

Page 18: CHAPTER 1: DISTRIBUTION THEORY

• Normal distribution� It describes the distribution behavior of averaged statistics

in practice.� N(µ, σ2)

fX(x) =1√

2πσ2exp{−(x− µ)2

2σ2 }

E[X] = µ, Var(X) = σ2

φX(t) = exp{itµ− σ2t2/2}

Page 19: CHAPTER 1: DISTRIBUTION THEORY

• Gamma distribution� It is useful to describe the distribution of a positive

random variable with long right tail.� Gamma(θ, β)

fX(x) =1

βθΓ(θ)xθ−1 exp{− x

β}, x > 0

E[X] = θβ, Var(X) = θβ2

� θ is called the shape parameter and β is the scaleparameter.� θ = 1 reduces to the so-called exponential distribution,

Exp(β), and θ = n/2, β = 2 is equivalent to a chi-squareddistribution with n degrees of freedom.

Caution In some references, notation Gamma(θ, β) is used to refer to the distribution Gamma(θ, 1/β) as defined

here!

Page 20: CHAPTER 1: DISTRIBUTION THEORY

• Cauchy distribution� Cauchy(a, b)

fX(x) =1

bπ {1 + (x− a)2/b2}

� E[|X|] =∞, φX(t) = exp{iat− |bt|}� This distribution is often used as a counter example whose

mean does not exist.

Page 21: CHAPTER 1: DISTRIBUTION THEORY

Algebra of Random Variables

Page 22: CHAPTER 1: DISTRIBUTION THEORY

• Assumption� We assume that X and Y are independent with distribution

functions FX and FY, respectively.� We are interested in the distribution or density of

X + Y, XY, X/Y.

� For XY and X/Y, we assume Y to be positive forconvenience, although this is not necessary.

Page 23: CHAPTER 1: DISTRIBUTION THEORY

• Summation of X and Y� Derivation:

FX+Y(z) = E[I(X + Y ≤ z)] = EY[EX[I(X ≤ z− Y)|Y]]

= EY[FX(z− Y)] =

∫FX(z− y)dFY(y)

� Convolution formula (density refers to the situation whenX and Y are continuous):

FX+Y(z) =

∫FY(z− x)dFX(x) ≡ FX ∗ FY(z)

fX ∗ fY(z) ≡∫

fX(z− y)fY(y)dy =

∫fY(z− x)fX(x)dx

Page 24: CHAPTER 1: DISTRIBUTION THEORY

• Product and quotient of X and Y (Y > 0)

� FXY(z) = E[E[I(XY ≤ z)|Y]] =∫

FX(z/y)dFY(y)

fXY(z) =

∫fX(z/y)/yfY(y)dy

� FX/Y(z) = E[E[I(X/Y ≤ z)|Y]] =∫

FX(yz)dFY(y)

fX/Y(z) =

∫fX(yz)yfY(y)dy

In-class exercise How to obtain the distributions if Y may not be positive?

Page 25: CHAPTER 1: DISTRIBUTION THEORY

• Application of algebra formulae� N(µ1, σ

21) + N(µ2, σ

22) ∼ N(µ1 + µ2, σ

21 + σ2

2)

� Gamma(r1, θ) + Gamma(r2, θ) ∼ Gamma(r1 + r2, θ)

� Poisson(λ1) + Poisson(λ2) ∼ Poisson(λ1 + λ2)

� Negative Binomial(m1, p) + Negative Binomial(m2, p)

∼ Negative Binomial(m1 + m2, p)

In-class exercise Verify the summation for Poisson r.v.’s.

Page 26: CHAPTER 1: DISTRIBUTION THEORY

• Deriving the summation of R.V.s using characteristicfunctions

� Key property If X and Y are independent, then

φX+Y(t) = φX(t)φY(t)

� For example, X and Y are normal withφX(t) = exp{iµ1t− σ2

1t2/2}, φY(t) = exp{iµ2t− σ22t2/2}

⇒φX+Y(t) = exp{i(µ1 + µ2)t− (σ21 + σ2

2)t2/2}

Page 27: CHAPTER 1: DISTRIBUTION THEORY

•More distributions generated from simple random variables

� Assume X ∼ N(0, 1), Y ∼ χ2m and Z ∼ χ2

n are independent.Then

X√Y/m

∼ Student’s t(m),

Y/mZ/n

∼ Snedecor’s Fm,n,

YY + Z

∼ Beta(m/2,n/2).

Page 28: CHAPTER 1: DISTRIBUTION THEORY

• Densities of t−, F− and Beta− distributions

ft(m)(x) =Γ((m + 1)/2)√πmΓ(m/2)

1(1 + x2/m)(m+1)/2 I(−∞,∞)(x)

fFm,n(x) =Γ(m + n)/2

Γ(m/2)Γ(n/2)

(m/n)m/2xm/2−1

(1 + mx/n)(m+n)/2 I(0,∞)(x)

�fBeta(a,b)(x) =

Γ(a + b)

Γ(a)Γ(b)xa−1(1− x)b−1I(x ∈ (0, 1))

� It is always recommended that an indicator function isincluded in the density expression to indicate the supportof r.v.

Page 29: CHAPTER 1: DISTRIBUTION THEORY

• Exponential vs Beta- distributions

� Assume Y1, ...,Yn+1 are i.i.d Exp(θ).

� Then Zi = Y1+...+YiY1+...+Yn+1

∼ Beta(i,n− i + 1).

� In fact, (Z1, . . . ,Zn) has the same joint distribution as thatof the order statistics (ξn:1, ..., ξn:n) of n independentUniform(0,1) random variables.

Page 30: CHAPTER 1: DISTRIBUTION THEORY

Transformation of Random Vector

Page 31: CHAPTER 1: DISTRIBUTION THEORY

Theorem 1.3 Suppose that X is k-dimension random vectorwith density function fX(x1, ..., xk). Let g be a one-to-one andcontinuously differentiable map from Rk to Rk. Then Y = g(X)is a random vector with density function

fX(g−1(y1, ..., yk))|Jg−1(y1, ..., yk)|,

where g−1 is the inverse of g and Jg−1 is the Jacobian of g−1.In-class exercise Prove it by transformation in multivariate integration.

Page 32: CHAPTER 1: DISTRIBUTION THEORY

• An application

� let R2 ∼ Exp{2},R > 0 and Θ ∼ Uniform(0, 2π) beindependent;� X = R cos Θ and Y = R sin Θ are two independent

standard normal random variables;� it can be applied to simulate normally distributed data.

In-class exercise Prove this statement.

In-class remark It is fundamental to have a mechanism to generate data from U(0, 1) (pseudo random

number generator). To generate data from any distribution FX , generate data from F−1X (U) where

U ∼ U(0, 1). Alternative algorithms include acceptance-rejection method, MCMC and etc.

Page 33: CHAPTER 1: DISTRIBUTION THEORY

Multivariate Normal Distribution

Page 34: CHAPTER 1: DISTRIBUTION THEORY

• DefinitionY = (Y1, ...,Yn)′ is said to have a multivariate normaldistribution with mean vector µ = (µ1, ..., µn)′ andnon-degenerate covariance matrix Σn×n if its joint density is

fY(y1, ..., yn) =1

(2π)n/2|Σ|1/2 exp{−12

(y− µ)′Σ−1(y− µ)}

Page 35: CHAPTER 1: DISTRIBUTION THEORY

• Characteristic function

φY(t)

= E[eit′Y]

= (2π)− n

2 |Σ|−12∫

exp{it′y−1

2(y− µ)

′Σ−1

(y− µ)}dy

= (√

2π)− n

2 |Σ|−12∫

exp{−y′Σ−1y

2+ (it + Σ

−1µ)′y−

µ′Σ−1µ

2}dy

=exp{−µ′Σ−1µ/2}(√

2π)n/2|Σ|1/2

∫exp{−

1

2(y− Σit− µ)

′Σ−1

(y− Σit− µ)

+(Σit + µ)′Σ−1(Σit + µ)

2

}dy

= exp

{−µ′Σ−1µ

2+

1

2(Σit + µ)

′Σ−1

(Σit + µ)

}

= exp{it′µ−1

2t′Σt}.

Page 36: CHAPTER 1: DISTRIBUTION THEORY

• Some remarks� multivariate normal distribution is uniquely determined

by µ and Σ.� for standard multivariate normal distribution,

φX(t) = exp{−t′t/2}.

� the moment generating function for Y is

exp{t′µ+ t′Σt/2}.

In-class exercise Derive the moments of the standard normal distribution.

Page 37: CHAPTER 1: DISTRIBUTION THEORY

• Linear transformation of normal r.v.Theorem 1.4 If Y = An×kXk×1 where X ∼ N(0, I) (standardmultivariate normal distribution), then Y’s characteristicfunction is given by

φY(t) = exp{−t′Σt/2

}, t = (t1, ..., tn)′ ∈ Rk,

where Σ = AA′ so rank(Σ) = rank(A), i.e., Y ∼ N(0,Σ).Conversely, if φY(t) = exp{−t′Σt/2}with Σn×n ≥ 0 of rank k,i.e., Y ∼ N(0,Σ), then there exits an A matrix with rank(A) = ksuch that

Y = An×kXk×1 with X ∼ N(0, Ik×k).

Page 38: CHAPTER 1: DISTRIBUTION THEORY

Proof of Theorem 1.4For the first part, since Y = AX and X ∼ N(0, Ik×k), we obtain

φY(t) = E[exp{it′(AX)}]

= E[exp{i(A′t)′X}]

= exp{−(A′t)′(A′t)/2}

= exp{−t′AA′t/2}.

Thus, Y ∼ N(0,Σ) with Σ = AA′ . Clearly, rank(Σ) = rank(AA′) = rank(A).

Page 39: CHAPTER 1: DISTRIBUTION THEORY

For the second part, suppose φY(t) = exp{−t′Σt/2}. There exists an orthogonal matrix, O, such that

Σ = O′DO, D = diag((d1, ..., dk, 0, ..., 0)′), d1, ..., dk > 0.

Define Z = OY thenφZ(t) = E[exp{it′(OY)}] = E[exp{i(O′t)′Y}]

= exp{−(O′t)′Σ(O′t)

2} = exp{−d1t2

1/2− ...− dkt2k/2}.

Thus, Z1, ..., Zk are independent N(0, d1), ...,N(0, dk) but Zk+1, ..., Zn are zeros.Let Xi = Zi/

√di, i = 1, ..., k and write O′ = (Bn×k, Cn×(n−k)). Clearly,

Y = O′Z = Bdiag{√

d1, ...,√

dk}X

so the result holds by letting A = Bdiag{√

d1, ...,√

dk}.

In-class remark Note that this proof also shows the way to generate data from N(0,Σ).

Page 40: CHAPTER 1: DISTRIBUTION THEORY

• Conditional normal distributionsTheorem 1.5 Suppose that Y = (Y1, ...,Yk,Yk+1, ...,Yn)′ has amultivariate normal distribution with mean µ = (µ(1)

′, µ(2)

′)′

and a non-degenerate covariance matrix

Σ =

(Σ11 Σ12Σ21 Σ22

).

Let Y(1) = (Y1, ...,Yk)′ and Y(2) = (Yk+1, ...,Yn)′. Then

(i) Y(1) ∼ Nk(µ(1),Σ11).

(ii) Y(1) and Y(2) are independent iff Σ12 = Σ21 = 0.(iii) For any matrix Am×n, AY has a multivariate normaldistribution with mean Aµ and covariance AΣA′.(iv) The conditional distribution of Y(1) given Y(2) is

Y(1)|Y(2) ∼ Nk(µ(1) + Σ12Σ−1

22 (Y(2) − µ(2)),Σ11 − Σ12Σ−122 Σ21).

Page 41: CHAPTER 1: DISTRIBUTION THEORY

Proof of Theorem 1.5We prove (iii) first. Since φY(t) = exp{it′µ− t′Σt/2},

φAY(t) = E[eit′AY]

= E[ei(A′t)′Y]

= exp{i(A′t)′µ− (A′t)′Σ(A′t)/2}

= exp{it′(Aµ)− t′(AΣA′)t/2}.

Thus, AY ∼ N(Aµ,AΣA′).

Page 42: CHAPTER 1: DISTRIBUTION THEORY

(i) follows from (iii) by letting A =(

Ik×k 0k×(n−k))

.To prove (ii). Note

φY(t) = exp[

it(1)′µ(1)

+ it(2)′µ(2)

−1

2

{t(1)′

Σ11t(1)+ 2t(1)′

Σ12t(2)+ t(2)′

Σ22t(2)}].

Thus, Y(1) and Y(2) are independent iff the characteristics function is a product of separate functions for t(1) and

t(2) iff Σ12 = 0.

Page 43: CHAPTER 1: DISTRIBUTION THEORY

To prove (iv), we consider

Z(1)= Y(1) − µ(1) − Σ12Σ

−122 (Y(2) − µ(2)

).

From (iii), Z(1) ∼ N(0,ΣZ) with

ΣZ = Cov(Y(1), Y(1)

)− 2Σ12Σ−122 Cov(Y(2)

, Y(1)) + Σ12Σ

−122 Cov(Y(2)

, Y(2))Σ−122 Σ21

= Σ11 − Σ12Σ−122 Σ21.

On the other hand,

Cov(Z(1), Y(2)

) = Cov(Y(1), Y(2)

)− Σ12Σ−122 Cov(Y(2)

, Y(2)) = 0

so Z(1) is independent of Y(2) by (ii).Therefore, the conditional distribution Z(1) given Y(2) is the same as the unconditional distribution of Z(1) , whichis N(0,ΣZ). I.e.,

Z(1)|Y(2) ∼ Z(1) ∼ N(0,Σ11 − Σ12Σ−122 Σ21).

The result holds.

In-class remark The constructed Z has a geometric interpretation as (Y(1) − µ(1)) subtracting its project on the

normalized variable Σ−1/222 (Y(2) − µ(2)).

Page 44: CHAPTER 1: DISTRIBUTION THEORY

• An example of measurement error model

� X and U are independent with X ∼ N(0, σ2x) and

U ∼ N(0, σ2u)

� Additive measurement error model assumes

Y = X + U.

� Note that the covariance of (X,Y) is(σ2

x σ2x

σ2x σ2

x + σ2u

)so from (iv),

X|Y ∼ N(σ2

xσ2

x + σ2u

Y, σ2x −

σ4x

σ2x + σ2

u) ≡ N(λY, σ2

x(1− λ)).

� Here, λ = σ2x/σ

2y is called (reliability coefficient).

Page 45: CHAPTER 1: DISTRIBUTION THEORY

• Quadratic form of normal random variables

� Summation of independent normal r.v.s:

N(0, 1)2 + ...+ N(0, 1)2 ∼ χ2n ≡ Gamma(n/2, 2).

� If Y ∼ N(0,Σn×n) with Σ > 0, then

Y′Σ−1Y ∼ χ2n.

Proof: Y can be expressed as AX where X ∼ N(0, I) and A = Σ−1/2 . Thus,

Y′Σ−1Y = X′A′(AA′)−1AX = X′X ∼ χ2n.

In-class exercise What is the distribution of Y′BY for an n× n symmetric matrix B?

Page 46: CHAPTER 1: DISTRIBUTION THEORY

• Noncentral Chi-square distribution

� AssumeX ∼ N(µ, In×n).

� DefineY = X′X ∼ χ2

n(δ)

and δ = µ′µ.� Y has a noncentral chi-square distribution and its density

function is a Poisson-probability weighted summation ofgamma densities:

fY(y) =

∞∑k=0

exp{−δ/2}(δ/2)k

k!Gamma(y; (2k + n)/2, 2).

In-class remark Equivalently, we generate S from Poisson(δ/2) and given S = k, we generate

Y ∼ Gamma((2k + n)/2, 2).

Page 47: CHAPTER 1: DISTRIBUTION THEORY

•More about normal distributions� noncentral t-distribution: N(δ, 1)/

√χ2

n/nnoncentral F-distribution: χ2

n(δ)/n/(χ2m/m)

� One trick to evaluate E[X′AX] when X ∼ N(µ,Σ):

E[X′AX] = E[tr(X′AX)] = E[tr(AXX′)]= tr(AE[XX′]) = tr(A(µµ′ + Σ))

= µ′Aµ+ tr(AΣ)

Page 48: CHAPTER 1: DISTRIBUTION THEORY

Distribution Family

Page 49: CHAPTER 1: DISTRIBUTION THEORY

• Location-scale family� aX + b: fX((x− b)/a)/a, a > 0, b ∈ R with mean aE[X] + b

and variance a2Var(X)

� Examples include normal distribution, uniformdistribution, gamma distributions etc.

Page 50: CHAPTER 1: DISTRIBUTION THEORY

• Exponential family

� Exponential family includes many commonly useddistributions such as binomial, poisson distributions fordiscrete variables and normal distribution, gammadistribution, Beta distribution for continuous variables.� The distribution within this family has a general

expression of densities, which share some family-specificproperties.

Page 51: CHAPTER 1: DISTRIBUTION THEORY

• General form of densities in an exponential family

� {pθ(x)}, which consists of densities indexed by θ (one ormore parameters), is said to form an s-parameterexponential family if

pθ(x) = exp

{s∑

k=1

ηk(θ)Tk(x)− B(θ)

}h(x),

where h(x) ≥ 0 and

exp{B(θ)} =

∫exp{

s∑k=1

ηk(θ)Tk(x)}h(x)dµ(x) <∞.

� When (η1(θ), ..., ηs(θ))′ is exactly the same as θ, the above

form is called the canonical form of the exponential family.

Page 52: CHAPTER 1: DISTRIBUTION THEORY

• Examples� X1, ...,Xn ∼ N(µ, σ2):

exp

σ2

n∑i=1

xi −1

2σ2

n∑i=1

x2i −

n2σ2µ

2

}1

(√

2πσ)n.

Its canonical parameterization is θ1 = µ/σ2 and θ2 = 1/σ2.� X ∼ Binomial(n, p):

exp{x logp

1− p+ n log(1− p)}

(nx

).

Its canonical parameterization is θ = log p/(1− p).� X ∼ Poisson(λ):

P(X = x) = exp{x logλ− λ}/x!

Its canonical parameterization is θ = logλ.

Page 53: CHAPTER 1: DISTRIBUTION THEORY

•Moment generating function in the exponential familyTheorem 1.6 Suppose the densities of an exponential familycan be written as the canonical form

exp{s∑

k=1

ηkTk(x)− A(η)}h(x),

where η = (η1, ..., ηs)′. Then the moment generating function

(MGF) for (T1, ...,Ts), defined as

MT(t1, ..., ts) = E [exp{t1T1 + ...+ tsTs}]

for t = (t1, ..., ts)′, is given by

MT(t) = exp{A(η + t)− A(η)}.

Page 54: CHAPTER 1: DISTRIBUTION THEORY

Proof of Theorem 1.6Since

exp{A(η)} =

∫exp{

s∑k=1

ηiTi(x)}h(x)dµ(x),

we haveMT(t) = E [exp{t1T1 + ... + tsTs}]

=

∫exp{

s∑k=1

(ηi + ti)Ti(x)− A(η)}h(x)dµ(x)

= exp{A(η + t)− A(η)}.

Page 55: CHAPTER 1: DISTRIBUTION THEORY

•MGF and cumulant generating function (CGF)� MGF can be useful to calculate the moments of (T1, ...,Ts).� Define the cumulant generating function (CGF)

KT(t1, ..., ts) = log MT(t1, ..., ts) = A(η + t)− A(η).

The coefficients in its Taylor expansion are called thecumulants for (T1, ...,Ts).� The first two cumulants are the mean and variance.

Page 56: CHAPTER 1: DISTRIBUTION THEORY

• Examples of using MGF

� Normal distribution: η = µ/σ2 and

A(η) =1

2σ2µ2 = η2σ2/2.

Thus,

MT(t) = exp{σ2

2((η + t)2 − η2)} = exp{µt + t2σ2/2}.

� For X ∼ N(0, σ2), since T = X, from the Taylor expansion,we have

E[X2r+1] = 0,E[X2r] = 1 · 2 · · · (2r− 1)σ2r, r = 1, 2, ...

Page 57: CHAPTER 1: DISTRIBUTION THEORY

� Gamma distribution has a canonical form

exp{−x/b + (a− 1) log x− log(Γ(a)ba)}I(x > 0).

� Treat b as the only parameter (a is a constant) then thecanonical parameterization is η = −1/b with T = X.� Note A(η) = log(Γ(a)ba) = a log(−1/η) + log Γ(a). we

obtain

MX(t) = exp{a logη

η + t} = (1− bt)−a.

This gives

E[X] = ab,E[X2] = ab2 + (ab)2, ...

Page 58: CHAPTER 1: DISTRIBUTION THEORY

READING MATERIALS: Lehmann and Casella, Sections 1.4and 1.5