Paul Cornwell March 31, 2011 1. Let X 1,…,X n be independent, identically distributed random...

Post on 23-Dec-2015

216 views 0 download

Transcript of Paul Cornwell March 31, 2011 1. Let X 1,…,X n be independent, identically distributed random...

1

The Central Limit Theorem

Paul CornwellMarch 31, 2011

2

Let X1,…,Xn be independent, identically distributed random variables with positive variance. Averages of these variables will be approximately normally distributed with mean μ and standard deviation σ/√n when n is large.

Statement of Theorem

3

How large of a sample size is required for the Central Limit Theorem (CLT) approximation to be good?

What is a ‘good’ approximation?

Questions

4

Permits analysis of random variables even when underlying distribution is unknown

Estimating parameters

Hypothesis Testing

Polling

Importance

5

Performing a hypothesis test to determine if set of data came from normal

Considerations◦ Power: probability that a test will reject the null

hypothesis when it is false

◦ Ease of Use

Testing for Normality

6

Problems◦ No test is desirable in every situation (no

universally most powerful test)

◦ Some lack ability to verify for composite hypothesis of normality (i.e. nonstandard normal)

◦ The reliability of tests is sensitive to sample size; with enough data, null hypothesis will be rejected

Testing for Normality

7

Symmetric

Unimodal

Bell-shaped

Continuous

Characteristics of Distribution

8

Skewness: Measures the asymmetry of a distribution.◦ Defined as the third standardized moment◦ Skew of normal distribution is 0

Closeness to Normal

3

1 E

X 3

1

3

)1( sn

XXn

ii

9

Kurtosis: Measures peakedness or heaviness of the tails.◦ Defined as the fourth standardized moment◦ Kurtosis of normal distribution is 3

Closeness to Normal

4

2 E

n

x

41

4

)1( sn

XXn

ii

10

Cumulative distribution function:

Binomial Distribution

X

i

iniin pppnxF

0

)1(C),;(

)1(][Var

][E

pnpX

npX

11

Binomial Distribution*paramete

rsKurtosi

sSkewnes

s% outside 1.96*sd

K-S distanc

e

MeanStd Dev

n = 20p = .2

-.0014(.25)

.3325(1.5)

.0434 .128 3.99991.786

n = 25p = .2

.002 .3013 .0743 .116 5.00072.002

n = 30p = .2

.0235 .2786 .0363 .106 5.9972.188

n = 50p = .2

.0106 .209 .0496 .083 10.0012.832

n = 100p = .2

.005 .149 .05988 .0574 19.9974.0055

*from R

12

Cumulative distribution function:

Uniform Distribution

ab

axbaxF

),;(

12

)(][Var

2][E

2abX

baX

13

Uniform Distribution*parameters Kurtosi

sSkewnes

s%

outside 1.96*sd

K-S distanc

e

MeanStd Dev

n = 5(a,b) = (0,1)

-.236(-1.2)

.004(0)

.0477 .0061 .4998.1289 (.129)

n = 5(a,b) = (0,50)

-.234 0 .04785 .0058 24.996.468 (6.455)

n = 5(a,b) = (0, .1)

-.238 -.0008 .048 .0060 .0500.0129 (.0129)

n = 3(a,b) = (0,50)

-.397 -.001 .0468 .01 24.998.326 (8.333)

*from R

14

Cumulative distribution function:

Exponential Distribution

xexF 1);(

2

1]Var[

1]E[

X

X

15

Exponential Distribution*

parameters

Kurtosis Skewness

% outside 1.96*sd

K-S distanc

e

MeanStd Dev

n = 5λ = 1

1.239(6)

.904(2)

.0434 .0598 .9995.4473 (.4472)

n = 10 .597 .630 .045 .042 1.0005.316 (.316)

n = 15 .396 .515 .0464 .034 .9997.258 (.2581)

*from R

16

Find n values for more distributions

Refine criteria for quality of approximation

Explore meanless distributions

Classify distributions in order to have more general guidelines for minimum sample size

For Next Time…

17

The Central Limit Theorem (Pt 2)

Paul CornwellMay 2, 2011

18

Central Limit Theorem: Averages of i.i.d. variables become normally distributed as sample size increases

Rate of converge depends on underlying distribution

What sample size is needed to produce a good approximation from the CLT?

Review

19

Real-life applications of the Central Limit Theorem

What does kurtosis tell us about a distribution?

What is the rationale for requiring np ≥ 5?

What about distributions with no mean?

Questions

20

Probability for total distance covered in a random walk tends towards normal

Hypothesis testing

Confidence intervals (polling)

Signal processing, noise cancellation

Applications of Theorem

21

Measures the “peakedness” of a distribution

Higher peaks means fatter tails

Kurtosis

3E

4

2

n

x

22

Traditional assumption for normality with binomial is np > 5 or 10

Skewness of binomial distribution increases as p moves away from .5

Larger n is required for convergence for skewed distributions

Why np?

23

Has no moments (including mean, variance)

Distribution of averages looks like regular distribution

CLT does not apply

Cauchy Distribution

)1(

1)(

2xxf

24

α = β = 1/3

Distribution is symmetric and bimodal

Convergence to normal is fast in averages

Beta Distribution

25

Heavier-tailed, bell-shaped curve

Approaches normal distribution as degrees of freedom increase

Student’s t Distribution

26

4 statistics: K-S distance, tail probabilities, skewness and kurtosis

Different thresholds for “adequate” and “superior” approximations

Both are fairly conservative

Criteria

27

Adequate Approximation

Distribution ∣Kurtosis∣ <.5

∣Skewness∣ <.25

Tail Prob. .04<x<.0

6

K-S Distance

<.05

max

Uniform 3 1 2 2 3

Beta (α=β=1/3)

4 1 3 3 4

Exponential 12 64 5 8 64

Binomial (p=.1)

11 114 14 332 332

Binomial (p=.5)

4 1 12 68 68

Student’s t with 2.5 df

NA NA 13 20 20

Student’s twith 4.1 df

120 1 1 2 120

28

Stronger Approximation

Distribution ∣Kurtosis∣ <.3

∣Skewness∣ <.15

Tail Prob. .04<x<.0

6

K-S Distance

<.02

max

Uniform 4 1 2 2 4

Beta (α=β=1/3)

6 1 3 4 6

Exponential 20 178 5 45 178

Binomial (p=.1)

18 317 14 1850 1850

Binomial (p=.5)

7 1 12 390 390

Student’s t with 2.5 df

NA NA 13 320 320

Student’s twith 4.1 df

200 1 1 5 200

29

Skewness is difficult to shake

Tail probabilities are fairly accurate for small sample sizes

Traditional recommendation is small for many common distributions

Conclusions