The inverse of the cumulative standard normal probability ...dominicd/dominici-invnorm.pdf · The...

14
The inverse of the cumulative standard normal probability function. Diego E. Dominici * Abstract Some properties of the inverse of the function N (x)= 1 2π x -∞ e - t 2 2 dt are studied. Its derivatives, integrals and asymptotic behavior are pre- sented. 1 Introduction It would be difficult to overestimate the importance of the standard normal (or Gauss) distribution. It finds widespread applications in almost every scientific discipline, e.g., probability theory, the theory of errors, heat conduction, biol- ogy, economics, physics, neural networks [10], etc. It plays a fundamental role in financial mathematics, being part of the Black-Scholes formula [2], and its inverse is used in computing the implied volatility of an option [9]. Yet, little is known about the properties of the inverse function, e.g., series expansions, asymptotic behavior, integral representations. The major work done has been in computing fast and accurate algorithms for numerical calculations [1]. Over the years a few articles have appeared with analytical studies of the closely related error function erf(x)= 2 π x 0 e -t 2 dt and its complement erfc(x)= 2 π x e -t 2 dt . Philip [11] introduced the notation “inverfc(x)” to denote the inverse of the complementary error function. He gave the first terms in the power series for inverfc(x), asymptotic formulas for small x in terms of continued logarithms, * Department of Mathematics, Statistics and computer Science, University of Illi- nois at Chicago (m/c 249), 851 South Morgan Street, Chicago, IL 60607-7045, USA ([email protected]) 1

Transcript of The inverse of the cumulative standard normal probability ...dominicd/dominici-invnorm.pdf · The...

Page 1: The inverse of the cumulative standard normal probability ...dominicd/dominici-invnorm.pdf · The inverse of the cumulative standard normal probability function. ... erf x √ 2 +1

The inverse of the cumulative standard normal

probability function.

Diego E. Dominici∗

Abstract

Some properties of the inverse of the function N(x) = 1√2π

∫ x

−∞ e−t22 dt

are studied. Its derivatives, integrals and asymptotic behavior are pre-sented.

1 Introduction

It would be difficult to overestimate the importance of the standard normal (orGauss) distribution. It finds widespread applications in almost every scientificdiscipline, e.g., probability theory, the theory of errors, heat conduction, biol-ogy, economics, physics, neural networks [10], etc. It plays a fundamental rolein financial mathematics, being part of the Black-Scholes formula [2], and itsinverse is used in computing the implied volatility of an option [9]. Yet, littleis known about the properties of the inverse function, e.g., series expansions,asymptotic behavior, integral representations. The major work done has beenin computing fast and accurate algorithms for numerical calculations [1].

Over the years a few articles have appeared with analytical studies of theclosely related error function

erf(x) =2√π

x∫

0

e−t2dt

and its complement

erfc(x) =2√π

∞∫

x

e−t2dt .

Philip [11] introduced the notation “inverfc(x)” to denote the inverse of thecomplementary error function. He gave the first terms in the power series forinverfc(x), asymptotic formulas for small x in terms of continued logarithms,

∗Department of Mathematics, Statistics and computer Science, University of Illi-nois at Chicago (m/c 249), 851 South Morgan Street, Chicago, IL 60607-7045, USA([email protected])

1

Page 2: The inverse of the cumulative standard normal probability ...dominicd/dominici-invnorm.pdf · The inverse of the cumulative standard normal probability function. ... erf x √ 2 +1

and some expressions for the derivatives and integrals. Carlitz [3], studied thearithmetic properties of the coefficients in the power series of inverfc(x). Strecok[12] computed the first 200 terms in the series of inverfc(x), and some expansionsin series of Chebyshev polynomials. Finally, Fettis [6] studied inverfc(x) forsmall x, using an iterative sequence of logarithms.

The purpose of this paper is to present some new results on the derivatives,integrals, and asymptotics of the inverse of the cumulative standard normalprobability function

N(x) =1√2π

∫ x

−∞e−

t22 dt

which we call S(x). The function N(x) is related to other important specialfunctions like the error function erf(x)

N(x) =12

[erf

(x√2

)+ 1

],

the parabolic cylinder function Dν(x)

N(x) =12

[2 −

√2π

e−x24 D−1(x)

],

and the incomplete Gamma function Γ(ν; x)

N(x) =12

[2 −

1√π

Γ(12,x2

2)]

.

In section 2 we derive an ODE satisfied by S(x), and solve it using a powerseries. We introduce a family of polynomials Pn related to the calculation ofhigher derivatives of S(x). In section 3 we study some properties of the Pn, suchas relations between coefficients, recurrences, and generating functions. We alsoderive a general formula for Pn using the idea of “nested derivatives”, and wecompare the Pn with the Hermite polynomials Hn.

In section 4 we extend the definition of the Pn to n < 0 and use them tocalculate the integrals of S(x). We also compute the integrals of powers of S(x)on the interval [0, 1]. Section 5 is dedicated to the asymptotic expansions ofS(x) for x → 0, x → 1, using the function Lambert W.

Finally, appendix A contains the first 20 non-zero coefficients in the seriesof S(x), and the first 10 polynomials Pn.

2 Derivatives

Definition 1 Let S(x) denote the inverse of

N(x) =1√2π

∫ x

−∞e−

t22 dt

2

Page 3: The inverse of the cumulative standard normal probability ...dominicd/dominici-invnorm.pdf · The inverse of the cumulative standard normal probability function. ... erf x √ 2 +1

satisfyingS ◦ N(x) = N ◦ S(x) = x (1)

Proposition 2 S(x) satisfies the IVP

S′′ = S(S′)2, S(

12

)= 0, S′ ( 1

2

)=

√2π (2)

Proof. Since N(0) = 1/2, in follows that S(1/2) = 0. From (1) we get

S′[N(x)] =1

N ′(x)=

√2πe

x22 =

√2πe

S2[N(x)]2

Substituting N(x) = y we have

S′(y) =√

2πeS2(y)

2 , S′(12) =

√2π

Differentiating ln[S′(y)], we get (2).

Proposition 3 The derivatives of S(x) obey the recurrence formula

S(n+2)(x) =n∑

i=0

i∑

j=0

(n

i

)(i

j

)S(n−i)(x)S(i−j+1)(x)S(j+1)(x), n ≥ 0.

Proof. Taking the nth derivative of (2), and using Leibnitz’s theorem, we have

S(n+2) =n∑

i=0

(n

i

)S(n−i)(S′S′)(i)

=n∑

i=0

(n

i

)S(n−i)

i∑

j=0

(i

j

)S(i−j+1)S(j+1)

=n∑

i=0

i∑

j=0

(n

i

)(i

j

)S(n−i)S(i−j+1)S(j+1).

Corollary 4 If Dn = dnSdxn ( 1

2 ) , then

D2n = 0, n ≥ 0.

PuttingDn = (2π)

n2 Cn, (3)

we can writeS(x) =

n≥0

(2π)2n+1

2C2n+1

(2n + 1)!(x −

12)2n+1

whereC1 = 1, C3 = 1, C5 = 7, C7 = 127, . . . .

3

Page 4: The inverse of the cumulative standard normal probability ...dominicd/dominici-invnorm.pdf · The inverse of the cumulative standard normal probability function. ... erf x √ 2 +1

Proposition 5S(n) = Pn−1(S)(S′)n n ≥ 1 (4)

where Pn(x) is a polynomial of degree n satisfying the forward recurrence

P0(x) = 1, Pn(x) = P ′n−1(x) + nxPn−1(x), n ≥ 1 (5)

so that

P1(x) = x, P2(x) = 1 + 2x2, P3(x) = 7x + 6x3, . . . .

Proof. We use induction on n. For n = 2 the result follows from (2). If weassume the result is true for n then

S(n+1) = [Pn−1(S)(S′)n]′

= P ′n−1(S)S′(S′)n + Pn−1(S)n(S′)n−1S′′

= P ′n−1(S)(S′)n+1 + Pn−1(S)n(S′)n−1S(S′)2

= [P ′n−1(S) + nSPn−1(S)](S′)n+1

= Pn(S)(S′)n+1

Since Pn−1(x) is a polynomial of degree n − 1 by hypothesis, is clear that

Pn(x) = P ′n−1(x) + nxPn−1(x)

is a polynomial of degree n.

Corollary 6 Setting x = 0 in (4), we obtain

Cn = Pn−1(0)

3 The polynomials Pn(x)

Lemma 7 If we write

Pn(x) =n∑

k=0

Qnkxk

we have

Qn0 = Qn−1

1

Qnk = nQn−1

k−1 + (k + 1)Qn−1k+1 , k = 1, . . . , n − 2 (6)

Qnk = nQn−1

k k = n − 1, n

In particular,Qn

n = n!

4

Page 5: The inverse of the cumulative standard normal probability ...dominicd/dominici-invnorm.pdf · The inverse of the cumulative standard normal probability function. ... erf x √ 2 +1

Proof.

n∑

k=0

Qnkxk

= Pn =d

dxPn−1 + nxPn−1

=n−1∑

k=0

Qn−1k kxk−1 +

n−1∑

k=0

nQn−1k xk+1

=n−2∑

k=0

Qn−1k+1 (k + 1)xk +

n∑

k=1

nQn−1k−1xk

Corollary 8 In matrix form (6) reads A(n)Qn−1 = Qn, where A(n) ∈ <(n+1) × n

is a tridiagonal matrix given by

A(n)i,j =

i, j = i + 1, i = 1, . . . , n − 1n, j = i − 1, i = 2, . . . , n + 1

0, ow

and

Qn =

Qn0

Qn1

Qn2...

Qnn

With the help of these matrices, we have an expression for the coefficientsof Pn(x)

Qn =n∏

k=1

A(n−k+1) = A(n)A(n−1) · · ·A(1)

Proposition 9 The polynomials Pn(x) satisfy the recurrence relation

Pn+1(x) =n∑

i=0

i∑

j=0

(n

i

)(i

j

)Pn−i−1(x)Pi−j (x)Pj(x)

5

Page 6: The inverse of the cumulative standard normal probability ...dominicd/dominici-invnorm.pdf · The inverse of the cumulative standard normal probability function. ... erf x √ 2 +1

Proof.

Pn+1(S)(S′)n+2

=S(n+2) =n∑

i=0

i∑

j=0

(n

i

)(i

j

)S(n−i)S(i−j+1)S(j+1)

=n∑

i=0

i∑

j=0

(n

i

)(i

j

)Pn−i−1(S)(S′)n−iPi−j(S)(S′)i−j+1Pj(S)(S′)j+1

= (S′)n+2n∑

i=0

i∑

j=0

(n

i

)(i

j

)Pn−i−1(S)Pi−j(S)Pj(S).

Proposition 10 The zeros of the polynomials Pn(x) are purely imaginary forn ≥ 1

Lemma 11 Proof. For n = 1 the result is obviously true. Assuming that it istrue for n, and that Pn(x) is written like

Pn(x) = n!n∏

k=1

(z − zk), Re(zk) = 0, 1 ≤ k ≤ n (7)

we have two possibilities for z∗, Pn+1(z∗) = 0:

1. z∗ = zk, for some 1 ≤ k ≤ n.

In this case, Re(z∗) = 0 and the lemma is proved.

2. z∗ 6= zk, for all 1 ≤ k ≤ n.

From (5) we get

Pn+1(x)Pn(x)

=d

dxln [Pn(x)] + (n + 1)x

=n∑

k=1

1z − zk

+ (n + 1)x

using (7). After evaluating at z = z∗ we obtain

0 =n∑

k=1

1z∗ − zk

+ (n + 1)z∗

6

Page 7: The inverse of the cumulative standard normal probability ...dominicd/dominici-invnorm.pdf · The inverse of the cumulative standard normal probability function. ... erf x √ 2 +1

and taking Re(•)

0 = Re

[n∑

k=1

1z∗ − zk

+ (n + 1)z∗]

=n∑

k=1

Re (z∗ − zk)|z∗ − zk|2

+ (n + 1) Re(z∗)

= Re(z∗)

[n∑

k=1

1

|z∗ − zk|2+ (n + 1)

]

which implies that Re(z∗) = 0.

Proposition 12 The exponential generating function of the polynomials Pn(x)is ∑

k≥0

Pk(x)tk

k!= e

12 S2[N(x)+tN ′(x)]− x2

2

Proof. Since

F (x, t) =∑

k≥0

Pk(x)tk

k!

= 1 +∑

k≥1

d

dxPk−1(x)

tk

k!+

k≥1

kxPk−1(x)tk

k!

= 1 +∑

k≥0

d

dxPk(x)

tk+1

(k + 1)k!+

k≥0

xPk(x)tk+1

k!

= 1 +∫ t

0

∂xF (x, s)ds + xtF (x, t)

it follows that F (x, t) satisfies the differential-integral equation

1 + (xt − 1)F (x, t) +∂

∂x

∫ t

0

F (x, s)ds = 0 (8)

Differentiating (8) with respect to t we get

xF (x, t) + (xt − 1)∂

∂tF (x, t) +

∂xF (x, t) = 0

whose general solution is of the form

F (x, t) = e−x22 G

(te−

x22 +

√2π

[N(x) − 1

2

])

for some function G(z).

7

Page 8: The inverse of the cumulative standard normal probability ...dominicd/dominici-invnorm.pdf · The inverse of the cumulative standard normal probability function. ... erf x √ 2 +1

From (8) we know that F (x, 0) = 1, and hence

G

(√2π

[N(x) − 1

2

])= e

x22 ,

which implies that

G(z) = e12 S2

(z√2π

+ 12

).

Therefore,F (x, t) = e

12 S2[N(x)+tN ′(x)]−x2

2

Definition 13 We define the “nested derivative” D(n) by

D(0)[f ](x) ≡ 1

D(n)[f ] (x) =d

dx

{f(x) × D(n−1)[f ] (x)

}, n ≥ 1

Example 14 The following examples serve to illustrate the calculation of the“nested derivatives” for some elementary functions.

1.D(n) [eax] = n!anenax

2.D(n) [x] = 1

3.D(n)

[x2

]= (n + 1)!xn

Proposition 15 With the help of the “nested derivatives” we can represent thepolynomials Pn(x) by

Pn(x) = e−n2 x2

D(n)[e

x22

]

Proof. We use induction on n. For n = 0 the result follows from the definitionof D(n). Assuming the result is true for n − 1

Pn(x) = P ′n−1(x) + nxPn−1(x)

=d

dx[e−

(n−1)2 x2

D(n−1)(ex22 )] + nxe−

(n−1)2 x2

D(n−1)(ex22 )

= −(n − 1)xe−(n−1)

2 x2D(n−1)(e

x22 ) + e−

(n−1)2 x2 d

dx[D(n)(e

x22 )]+

+ nxe−(n−1)

2 x2D(n−1)(e

x22 )

= e−(n−1)

2 x2[xD(n−1)(e

x22 ) +

d

dxD(n−1)(e

x22 )

]

= e−(n−1)

2 x2e−

12 x2 d

dx

[e

12 x2

D(n−1)(ex22 )

]

= e−n2 x2

D(n)[e

x22

]

8

Page 9: The inverse of the cumulative standard normal probability ...dominicd/dominici-invnorm.pdf · The inverse of the cumulative standard normal probability function. ... erf x √ 2 +1

Summary 16 We conclude this section by comparing the properties of Pn(x)with the well known formulas for the Hermite polynomials Hn(x) [8]. Sincethe Hn are deeply related to the function N(x), we would expect to see somesimilarities between the Hn and the Pn.

Pn(x) Hn(x)

Pn(x) = e−n2 x2

D(n)(ex22 ) Hn(x) = (−1)nex2 dn

dxn (e−x2)

∑k≥0

Pk(x) tk

k! = e12 S2[N(x)+tN ′(x)]−x2

2∑k≥0

Hk(x) tk

k! = e2xt−t2

Pn(x) = P ′n−1(x) + nxPn−1(x) Hn(x) = −H ′

n−1(x) + 2xHn−1(x)

S(n) = Pn−1(S)(S′)n N (n) =(

−1√2

)n−1

Hn−1

(x√2

)N ′

4 Integrals of S(x)

Definition 17 We define the nth repeated integral of S(x) by

S(−n)(x) =

x∫

0

x1∫

0

· · ·xn−1∫

0

S(xn) dxndxn−1 . . . dx1, n ≥ 1.

Lemma 18 The backward recurrence formula for the Pn(x) is

Pn−1(x) = e−n2 x2

Pn−1(0) +

x∫

0

en2 t2Pn(t)dt

(9)

Proof. It follows immediately from solving the ODE for Pn−1 in terms of Pn.

Proposition 19 Using (9) to define Pn(x) for n < 0 yields

P−1(x) = x

P−2(x) = −1 (10)

P−3(x) = −√

πex2N

(√2x

)

and the relationS(n) = Pn−1(S)(S′)n

still holds.

9

Page 10: The inverse of the cumulative standard normal probability ...dominicd/dominici-invnorm.pdf · The inverse of the cumulative standard normal probability function. ... erf x √ 2 +1

Proof. For n = 0 we have

P−1(x) = P−1(0) + x

S = S(0) = P−1(S)

so P−1(0) = 0. For n = −1

P−2(x) = ex22 [P−2(0) + 1] − 1

We can calculate S(−1) explicitly by

S(−1)(x) =

x∫

0

S(t)dt =1√2π

S(x)∫

−∞

S[N(z)]e−z22 dz

=1√2π

S(x)∫

−∞

ze−z22 dz

= − 1√2π

e−S(x)2

2 = −(S′)−1

Hence, P−2(0) = −1.Finally, for n = −2

P−3(x) = ex2[P−3(0) −

√πN(

√2x) +

√π

2

]

A similar calculation as the one above, making a change of variables t = N(z)in the integral of S(−1)(x) yields

S(−2)(x) = − 12√

πN [

√2S(x)]

and we conclude that P−3(0) = −√

π2 .

Corollary 20 The function S(x) satisfies the functional relations

S′(x)S(−1)(x) = −1 (11)

S[−2

√πS(−2)(x)

]=

√2S(x)

Corollary 21 The coefficients Cn satisfy the recurrence relation

Cn+1 =n−1∑

j=0

(n

j + 1

)CjCn−j , n ≥ 1

10

Page 11: The inverse of the cumulative standard normal probability ...dominicd/dominici-invnorm.pdf · The inverse of the cumulative standard normal probability function. ... erf x √ 2 +1

Proof. Taking a derivative in (11), we have

0 =n∑

k=0

(n

k

)S(k−1)S(n−k+1)

= S(−1)S(n+1) +n−1∑

j=0

(n

j + 1

)S(j)S(n−j)

Evaluating at x = 12 , we obtain

n−1∑

j=0

(n

j + 1

)S(j)

(12

)S(n−j)

(12

)=

1√2π

S(n+1)

(12

)

and the result follows from (3) after dividing both sides by (2π)n.

Proposition 22 The integral of the nth power of S(x) is given by the formula

1∫

0

Sn(x)dx =

k∏i=1

(2i + 1), n = 2k, k ≥ 1

0, n = 2k + 1, k ≥ 0

Proof.1∫

0

Sn(x)dx =

∞∫

−∞

znN ′(z)dz

=1√2π

∞∫

−∞

zne−12 z2

dz

The last integral can be computed exactly [7], to yield the result.

5 Asymptotics

Definition 23 We’ll denote by LW(x) the function Lambert W [4],

LW(x)eLW(x) = x (12)

This function has the series representation [5]

LW(x) =∑

n≥1

(−n)n−1

n!xn,

the derivatived

dxLW =

LW(x)x[1 + LW(x)]

if x 6= 0,

and it has the asymptotic behavior

LW(x) ∼ ln(x) − ln[ln(x)] x → ∞.

11

Page 12: The inverse of the cumulative standard normal probability ...dominicd/dominici-invnorm.pdf · The inverse of the cumulative standard normal probability function. ... erf x √ 2 +1

Proposition 24

S(x) ∼ g0(x) = −

√LW

(1

2πx2

), x → 0

S(x) ∼ g1(x) =

√LW

(1

2π(x − 1)2

), x → 1

Both functions g0(x) and g1(x) satisfy the ODE

g′′ = g(g′)2[1 +

2g2(1 + g2)

]∼ g(g′)2, for |g| → ∞

Proof.

N(x) ∼ 1√2π

e−x22

1x

, x → −∞

t ∼ 1√2π

e−S(t)2

21

S(t), t → 0

S(t)eS(t)2

2 ∼ 1√2πt

, t → 0

S2(t)eS2(t) ∼1

2πt2, t → 0

Using the definition of LW(x) we have

S2(t) ∼ LW(

12πt2

), t → 0

or

S(t) ∼ −

√LW

(1

2πt2

), t → 0

The case x → 1 is completely analogous.

Corollary 25 Combining the above expressions, we get

S(x) ∼ (2x − 1)

√LW

(1

2πx2(x − 1)2

), x → 0, x → 1 (13)

Conclusion 26 We have presented some results on the operations of the cal-culus, intrarelationships, and asymptotic analysis of the function S(x) and afamily of polynomials associated with it. The objective of this work is to providea reference for researchers from different disciplines, who may encounter theNormal distribution in one of its many applications.

Acknowledgement 27 We thank Professor Charles Knessl for his valuablecomments on earlier drafts. This work was supported in part by NSF grant99-73231, provided by Professor Floyd Hanson. We wish to thank him for hisgenerous sponsorship.

12

Page 13: The inverse of the cumulative standard normal probability ...dominicd/dominici-invnorm.pdf · The inverse of the cumulative standard normal probability function. ... erf x √ 2 +1

6 Appendix

The first 10 Pn(x) are

P0(x) = 1P1(x) = x

P2(x) = 1 + 2x2

P3(x) = 7x + 6x3

P4(x) = 7 + 46x2 + 24x4

P5(x) = 127x + 326x3 + 120x5

P6(x) = 127 + 1740x2 + 2556x4 + 720x6

P7(x) = 4369x + 22404x3 + 22212x5 + 5040x7

P8(x) = 4369 + 102164x2 + 290292x4 + 212976x6 + 40320x8

P9(x) = 243649x + 2080644x3 + 3890484x5 + 2239344x7 + 362880x9

P10(x) = 243649 + 8678422x2 + 40258860x4 + 54580248x6+

+ 25659360x8 + 3628800x10

The first few odd Cn aren Cn

1 13 15 77 1279 436911 24364913 2003698315 228035686317 34314143376119 6596724120000121 1577346142379376723 459122712323094540725 159835173324760985284927 65578224979953171437548929 31316040486497385233866978331 17220166851265734645512645734333 10802634947676204112783980061728135 7668370196972678030742096890473344137 6115467419532433012529577853117243872739 5444102953057402868740275358627854939660741 53789884101606550209324949796685518122943569

13

Page 14: The inverse of the cumulative standard normal probability ...dominicd/dominici-invnorm.pdf · The inverse of the cumulative standard normal probability function. ... erf x √ 2 +1

References

[1] J. M. Blair, C. A. Edwards, and J. H. Johnson. Rational Chebyshev approx-imations for the inverse of the error function. Math. Comp., 30(136):827–830, 1976.

[2] Fisher Black and Myron S Scholes. The pricing of options and corporateliabilities. Journal of Political Economy, 81(3):637–654, 1973.

[3] L. Carlitz. The inverse of the error function. Pacific J. Math., 13:459–470,1963.

[4] R. M. Corless, G. H. Gonnet, D. E. G. Hare, D. J. Jeffrey, and D. E. Knuth.On the Lambert W function. Adv. Comput. Math., 5(4):329–359, 1996.

[5] Robert M. Corless, David J. Jeffrey, and Donald E. Knuth. A sequenceof series for the Lambert W function. In Proceedings of the 1997 Inter-national Symposium on Symbolic and Algebraic Computation (Kihei, HI),pages 197–204 (electronic), New York, 1997. ACM.

[6] Henry E. Fettis. A stable algorithm for computing the inverse error functionin the “tail-end” region. Math. Comp., 28:585–587, 1974.

[7] Gradshteyn, I. S. and Ryzhik, I. M., Table of integrals, series, and prod-ucts, Sixth edition, Translated from the Russian, Academic Press Inc., SanDiego, CA, 2000.

[8] N. N. Lebedev. Special functions and their applications. Prentice-Hall Inc.,Englewood Cliffs, N.J., 1965.

[9] C.F. Lee and A. Tucker. An alternative method for obtaining the impliedstandard deviation. The Journal of Financial Engineering, 1:369–375, 1992.

[10] A. Menon, K. Mehrotra, C. K. Mohan, and S. Ranka. Characterization ofa class of sigmoid functions with applications to neural networks. NeuralNetworks, 9(5):819–835, 1996.

[11] J. R. Philip. The function inverfc θ. Austral. J. Phys., 13:13–20, 1960.

[12] Anthony Strecok. On the calculation of the inverse of the error function.Math. Comp., 22:144–158, 1968.

14