Dave Goldsman - gatech.edu

Post on 08-Feb-2022

10 views 0 download

Transcript of Dave Goldsman - gatech.edu

4. Distributions

Dave Goldsman

H. Milton Stewart School of Industrial and Systems EngineeringGeorgia Institute of Technology

8/5/20

ISYE 6739 — Goldsman 8/5/20 1 / 108

Outline

1 Bernoulli and Binomial Distributions2 Hypergeometric Distribution3 Geometric and Negative Binomial Distributions4 Poisson Distribution5 Uniform, Exponential, and Friends6 Other Continuous Distributions7 Normal Distribution: Basics8 Standard Normal Distribution9 Sample Mean of Normals

10 The Central Limit Theorem + Proof11 Central Limit Theorem Examples12 Extensions — Multivariate Normal Distribution13 Extensions — Lognormal Distribution14 Computer Stuff

ISYE 6739 — Goldsman 8/5/20 2 / 108

Bernoulli and Binomial Distributions

Lesson 4.1 — Bernoulli and Binomial Distributions

Goal: We’ll discuss lots of interesting distributions in this module.

The module will be a compendium of results, some of which we’ll prove, andsome of which we’ve already seen previously.

Special emphasis will be placed on the Normal distribution, because it’s soimportant and has so many implications, including the Central LimitTheorem.

In the next few lessons we’ll discuss some important discrete distributions:

Bernoulli and Binomial Distributions

Hypergeometric Distribution

Geometric and Negative Binomial Distributions

Poisson Distribution

ISYE 6739 — Goldsman 8/5/20 3 / 108

Bernoulli and Binomial Distributions

Lesson 4.1 — Bernoulli and Binomial Distributions

Goal: We’ll discuss lots of interesting distributions in this module.

The module will be a compendium of results, some of which we’ll prove, andsome of which we’ve already seen previously.

Special emphasis will be placed on the Normal distribution, because it’s soimportant and has so many implications, including the Central LimitTheorem.

In the next few lessons we’ll discuss some important discrete distributions:

Bernoulli and Binomial Distributions

Hypergeometric Distribution

Geometric and Negative Binomial Distributions

Poisson Distribution

ISYE 6739 — Goldsman 8/5/20 3 / 108

Bernoulli and Binomial Distributions

Lesson 4.1 — Bernoulli and Binomial Distributions

Goal: We’ll discuss lots of interesting distributions in this module.

The module will be a compendium of results, some of which we’ll prove, andsome of which we’ve already seen previously.

Special emphasis will be placed on the Normal distribution, because it’s soimportant and has so many implications, including the Central LimitTheorem.

In the next few lessons we’ll discuss some important discrete distributions:

Bernoulli and Binomial Distributions

Hypergeometric Distribution

Geometric and Negative Binomial Distributions

Poisson Distribution

ISYE 6739 — Goldsman 8/5/20 3 / 108

Bernoulli and Binomial Distributions

Lesson 4.1 — Bernoulli and Binomial Distributions

Goal: We’ll discuss lots of interesting distributions in this module.

The module will be a compendium of results, some of which we’ll prove, andsome of which we’ve already seen previously.

Special emphasis will be placed on the Normal distribution, because it’s soimportant and has so many implications, including the Central LimitTheorem.

In the next few lessons we’ll discuss some important discrete distributions:

Bernoulli and Binomial Distributions

Hypergeometric Distribution

Geometric and Negative Binomial Distributions

Poisson Distribution

ISYE 6739 — Goldsman 8/5/20 3 / 108

Bernoulli and Binomial Distributions

Lesson 4.1 — Bernoulli and Binomial Distributions

Goal: We’ll discuss lots of interesting distributions in this module.

The module will be a compendium of results, some of which we’ll prove, andsome of which we’ve already seen previously.

Special emphasis will be placed on the Normal distribution, because it’s soimportant and has so many implications, including the Central LimitTheorem.

In the next few lessons we’ll discuss some important discrete distributions:

Bernoulli and Binomial Distributions

Hypergeometric Distribution

Geometric and Negative Binomial Distributions

Poisson Distribution

ISYE 6739 — Goldsman 8/5/20 3 / 108

Bernoulli and Binomial Distributions

Definition: The Bernoulli distribution with parameter p is given by

X =

{1 w.p. p (“success”)

0 w.p. q (“failure”)

Recall: E[X] = p, Var(X) = pq, and MX(t) = pet + q.

Definition: The Binomial distribution with parameters n and p is givenby

P (Y = k) =

(n

k

)pkqn−k, k = 0, 1, . . . , n.

ISYE 6739 — Goldsman 8/5/20 4 / 108

Bernoulli and Binomial Distributions

Definition: The Bernoulli distribution with parameter p is given by

X =

{1 w.p. p (“success”)

0 w.p. q (“failure”)

Recall: E[X] = p, Var(X) = pq, and MX(t) = pet + q.

Definition: The Binomial distribution with parameters n and p is givenby

P (Y = k) =

(n

k

)pkqn−k, k = 0, 1, . . . , n.

ISYE 6739 — Goldsman 8/5/20 4 / 108

Bernoulli and Binomial Distributions

Definition: The Bernoulli distribution with parameter p is given by

X =

{1 w.p. p (“success”)

0 w.p. q (“failure”)

Recall: E[X] = p, Var(X) = pq, and MX(t) = pet + q.

Definition: The Binomial distribution with parameters n and p is givenby

P (Y = k) =

(n

k

)pkqn−k, k = 0, 1, . . . , n.

ISYE 6739 — Goldsman 8/5/20 4 / 108

Bernoulli and Binomial Distributions

Example: Toss 2 dice and take the sum; repeat 5 times. Let Y be the numberof 7’s you see.

Y ∼ Bin(5, 1/6). Then, e.g.,

P (Y = 4) =

(5

4

)(1

6

)4(5

6

)5−4. 2

Theorem: X1, . . . , Xniid∼ Bern(p)⇒ Y ≡

∑ni=1Xi ∼ Bin(n, p).

Proof: This kind of result can easily be proved by a moment generatingfunction uniqueness argument, as we mentioned in previous modules. 2

Think of the Binomial as the number of successes from n Bern(p) trials.

ISYE 6739 — Goldsman 8/5/20 5 / 108

Bernoulli and Binomial Distributions

Example: Toss 2 dice and take the sum; repeat 5 times. Let Y be the numberof 7’s you see. Y ∼ Bin(5, 1/6). Then, e.g.,

P (Y = 4) =

(5

4

)(1

6

)4(5

6

)5−4. 2

Theorem: X1, . . . , Xniid∼ Bern(p)⇒ Y ≡

∑ni=1Xi ∼ Bin(n, p).

Proof: This kind of result can easily be proved by a moment generatingfunction uniqueness argument, as we mentioned in previous modules. 2

Think of the Binomial as the number of successes from n Bern(p) trials.

ISYE 6739 — Goldsman 8/5/20 5 / 108

Bernoulli and Binomial Distributions

Example: Toss 2 dice and take the sum; repeat 5 times. Let Y be the numberof 7’s you see. Y ∼ Bin(5, 1/6). Then, e.g.,

P (Y = 4) =

(5

4

)(1

6

)4(5

6

)5−4. 2

Theorem: X1, . . . , Xniid∼ Bern(p)⇒ Y ≡

∑ni=1Xi ∼ Bin(n, p).

Proof: This kind of result can easily be proved by a moment generatingfunction uniqueness argument, as we mentioned in previous modules. 2

Think of the Binomial as the number of successes from n Bern(p) trials.

ISYE 6739 — Goldsman 8/5/20 5 / 108

Bernoulli and Binomial Distributions

Example: Toss 2 dice and take the sum; repeat 5 times. Let Y be the numberof 7’s you see. Y ∼ Bin(5, 1/6). Then, e.g.,

P (Y = 4) =

(5

4

)(1

6

)4(5

6

)5−4. 2

Theorem: X1, . . . , Xniid∼ Bern(p)⇒ Y ≡

∑ni=1Xi ∼ Bin(n, p).

Proof: This kind of result can easily be proved by a moment generatingfunction uniqueness argument, as we mentioned in previous modules. 2

Think of the Binomial as the number of successes from n Bern(p) trials.

ISYE 6739 — Goldsman 8/5/20 5 / 108

Bernoulli and Binomial Distributions

Example: Toss 2 dice and take the sum; repeat 5 times. Let Y be the numberof 7’s you see. Y ∼ Bin(5, 1/6). Then, e.g.,

P (Y = 4) =

(5

4

)(1

6

)4(5

6

)5−4. 2

Theorem: X1, . . . , Xniid∼ Bern(p)⇒ Y ≡

∑ni=1Xi ∼ Bin(n, p).

Proof: This kind of result can easily be proved by a moment generatingfunction uniqueness argument, as we mentioned in previous modules. 2

Think of the Binomial as the number of successes from n Bern(p) trials.

ISYE 6739 — Goldsman 8/5/20 5 / 108

Bernoulli and Binomial Distributions

Example: Toss 2 dice and take the sum; repeat 5 times. Let Y be the numberof 7’s you see. Y ∼ Bin(5, 1/6). Then, e.g.,

P (Y = 4) =

(5

4

)(1

6

)4(5

6

)5−4. 2

Theorem: X1, . . . , Xniid∼ Bern(p)⇒ Y ≡

∑ni=1Xi ∼ Bin(n, p).

Proof: This kind of result can easily be proved by a moment generatingfunction uniqueness argument, as we mentioned in previous modules. 2

Think of the Binomial as the number of successes from n Bern(p) trials.

ISYE 6739 — Goldsman 8/5/20 5 / 108

Bernoulli and Binomial Distributions

Theorem: Y ∼ Bin(n, p) implies

E[Y ] = E[ n∑i=1

Xi

]=

n∑i=1

E[Xi] = np.

Similarly,Var(Y ) = npq.

We’ve already seen that

MY (t) = (pet + q)n.

Theorem: Certain Binomials add up: If Y1, . . . , Yk are independent andYi ∼ Bin(ni, p), then

k∑i=1

Yi ∼ Bin( k∑i=1

ni, p).

ISYE 6739 — Goldsman 8/5/20 6 / 108

Bernoulli and Binomial Distributions

Theorem: Y ∼ Bin(n, p) implies

E[Y ] = E[ n∑i=1

Xi

]=

n∑i=1

E[Xi] = np.

Similarly,Var(Y ) = npq.

We’ve already seen that

MY (t) = (pet + q)n.

Theorem: Certain Binomials add up: If Y1, . . . , Yk are independent andYi ∼ Bin(ni, p), then

k∑i=1

Yi ∼ Bin( k∑i=1

ni, p).

ISYE 6739 — Goldsman 8/5/20 6 / 108

Bernoulli and Binomial Distributions

Theorem: Y ∼ Bin(n, p) implies

E[Y ] = E[ n∑i=1

Xi

]=

n∑i=1

E[Xi] = np.

Similarly,Var(Y ) = npq.

We’ve already seen that

MY (t) = (pet + q)n.

Theorem: Certain Binomials add up: If Y1, . . . , Yk are independent andYi ∼ Bin(ni, p), then

k∑i=1

Yi ∼ Bin( k∑i=1

ni, p).

ISYE 6739 — Goldsman 8/5/20 6 / 108

Bernoulli and Binomial Distributions

Theorem: Y ∼ Bin(n, p) implies

E[Y ] = E[ n∑i=1

Xi

]=

n∑i=1

E[Xi] = np.

Similarly,Var(Y ) = npq.

We’ve already seen that

MY (t) = (pet + q)n.

Theorem: Certain Binomials add up: If Y1, . . . , Yk are independent andYi ∼ Bin(ni, p),

then

k∑i=1

Yi ∼ Bin( k∑i=1

ni, p).

ISYE 6739 — Goldsman 8/5/20 6 / 108

Bernoulli and Binomial Distributions

Theorem: Y ∼ Bin(n, p) implies

E[Y ] = E[ n∑i=1

Xi

]=

n∑i=1

E[Xi] = np.

Similarly,Var(Y ) = npq.

We’ve already seen that

MY (t) = (pet + q)n.

Theorem: Certain Binomials add up: If Y1, . . . , Yk are independent andYi ∼ Bin(ni, p), then

k∑i=1

Yi ∼ Bin( k∑i=1

ni, p).

ISYE 6739 — Goldsman 8/5/20 6 / 108

Hypergeometric Distribution

1 Bernoulli and Binomial Distributions

2 Hypergeometric Distribution3 Geometric and Negative Binomial Distributions

4 Poisson Distribution5 Uniform, Exponential, and Friends6 Other Continuous Distributions7 Normal Distribution: Basics8 Standard Normal Distribution9 Sample Mean of Normals

10 The Central Limit Theorem + Proof

11 Central Limit Theorem Examples

12 Extensions — Multivariate Normal Distribution13 Extensions — Lognormal Distribution

14 Computer Stuff

ISYE 6739 — Goldsman 8/5/20 7 / 108

Hypergeometric Distribution

Lesson 4.2 — Hypergeometric Distribution

Definition: You have a objects of type 1 and b objects of type 2.

Select n objects without replacement from the a+ b.

Let X be the number of type 1’s selected. Then X has the Hypergeometricdistribution with pmf

P (X = k) =

(ak

)(b

n−k)(

a+bn

) , k = 0, 1, . . . ,min(a, n− b, n).

Example: 25 sox in a box. 15 red, 10 blue. Pick 7 w/o replacement.

P (exactly 3 reds are picked) =

(153

)(104

)(257

) . 2

ISYE 6739 — Goldsman 8/5/20 8 / 108

Hypergeometric Distribution

Lesson 4.2 — Hypergeometric Distribution

Definition: You have a objects of type 1 and b objects of type 2.

Select n objects without replacement from the a+ b.

Let X be the number of type 1’s selected. Then X has the Hypergeometricdistribution with pmf

P (X = k) =

(ak

)(b

n−k)(

a+bn

) , k = 0, 1, . . . ,min(a, n− b, n).

Example: 25 sox in a box. 15 red, 10 blue. Pick 7 w/o replacement.

P (exactly 3 reds are picked) =

(153

)(104

)(257

) . 2

ISYE 6739 — Goldsman 8/5/20 8 / 108

Hypergeometric Distribution

Lesson 4.2 — Hypergeometric Distribution

Definition: You have a objects of type 1 and b objects of type 2.

Select n objects without replacement from the a+ b.

Let X be the number of type 1’s selected. Then X has the Hypergeometricdistribution with pmf

P (X = k) =

(ak

)(b

n−k)(

a+bn

) , k = 0, 1, . . . ,min(a, n− b, n).

Example: 25 sox in a box. 15 red, 10 blue. Pick 7 w/o replacement.

P (exactly 3 reds are picked) =

(153

)(104

)(257

) . 2

ISYE 6739 — Goldsman 8/5/20 8 / 108

Hypergeometric Distribution

Lesson 4.2 — Hypergeometric Distribution

Definition: You have a objects of type 1 and b objects of type 2.

Select n objects without replacement from the a+ b.

Let X be the number of type 1’s selected. Then X has the Hypergeometricdistribution with pmf

P (X = k) =

(ak

)(b

n−k)(

a+bn

) , k = 0, 1, . . . ,min(a, n− b, n).

Example: 25 sox in a box. 15 red, 10 blue. Pick 7 w/o replacement.

P (exactly 3 reds are picked) =

(153

)(104

)(257

) . 2

ISYE 6739 — Goldsman 8/5/20 8 / 108

Hypergeometric Distribution

Lesson 4.2 — Hypergeometric Distribution

Definition: You have a objects of type 1 and b objects of type 2.

Select n objects without replacement from the a+ b.

Let X be the number of type 1’s selected. Then X has the Hypergeometricdistribution with pmf

P (X = k) =

(ak

)(b

n−k)(

a+bn

) , k = 0, 1, . . . ,min(a, n− b, n).

Example: 25 sox in a box. 15 red, 10 blue. Pick 7 w/o replacement.

P (exactly 3 reds are picked) =

(153

)(104

)(257

) . 2

ISYE 6739 — Goldsman 8/5/20 8 / 108

Hypergeometric Distribution

Lesson 4.2 — Hypergeometric Distribution

Definition: You have a objects of type 1 and b objects of type 2.

Select n objects without replacement from the a+ b.

Let X be the number of type 1’s selected. Then X has the Hypergeometricdistribution with pmf

P (X = k) =

(ak

)(b

n−k)(

a+bn

) , k = 0, 1, . . . ,min(a, n− b, n).

Example: 25 sox in a box. 15 red, 10 blue. Pick 7 w/o replacement.

P (exactly 3 reds are picked) =

(153

)(104

)(257

) . 2

ISYE 6739 — Goldsman 8/5/20 8 / 108

Hypergeometric Distribution

Lesson 4.2 — Hypergeometric Distribution

Definition: You have a objects of type 1 and b objects of type 2.

Select n objects without replacement from the a+ b.

Let X be the number of type 1’s selected. Then X has the Hypergeometricdistribution with pmf

P (X = k) =

(ak

)(b

n−k)(

a+bn

) , k = 0, 1, . . . ,min(a, n− b, n).

Example: 25 sox in a box. 15 red, 10 blue. Pick 7 w/o replacement.

P (exactly 3 reds are picked) =

(153

)(104

)(257

) . 2

ISYE 6739 — Goldsman 8/5/20 8 / 108

Hypergeometric Distribution

Theorem: After some algebra, it turns out that

E[X] = n

(a

a+ b

)and

Var(X) = n

(a

a+ b

)(1− a

a+ b

)(a+ b− na+ b− 1

).

Remark: Here, aa+b here plays the role of p in the Binomial distribution.

And then the corresponding Y ∼ Bin(n, p) results would be

E[Y ] = n

(a

a+ b

)and

Var(Y ) = n

(a

a+ b

)(1− a

a+ b

).

So same mean as the Hypergeometric, but slightly larger variance.

ISYE 6739 — Goldsman 8/5/20 9 / 108

Hypergeometric Distribution

Theorem: After some algebra, it turns out that

E[X] = n

(a

a+ b

)and

Var(X) = n

(a

a+ b

)(1− a

a+ b

)(a+ b− na+ b− 1

).

Remark: Here, aa+b here plays the role of p in the Binomial distribution.

And then the corresponding Y ∼ Bin(n, p) results would be

E[Y ] = n

(a

a+ b

)and

Var(Y ) = n

(a

a+ b

)(1− a

a+ b

).

So same mean as the Hypergeometric, but slightly larger variance.

ISYE 6739 — Goldsman 8/5/20 9 / 108

Hypergeometric Distribution

Theorem: After some algebra, it turns out that

E[X] = n

(a

a+ b

)and

Var(X) = n

(a

a+ b

)(1− a

a+ b

)(a+ b− na+ b− 1

).

Remark: Here, aa+b here plays the role of p in the Binomial distribution.

And then the corresponding Y ∼ Bin(n, p) results would be

E[Y ] = n

(a

a+ b

)and

Var(Y ) = n

(a

a+ b

)(1− a

a+ b

).

So same mean as the Hypergeometric, but slightly larger variance.

ISYE 6739 — Goldsman 8/5/20 9 / 108

Hypergeometric Distribution

Theorem: After some algebra, it turns out that

E[X] = n

(a

a+ b

)and

Var(X) = n

(a

a+ b

)(1− a

a+ b

)(a+ b− na+ b− 1

).

Remark: Here, aa+b here plays the role of p in the Binomial distribution.

And then the corresponding Y ∼ Bin(n, p) results would be

E[Y ] = n

(a

a+ b

)and

Var(Y ) = n

(a

a+ b

)(1− a

a+ b

).

So same mean as the Hypergeometric, but slightly larger variance.

ISYE 6739 — Goldsman 8/5/20 9 / 108

Hypergeometric Distribution

Theorem: After some algebra, it turns out that

E[X] = n

(a

a+ b

)and

Var(X) = n

(a

a+ b

)(1− a

a+ b

)(a+ b− na+ b− 1

).

Remark: Here, aa+b here plays the role of p in the Binomial distribution.

And then the corresponding Y ∼ Bin(n, p) results would be

E[Y ] = n

(a

a+ b

)and

Var(Y ) = n

(a

a+ b

)(1− a

a+ b

).

So same mean as the Hypergeometric, but slightly larger variance.

ISYE 6739 — Goldsman 8/5/20 9 / 108

Hypergeometric Distribution

Theorem: After some algebra, it turns out that

E[X] = n

(a

a+ b

)and

Var(X) = n

(a

a+ b

)(1− a

a+ b

)(a+ b− na+ b− 1

).

Remark: Here, aa+b here plays the role of p in the Binomial distribution.

And then the corresponding Y ∼ Bin(n, p) results would be

E[Y ] = n

(a

a+ b

)and

Var(Y ) = n

(a

a+ b

)(1− a

a+ b

).

So same mean as the Hypergeometric, but slightly larger variance.ISYE 6739 — Goldsman 8/5/20 9 / 108

Geometric and Negative Binomial Distributions

1 Bernoulli and Binomial Distributions

2 Hypergeometric Distribution3 Geometric and Negative Binomial Distributions

4 Poisson Distribution5 Uniform, Exponential, and Friends6 Other Continuous Distributions7 Normal Distribution: Basics8 Standard Normal Distribution9 Sample Mean of Normals

10 The Central Limit Theorem + Proof

11 Central Limit Theorem Examples

12 Extensions — Multivariate Normal Distribution13 Extensions — Lognormal Distribution

14 Computer Stuff

ISYE 6739 — Goldsman 8/5/20 10 / 108

Geometric and Negative Binomial Distributions

Lesson 4.3 — Geometric and Negative Binomial Distributions

Definition: Suppose we consider an infinite sequence of independentBern(p) trials.

Let Z equal the number of trials until the first success is obtained. The eventZ = k corresponds to k − 1 failures, and then a success. Thus,

P (Z = k) = qk−1p, k = 1, 2, . . . ,

and we say that Z has the Geometric distribution with parameter p.

Notation: X ∼ Geom(p).

We’ll get the mean and variance of the Geometric via the mgf. . . .

ISYE 6739 — Goldsman 8/5/20 11 / 108

Geometric and Negative Binomial Distributions

Lesson 4.3 — Geometric and Negative Binomial Distributions

Definition: Suppose we consider an infinite sequence of independentBern(p) trials.

Let Z equal the number of trials until the first success is obtained. The eventZ = k corresponds to k − 1 failures, and then a success. Thus,

P (Z = k) = qk−1p, k = 1, 2, . . . ,

and we say that Z has the Geometric distribution with parameter p.

Notation: X ∼ Geom(p).

We’ll get the mean and variance of the Geometric via the mgf. . . .

ISYE 6739 — Goldsman 8/5/20 11 / 108

Geometric and Negative Binomial Distributions

Lesson 4.3 — Geometric and Negative Binomial Distributions

Definition: Suppose we consider an infinite sequence of independentBern(p) trials.

Let Z equal the number of trials until the first success is obtained. The eventZ = k corresponds to k − 1 failures, and then a success. Thus,

P (Z = k) = qk−1p, k = 1, 2, . . . ,

and we say that Z has the Geometric distribution with parameter p.

Notation: X ∼ Geom(p).

We’ll get the mean and variance of the Geometric via the mgf. . . .

ISYE 6739 — Goldsman 8/5/20 11 / 108

Geometric and Negative Binomial Distributions

Lesson 4.3 — Geometric and Negative Binomial Distributions

Definition: Suppose we consider an infinite sequence of independentBern(p) trials.

Let Z equal the number of trials until the first success is obtained. The eventZ = k corresponds to k − 1 failures, and then a success. Thus,

P (Z = k) = qk−1p, k = 1, 2, . . . ,

and we say that Z has the Geometric distribution with parameter p.

Notation: X ∼ Geom(p).

We’ll get the mean and variance of the Geometric via the mgf. . . .

ISYE 6739 — Goldsman 8/5/20 11 / 108

Geometric and Negative Binomial Distributions

Lesson 4.3 — Geometric and Negative Binomial Distributions

Definition: Suppose we consider an infinite sequence of independentBern(p) trials.

Let Z equal the number of trials until the first success is obtained. The eventZ = k corresponds to k − 1 failures, and then a success. Thus,

P (Z = k) = qk−1p, k = 1, 2, . . . ,

and we say that Z has the Geometric distribution with parameter p.

Notation: X ∼ Geom(p).

We’ll get the mean and variance of the Geometric via the mgf. . . .

ISYE 6739 — Goldsman 8/5/20 11 / 108

Geometric and Negative Binomial Distributions

Lesson 4.3 — Geometric and Negative Binomial Distributions

Definition: Suppose we consider an infinite sequence of independentBern(p) trials.

Let Z equal the number of trials until the first success is obtained. The eventZ = k corresponds to k − 1 failures, and then a success. Thus,

P (Z = k) = qk−1p, k = 1, 2, . . . ,

and we say that Z has the Geometric distribution with parameter p.

Notation: X ∼ Geom(p).

We’ll get the mean and variance of the Geometric via the mgf. . . .

ISYE 6739 — Goldsman 8/5/20 11 / 108

Geometric and Negative Binomial Distributions

Lesson 4.3 — Geometric and Negative Binomial Distributions

Definition: Suppose we consider an infinite sequence of independentBern(p) trials.

Let Z equal the number of trials until the first success is obtained. The eventZ = k corresponds to k − 1 failures, and then a success. Thus,

P (Z = k) = qk−1p, k = 1, 2, . . . ,

and we say that Z has the Geometric distribution with parameter p.

Notation: X ∼ Geom(p).

We’ll get the mean and variance of the Geometric via the mgf. . . .

ISYE 6739 — Goldsman 8/5/20 11 / 108

Geometric and Negative Binomial Distributions

Theorem: The mgf of the Geom(p) is

MZ(t) =pet

1− qet, for t < `n(1/q).

Proof:

MZ(t) = E[etZ ]

=

∞∑k=1

etkqk−1p

= pet∞∑k=0

(qet)k

=pet

1− qet, for qet < 1. 2

ISYE 6739 — Goldsman 8/5/20 12 / 108

Geometric and Negative Binomial Distributions

Theorem: The mgf of the Geom(p) is

MZ(t) =pet

1− qet, for t < `n(1/q).

Proof:

MZ(t) = E[etZ ]

=

∞∑k=1

etkqk−1p

= pet∞∑k=0

(qet)k

=pet

1− qet, for qet < 1. 2

ISYE 6739 — Goldsman 8/5/20 12 / 108

Geometric and Negative Binomial Distributions

Theorem: The mgf of the Geom(p) is

MZ(t) =pet

1− qet, for t < `n(1/q).

Proof:

MZ(t) = E[etZ ]

=

∞∑k=1

etkqk−1p

= pet∞∑k=0

(qet)k

=pet

1− qet, for qet < 1. 2

ISYE 6739 — Goldsman 8/5/20 12 / 108

Geometric and Negative Binomial Distributions

Theorem: The mgf of the Geom(p) is

MZ(t) =pet

1− qet, for t < `n(1/q).

Proof:

MZ(t) = E[etZ ]

=

∞∑k=1

etkqk−1p

= pet∞∑k=0

(qet)k

=pet

1− qet, for qet < 1. 2

ISYE 6739 — Goldsman 8/5/20 12 / 108

Geometric and Negative Binomial Distributions

Theorem: The mgf of the Geom(p) is

MZ(t) =pet

1− qet, for t < `n(1/q).

Proof:

MZ(t) = E[etZ ]

=

∞∑k=1

etkqk−1p

= pet∞∑k=0

(qet)k

=pet

1− qet, for qet < 1. 2

ISYE 6739 — Goldsman 8/5/20 12 / 108

Geometric and Negative Binomial Distributions

Corollary: E[Z] = 1/p.

Proof:

E[Z] =d

dtMZ(t)

∣∣∣t=0

=(1− qet)(pet)− (−qet)(pet)

(1− qet)2∣∣∣t=0

=pet

(1− qet)2∣∣∣t=0

=p

(1− q)2=

1

p. 2

Remark: We could also have proven this directly from the definition ofexpected value.

ISYE 6739 — Goldsman 8/5/20 13 / 108

Geometric and Negative Binomial Distributions

Corollary: E[Z] = 1/p.

Proof:

E[Z] =d

dtMZ(t)

∣∣∣t=0

=(1− qet)(pet)− (−qet)(pet)

(1− qet)2∣∣∣t=0

=pet

(1− qet)2∣∣∣t=0

=p

(1− q)2=

1

p. 2

Remark: We could also have proven this directly from the definition ofexpected value.

ISYE 6739 — Goldsman 8/5/20 13 / 108

Geometric and Negative Binomial Distributions

Corollary: E[Z] = 1/p.

Proof:

E[Z] =d

dtMZ(t)

∣∣∣t=0

=(1− qet)(pet)− (−qet)(pet)

(1− qet)2∣∣∣t=0

=pet

(1− qet)2∣∣∣t=0

=p

(1− q)2=

1

p. 2

Remark: We could also have proven this directly from the definition ofexpected value.

ISYE 6739 — Goldsman 8/5/20 13 / 108

Geometric and Negative Binomial Distributions

Corollary: E[Z] = 1/p.

Proof:

E[Z] =d

dtMZ(t)

∣∣∣t=0

=(1− qet)(pet)− (−qet)(pet)

(1− qet)2∣∣∣t=0

=pet

(1− qet)2∣∣∣t=0

=p

(1− q)2=

1

p. 2

Remark: We could also have proven this directly from the definition ofexpected value.

ISYE 6739 — Goldsman 8/5/20 13 / 108

Geometric and Negative Binomial Distributions

Corollary: E[Z] = 1/p.

Proof:

E[Z] =d

dtMZ(t)

∣∣∣t=0

=(1− qet)(pet)− (−qet)(pet)

(1− qet)2∣∣∣t=0

=pet

(1− qet)2∣∣∣t=0

=p

(1− q)2=

1

p. 2

Remark: We could also have proven this directly from the definition ofexpected value.

ISYE 6739 — Goldsman 8/5/20 13 / 108

Geometric and Negative Binomial Distributions

Corollary: E[Z] = 1/p.

Proof:

E[Z] =d

dtMZ(t)

∣∣∣t=0

=(1− qet)(pet)− (−qet)(pet)

(1− qet)2∣∣∣t=0

=pet

(1− qet)2∣∣∣t=0

=p

(1− q)2=

1

p. 2

Remark: We could also have proven this directly from the definition ofexpected value.

ISYE 6739 — Goldsman 8/5/20 13 / 108

Geometric and Negative Binomial Distributions

Similarly, after a lot of algebra,

E[Z2] =d2

dt2MZ(t)

∣∣∣t=0

=2− pp2

,

and thenVar(Z) = E[Z2]− (E[Z])2 = q/p2. 2

Example: Toss a die repeatedly. What’s the probability that we observe a 3for the first time on the 8th toss?

Answer: The number of tosses we need is Z ∼ Geom(1/6).P (Z = 8) = qk−1p = (5/6)7(1/6). 2

How many tosses would we expect to take?

Answer: E[Z] = 1/p = 6 tosses. 2

ISYE 6739 — Goldsman 8/5/20 14 / 108

Geometric and Negative Binomial Distributions

Similarly, after a lot of algebra,

E[Z2] =d2

dt2MZ(t)

∣∣∣t=0

=2− pp2

,

and thenVar(Z) = E[Z2]− (E[Z])2 = q/p2. 2

Example: Toss a die repeatedly. What’s the probability that we observe a 3for the first time on the 8th toss?

Answer: The number of tosses we need is Z ∼ Geom(1/6).P (Z = 8) = qk−1p = (5/6)7(1/6). 2

How many tosses would we expect to take?

Answer: E[Z] = 1/p = 6 tosses. 2

ISYE 6739 — Goldsman 8/5/20 14 / 108

Geometric and Negative Binomial Distributions

Similarly, after a lot of algebra,

E[Z2] =d2

dt2MZ(t)

∣∣∣t=0

=2− pp2

,

and thenVar(Z) = E[Z2]− (E[Z])2 = q/p2. 2

Example: Toss a die repeatedly. What’s the probability that we observe a 3for the first time on the 8th toss?

Answer: The number of tosses we need is Z ∼ Geom(1/6).P (Z = 8) = qk−1p = (5/6)7(1/6). 2

How many tosses would we expect to take?

Answer: E[Z] = 1/p = 6 tosses. 2

ISYE 6739 — Goldsman 8/5/20 14 / 108

Geometric and Negative Binomial Distributions

Similarly, after a lot of algebra,

E[Z2] =d2

dt2MZ(t)

∣∣∣t=0

=2− pp2

,

and thenVar(Z) = E[Z2]− (E[Z])2 = q/p2. 2

Example: Toss a die repeatedly. What’s the probability that we observe a 3for the first time on the 8th toss?

Answer: The number of tosses we need is Z ∼ Geom(1/6).P (Z = 8) = qk−1p = (5/6)7(1/6). 2

How many tosses would we expect to take?

Answer: E[Z] = 1/p = 6 tosses. 2

ISYE 6739 — Goldsman 8/5/20 14 / 108

Geometric and Negative Binomial Distributions

Similarly, after a lot of algebra,

E[Z2] =d2

dt2MZ(t)

∣∣∣t=0

=2− pp2

,

and thenVar(Z) = E[Z2]− (E[Z])2 = q/p2. 2

Example: Toss a die repeatedly. What’s the probability that we observe a 3for the first time on the 8th toss?

Answer: The number of tosses we need is Z ∼ Geom(1/6).P (Z = 8) = qk−1p

= (5/6)7(1/6). 2

How many tosses would we expect to take?

Answer: E[Z] = 1/p = 6 tosses. 2

ISYE 6739 — Goldsman 8/5/20 14 / 108

Geometric and Negative Binomial Distributions

Similarly, after a lot of algebra,

E[Z2] =d2

dt2MZ(t)

∣∣∣t=0

=2− pp2

,

and thenVar(Z) = E[Z2]− (E[Z])2 = q/p2. 2

Example: Toss a die repeatedly. What’s the probability that we observe a 3for the first time on the 8th toss?

Answer: The number of tosses we need is Z ∼ Geom(1/6).P (Z = 8) = qk−1p = (5/6)7(1/6). 2

How many tosses would we expect to take?

Answer: E[Z] = 1/p = 6 tosses. 2

ISYE 6739 — Goldsman 8/5/20 14 / 108

Geometric and Negative Binomial Distributions

Similarly, after a lot of algebra,

E[Z2] =d2

dt2MZ(t)

∣∣∣t=0

=2− pp2

,

and thenVar(Z) = E[Z2]− (E[Z])2 = q/p2. 2

Example: Toss a die repeatedly. What’s the probability that we observe a 3for the first time on the 8th toss?

Answer: The number of tosses we need is Z ∼ Geom(1/6).P (Z = 8) = qk−1p = (5/6)7(1/6). 2

How many tosses would we expect to take?

Answer: E[Z] = 1/p = 6 tosses. 2

ISYE 6739 — Goldsman 8/5/20 14 / 108

Geometric and Negative Binomial Distributions

Similarly, after a lot of algebra,

E[Z2] =d2

dt2MZ(t)

∣∣∣t=0

=2− pp2

,

and thenVar(Z) = E[Z2]− (E[Z])2 = q/p2. 2

Example: Toss a die repeatedly. What’s the probability that we observe a 3for the first time on the 8th toss?

Answer: The number of tosses we need is Z ∼ Geom(1/6).P (Z = 8) = qk−1p = (5/6)7(1/6). 2

How many tosses would we expect to take?

Answer: E[Z] = 1/p = 6 tosses. 2

ISYE 6739 — Goldsman 8/5/20 14 / 108

Geometric and Negative Binomial Distributions

Memoryless Property of Geometric

Theorem: Suppose Z ∼ Geom(p). Then for positive integers s, t, we have

P (Z > s+ t|Z > s) = P (Z > t).

Why is it called the Memoryless Property? If an event hasn’t occurred by times, the probability that it will occur after an additional t time units is the sameas the (unconditional) probability that it will occur after time t — it forgot thatit made it past time s!

ISYE 6739 — Goldsman 8/5/20 15 / 108

Geometric and Negative Binomial Distributions

Memoryless Property of Geometric

Theorem: Suppose Z ∼ Geom(p). Then for positive integers s, t, we have

P (Z > s+ t|Z > s) = P (Z > t).

Why is it called the Memoryless Property? If an event hasn’t occurred by times, the probability that it will occur after an additional t time units is the sameas the (unconditional) probability that it will occur after time t — it forgot thatit made it past time s!

ISYE 6739 — Goldsman 8/5/20 15 / 108

Geometric and Negative Binomial Distributions

Memoryless Property of Geometric

Theorem: Suppose Z ∼ Geom(p). Then for positive integers s, t, we have

P (Z > s+ t|Z > s) = P (Z > t).

Why is it called the Memoryless Property?

If an event hasn’t occurred by times, the probability that it will occur after an additional t time units is the sameas the (unconditional) probability that it will occur after time t — it forgot thatit made it past time s!

ISYE 6739 — Goldsman 8/5/20 15 / 108

Geometric and Negative Binomial Distributions

Memoryless Property of Geometric

Theorem: Suppose Z ∼ Geom(p). Then for positive integers s, t, we have

P (Z > s+ t|Z > s) = P (Z > t).

Why is it called the Memoryless Property? If an event hasn’t occurred by times, the probability that it will occur after an additional t time units is the sameas the (unconditional) probability that it will occur after time t — it forgot thatit made it past time s!

ISYE 6739 — Goldsman 8/5/20 15 / 108

Geometric and Negative Binomial Distributions

Proof: First of all, for any t = 0, 1, 2, . . ., the tail probability is

P (Z > t) = P (t Bern(p) failures in a row) = qt. (∗)

Then

P (Z > s+ t|Z > s) =P (Z > s+ t ∩ Z > s)

P (Z > s)

=P (Z > s+ t)

P (Z > s)

=qs+t

qs(by (∗))

= qt

= P (Z > t). 2

ISYE 6739 — Goldsman 8/5/20 16 / 108

Geometric and Negative Binomial Distributions

Proof: First of all, for any t = 0, 1, 2, . . ., the tail probability is

P (Z > t) = P (t Bern(p) failures in a row) = qt. (∗)

Then

P (Z > s+ t|Z > s)

=P (Z > s+ t ∩ Z > s)

P (Z > s)

=P (Z > s+ t)

P (Z > s)

=qs+t

qs(by (∗))

= qt

= P (Z > t). 2

ISYE 6739 — Goldsman 8/5/20 16 / 108

Geometric and Negative Binomial Distributions

Proof: First of all, for any t = 0, 1, 2, . . ., the tail probability is

P (Z > t) = P (t Bern(p) failures in a row) = qt. (∗)

Then

P (Z > s+ t|Z > s) =P (Z > s+ t ∩ Z > s)

P (Z > s)

=P (Z > s+ t)

P (Z > s)

=qs+t

qs(by (∗))

= qt

= P (Z > t). 2

ISYE 6739 — Goldsman 8/5/20 16 / 108

Geometric and Negative Binomial Distributions

Proof: First of all, for any t = 0, 1, 2, . . ., the tail probability is

P (Z > t) = P (t Bern(p) failures in a row) = qt. (∗)

Then

P (Z > s+ t|Z > s) =P (Z > s+ t ∩ Z > s)

P (Z > s)

=P (Z > s+ t)

P (Z > s)

=qs+t

qs(by (∗))

= qt

= P (Z > t). 2

ISYE 6739 — Goldsman 8/5/20 16 / 108

Geometric and Negative Binomial Distributions

Proof: First of all, for any t = 0, 1, 2, . . ., the tail probability is

P (Z > t) = P (t Bern(p) failures in a row) = qt. (∗)

Then

P (Z > s+ t|Z > s) =P (Z > s+ t ∩ Z > s)

P (Z > s)

=P (Z > s+ t)

P (Z > s)

=qs+t

qs(by (∗))

= qt

= P (Z > t). 2

ISYE 6739 — Goldsman 8/5/20 16 / 108

Geometric and Negative Binomial Distributions

Proof: First of all, for any t = 0, 1, 2, . . ., the tail probability is

P (Z > t) = P (t Bern(p) failures in a row) = qt. (∗)

Then

P (Z > s+ t|Z > s) =P (Z > s+ t ∩ Z > s)

P (Z > s)

=P (Z > s+ t)

P (Z > s)

=qs+t

qs(by (∗))

= qt

= P (Z > t). 2

ISYE 6739 — Goldsman 8/5/20 16 / 108

Geometric and Negative Binomial Distributions

Proof: First of all, for any t = 0, 1, 2, . . ., the tail probability is

P (Z > t) = P (t Bern(p) failures in a row) = qt. (∗)

Then

P (Z > s+ t|Z > s) =P (Z > s+ t ∩ Z > s)

P (Z > s)

=P (Z > s+ t)

P (Z > s)

=qs+t

qs(by (∗))

= qt

= P (Z > t). 2

ISYE 6739 — Goldsman 8/5/20 16 / 108

Geometric and Negative Binomial Distributions

Example: Let’s toss a die until a 5 appears for the first time.

Suppose thatwe’ve already made 4 tosses without success. What’s the probability thatwe’ll need more than 2 additional tosses before we observe a 5?

Let Z be the number of tosses required. By the Memoryless Property (withs = 4 and t = 2) and (∗), we want

P (Z > 6|Z > 4) = P (Z > 2) = (5/6)2. 2

Fun Fact: The Geom(p) is the only discrete distribution with thememoryless property.

Not-as-Fun Fact: Some books define the Geom(p) as the number ofBern(p) failures until you observe a success. # failures = # trials − 1. Youshould be aware of this inconsistency, but don’t worry about it now.

ISYE 6739 — Goldsman 8/5/20 17 / 108

Geometric and Negative Binomial Distributions

Example: Let’s toss a die until a 5 appears for the first time. Suppose thatwe’ve already made 4 tosses without success.

What’s the probability thatwe’ll need more than 2 additional tosses before we observe a 5?

Let Z be the number of tosses required. By the Memoryless Property (withs = 4 and t = 2) and (∗), we want

P (Z > 6|Z > 4) = P (Z > 2) = (5/6)2. 2

Fun Fact: The Geom(p) is the only discrete distribution with thememoryless property.

Not-as-Fun Fact: Some books define the Geom(p) as the number ofBern(p) failures until you observe a success. # failures = # trials − 1. Youshould be aware of this inconsistency, but don’t worry about it now.

ISYE 6739 — Goldsman 8/5/20 17 / 108

Geometric and Negative Binomial Distributions

Example: Let’s toss a die until a 5 appears for the first time. Suppose thatwe’ve already made 4 tosses without success. What’s the probability thatwe’ll need more than 2 additional tosses before we observe a 5?

Let Z be the number of tosses required. By the Memoryless Property (withs = 4 and t = 2) and (∗), we want

P (Z > 6|Z > 4) = P (Z > 2) = (5/6)2. 2

Fun Fact: The Geom(p) is the only discrete distribution with thememoryless property.

Not-as-Fun Fact: Some books define the Geom(p) as the number ofBern(p) failures until you observe a success. # failures = # trials − 1. Youshould be aware of this inconsistency, but don’t worry about it now.

ISYE 6739 — Goldsman 8/5/20 17 / 108

Geometric and Negative Binomial Distributions

Example: Let’s toss a die until a 5 appears for the first time. Suppose thatwe’ve already made 4 tosses without success. What’s the probability thatwe’ll need more than 2 additional tosses before we observe a 5?

Let Z be the number of tosses required.

By the Memoryless Property (withs = 4 and t = 2) and (∗), we want

P (Z > 6|Z > 4) = P (Z > 2) = (5/6)2. 2

Fun Fact: The Geom(p) is the only discrete distribution with thememoryless property.

Not-as-Fun Fact: Some books define the Geom(p) as the number ofBern(p) failures until you observe a success. # failures = # trials − 1. Youshould be aware of this inconsistency, but don’t worry about it now.

ISYE 6739 — Goldsman 8/5/20 17 / 108

Geometric and Negative Binomial Distributions

Example: Let’s toss a die until a 5 appears for the first time. Suppose thatwe’ve already made 4 tosses without success. What’s the probability thatwe’ll need more than 2 additional tosses before we observe a 5?

Let Z be the number of tosses required. By the Memoryless Property (withs = 4 and t = 2) and (∗), we want

P (Z > 6|Z > 4) = P (Z > 2) = (5/6)2. 2

Fun Fact: The Geom(p) is the only discrete distribution with thememoryless property.

Not-as-Fun Fact: Some books define the Geom(p) as the number ofBern(p) failures until you observe a success. # failures = # trials − 1. Youshould be aware of this inconsistency, but don’t worry about it now.

ISYE 6739 — Goldsman 8/5/20 17 / 108

Geometric and Negative Binomial Distributions

Example: Let’s toss a die until a 5 appears for the first time. Suppose thatwe’ve already made 4 tosses without success. What’s the probability thatwe’ll need more than 2 additional tosses before we observe a 5?

Let Z be the number of tosses required. By the Memoryless Property (withs = 4 and t = 2) and (∗), we want

P (Z > 6|Z > 4) = P (Z > 2) = (5/6)2. 2

Fun Fact: The Geom(p) is the only discrete distribution with thememoryless property.

Not-as-Fun Fact: Some books define the Geom(p) as the number ofBern(p) failures until you observe a success. # failures = # trials − 1. Youshould be aware of this inconsistency, but don’t worry about it now.

ISYE 6739 — Goldsman 8/5/20 17 / 108

Geometric and Negative Binomial Distributions

Example: Let’s toss a die until a 5 appears for the first time. Suppose thatwe’ve already made 4 tosses without success. What’s the probability thatwe’ll need more than 2 additional tosses before we observe a 5?

Let Z be the number of tosses required. By the Memoryless Property (withs = 4 and t = 2) and (∗), we want

P (Z > 6|Z > 4) = P (Z > 2) = (5/6)2. 2

Fun Fact: The Geom(p) is the only discrete distribution with thememoryless property.

Not-as-Fun Fact: Some books define the Geom(p) as the number ofBern(p) failures until you observe a success. # failures = # trials − 1. Youshould be aware of this inconsistency, but don’t worry about it now.

ISYE 6739 — Goldsman 8/5/20 17 / 108

Geometric and Negative Binomial Distributions

Example: Let’s toss a die until a 5 appears for the first time. Suppose thatwe’ve already made 4 tosses without success. What’s the probability thatwe’ll need more than 2 additional tosses before we observe a 5?

Let Z be the number of tosses required. By the Memoryless Property (withs = 4 and t = 2) and (∗), we want

P (Z > 6|Z > 4) = P (Z > 2) = (5/6)2. 2

Fun Fact: The Geom(p) is the only discrete distribution with thememoryless property.

Not-as-Fun Fact: Some books define the Geom(p) as the number ofBern(p) failures until you observe a success. # failures = # trials − 1. Youshould be aware of this inconsistency, but don’t worry about it now.

ISYE 6739 — Goldsman 8/5/20 17 / 108

Geometric and Negative Binomial Distributions

Definition: Suppose we consider an infinite sequence of independentBern(p) trials.

Now let W equal the number of trials until the rth success is obtained.W = r, r + 1, . . .. The event W = k corresponds to exactly r − 1 successesby time k − 1, and then the rth success at time k.

We say that W has the Negative Binomial distribution (aka the Pascaldistribution) with parameters r and p.

Example: ‘FFFFSFS’ corresponds to W = 7 trials until the r = 2nd success.

Notation: W ∼ NegBin(r, p).

Remark: As with the Geom(p), the exact definition of the NegBin dependson what book you’re reading.

ISYE 6739 — Goldsman 8/5/20 18 / 108

Geometric and Negative Binomial Distributions

Definition: Suppose we consider an infinite sequence of independentBern(p) trials.

Now let W equal the number of trials until the rth success is obtained.W = r, r + 1, . . ..

The event W = k corresponds to exactly r − 1 successesby time k − 1, and then the rth success at time k.

We say that W has the Negative Binomial distribution (aka the Pascaldistribution) with parameters r and p.

Example: ‘FFFFSFS’ corresponds to W = 7 trials until the r = 2nd success.

Notation: W ∼ NegBin(r, p).

Remark: As with the Geom(p), the exact definition of the NegBin dependson what book you’re reading.

ISYE 6739 — Goldsman 8/5/20 18 / 108

Geometric and Negative Binomial Distributions

Definition: Suppose we consider an infinite sequence of independentBern(p) trials.

Now let W equal the number of trials until the rth success is obtained.W = r, r + 1, . . .. The event W = k corresponds to exactly r − 1 successesby time k − 1, and then the rth success at time k.

We say that W has the Negative Binomial distribution (aka the Pascaldistribution) with parameters r and p.

Example: ‘FFFFSFS’ corresponds to W = 7 trials until the r = 2nd success.

Notation: W ∼ NegBin(r, p).

Remark: As with the Geom(p), the exact definition of the NegBin dependson what book you’re reading.

ISYE 6739 — Goldsman 8/5/20 18 / 108

Geometric and Negative Binomial Distributions

Definition: Suppose we consider an infinite sequence of independentBern(p) trials.

Now let W equal the number of trials until the rth success is obtained.W = r, r + 1, . . .. The event W = k corresponds to exactly r − 1 successesby time k − 1, and then the rth success at time k.

We say that W has the Negative Binomial distribution (aka the Pascaldistribution) with parameters r and p.

Example: ‘FFFFSFS’ corresponds to W = 7 trials until the r = 2nd success.

Notation: W ∼ NegBin(r, p).

Remark: As with the Geom(p), the exact definition of the NegBin dependson what book you’re reading.

ISYE 6739 — Goldsman 8/5/20 18 / 108

Geometric and Negative Binomial Distributions

Definition: Suppose we consider an infinite sequence of independentBern(p) trials.

Now let W equal the number of trials until the rth success is obtained.W = r, r + 1, . . .. The event W = k corresponds to exactly r − 1 successesby time k − 1, and then the rth success at time k.

We say that W has the Negative Binomial distribution (aka the Pascaldistribution) with parameters r and p.

Example: ‘FFFFSFS’ corresponds to W = 7 trials until the r = 2nd success.

Notation: W ∼ NegBin(r, p).

Remark: As with the Geom(p), the exact definition of the NegBin dependson what book you’re reading.

ISYE 6739 — Goldsman 8/5/20 18 / 108

Geometric and Negative Binomial Distributions

Definition: Suppose we consider an infinite sequence of independentBern(p) trials.

Now let W equal the number of trials until the rth success is obtained.W = r, r + 1, . . .. The event W = k corresponds to exactly r − 1 successesby time k − 1, and then the rth success at time k.

We say that W has the Negative Binomial distribution (aka the Pascaldistribution) with parameters r and p.

Example: ‘FFFFSFS’ corresponds to W = 7 trials until the r = 2nd success.

Notation: W ∼ NegBin(r, p).

Remark: As with the Geom(p), the exact definition of the NegBin dependson what book you’re reading.

ISYE 6739 — Goldsman 8/5/20 18 / 108

Geometric and Negative Binomial Distributions

Definition: Suppose we consider an infinite sequence of independentBern(p) trials.

Now let W equal the number of trials until the rth success is obtained.W = r, r + 1, . . .. The event W = k corresponds to exactly r − 1 successesby time k − 1, and then the rth success at time k.

We say that W has the Negative Binomial distribution (aka the Pascaldistribution) with parameters r and p.

Example: ‘FFFFSFS’ corresponds to W = 7 trials until the r = 2nd success.

Notation: W ∼ NegBin(r, p).

Remark: As with the Geom(p), the exact definition of the NegBin dependson what book you’re reading.

ISYE 6739 — Goldsman 8/5/20 18 / 108

Geometric and Negative Binomial Distributions

Theorem: If Z1, . . . , Zriid∼ Geom(p), then W =

∑ri=1 Zi ∼ NegBin(r, p).

In other words, Geometrics add up to a NegBin.

Proof: Won’t do it here, but you can use the mgf technique. 2

Anyhow, it makes sense if you think of Zi as the number of trials after the(i− 1)st success up to and including the ith success.

Since the Zi’s are i.i.d., the above theorem gives:

E[W ] = rE[Zi] = r/p,

Var(W ) = rVar(Zi) = rq/p2,

MW (t) = [MZi(t)]r =

(pet

1− qet

)r.

ISYE 6739 — Goldsman 8/5/20 19 / 108

Geometric and Negative Binomial Distributions

Theorem: If Z1, . . . , Zriid∼ Geom(p), then W =

∑ri=1 Zi ∼ NegBin(r, p).

In other words, Geometrics add up to a NegBin.

Proof: Won’t do it here, but you can use the mgf technique. 2

Anyhow, it makes sense if you think of Zi as the number of trials after the(i− 1)st success up to and including the ith success.

Since the Zi’s are i.i.d., the above theorem gives:

E[W ] = rE[Zi] = r/p,

Var(W ) = rVar(Zi) = rq/p2,

MW (t) = [MZi(t)]r =

(pet

1− qet

)r.

ISYE 6739 — Goldsman 8/5/20 19 / 108

Geometric and Negative Binomial Distributions

Theorem: If Z1, . . . , Zriid∼ Geom(p), then W =

∑ri=1 Zi ∼ NegBin(r, p).

In other words, Geometrics add up to a NegBin.

Proof: Won’t do it here, but you can use the mgf technique. 2

Anyhow, it makes sense if you think of Zi as the number of trials after the(i− 1)st success up to and including the ith success.

Since the Zi’s are i.i.d., the above theorem gives:

E[W ] = rE[Zi] = r/p,

Var(W ) = rVar(Zi) = rq/p2,

MW (t) = [MZi(t)]r =

(pet

1− qet

)r.

ISYE 6739 — Goldsman 8/5/20 19 / 108

Geometric and Negative Binomial Distributions

Theorem: If Z1, . . . , Zriid∼ Geom(p), then W =

∑ri=1 Zi ∼ NegBin(r, p).

In other words, Geometrics add up to a NegBin.

Proof: Won’t do it here, but you can use the mgf technique. 2

Anyhow, it makes sense if you think of Zi as the number of trials after the(i− 1)st success up to and including the ith success.

Since the Zi’s are i.i.d., the above theorem gives:

E[W ] = rE[Zi] = r/p,

Var(W ) = rVar(Zi) = rq/p2,

MW (t) = [MZi(t)]r =

(pet

1− qet

)r.

ISYE 6739 — Goldsman 8/5/20 19 / 108

Geometric and Negative Binomial Distributions

Theorem: If Z1, . . . , Zriid∼ Geom(p), then W =

∑ri=1 Zi ∼ NegBin(r, p).

In other words, Geometrics add up to a NegBin.

Proof: Won’t do it here, but you can use the mgf technique. 2

Anyhow, it makes sense if you think of Zi as the number of trials after the(i− 1)st success up to and including the ith success.

Since the Zi’s are i.i.d., the above theorem gives:

E[W ] = rE[Zi] = r/p,

Var(W ) = rVar(Zi) = rq/p2,

MW (t) = [MZi(t)]r =

(pet

1− qet

)r.

ISYE 6739 — Goldsman 8/5/20 19 / 108

Geometric and Negative Binomial Distributions

Theorem: If Z1, . . . , Zriid∼ Geom(p), then W =

∑ri=1 Zi ∼ NegBin(r, p).

In other words, Geometrics add up to a NegBin.

Proof: Won’t do it here, but you can use the mgf technique. 2

Anyhow, it makes sense if you think of Zi as the number of trials after the(i− 1)st success up to and including the ith success.

Since the Zi’s are i.i.d., the above theorem gives:

E[W ] = rE[Zi] = r/p,

Var(W ) = rVar(Zi) = rq/p2,

MW (t) = [MZi(t)]r =

(pet

1− qet

)r.

ISYE 6739 — Goldsman 8/5/20 19 / 108

Geometric and Negative Binomial Distributions

Theorem: If Z1, . . . , Zriid∼ Geom(p), then W =

∑ri=1 Zi ∼ NegBin(r, p).

In other words, Geometrics add up to a NegBin.

Proof: Won’t do it here, but you can use the mgf technique. 2

Anyhow, it makes sense if you think of Zi as the number of trials after the(i− 1)st success up to and including the ith success.

Since the Zi’s are i.i.d., the above theorem gives:

E[W ] = rE[Zi] = r/p,

Var(W ) = rVar(Zi) = rq/p2,

MW (t) = [MZi(t)]r =

(pet

1− qet

)r.

ISYE 6739 — Goldsman 8/5/20 19 / 108

Geometric and Negative Binomial Distributions

Theorem: If Z1, . . . , Zriid∼ Geom(p), then W =

∑ri=1 Zi ∼ NegBin(r, p).

In other words, Geometrics add up to a NegBin.

Proof: Won’t do it here, but you can use the mgf technique. 2

Anyhow, it makes sense if you think of Zi as the number of trials after the(i− 1)st success up to and including the ith success.

Since the Zi’s are i.i.d., the above theorem gives:

E[W ] = rE[Zi] = r/p,

Var(W ) = rVar(Zi) = rq/p2,

MW (t) = [MZi(t)]r =

(pet

1− qet

)r.

ISYE 6739 — Goldsman 8/5/20 19 / 108

Geometric and Negative Binomial Distributions

Just to be complete, let’s get the pmf of W .

Note that W = k iff you getexactly r − 1 successes by time k − 1, and then the rth success at time k.So. . .

P (W = k) =

[(k − 1

r − 1

)pr−1qk−r

]p =

(k − 1

r − 1

)prqk−r, k = r, r + 1, . . .

Example: Toss a die until a 5 appears for the third time. What’s theprobability that we’ll need exactly 7 tosses?

Let W be the number of tosses required. Clearly, W ∼ NegBin(3, 1/6).

P (W = 7) =

(7− 1

3− 1

)(1/6)3(5/6)7−3.

ISYE 6739 — Goldsman 8/5/20 20 / 108

Geometric and Negative Binomial Distributions

Just to be complete, let’s get the pmf of W . Note that W = k iff you getexactly r − 1 successes by time k − 1, and then the rth success at time k.So. . .

P (W = k) =

[(k − 1

r − 1

)pr−1qk−r

]p =

(k − 1

r − 1

)prqk−r, k = r, r + 1, . . .

Example: Toss a die until a 5 appears for the third time. What’s theprobability that we’ll need exactly 7 tosses?

Let W be the number of tosses required. Clearly, W ∼ NegBin(3, 1/6).

P (W = 7) =

(7− 1

3− 1

)(1/6)3(5/6)7−3.

ISYE 6739 — Goldsman 8/5/20 20 / 108

Geometric and Negative Binomial Distributions

Just to be complete, let’s get the pmf of W . Note that W = k iff you getexactly r − 1 successes by time k − 1, and then the rth success at time k.So. . .

P (W = k) =

[(k − 1

r − 1

)pr−1qk−r

]p

=

(k − 1

r − 1

)prqk−r, k = r, r + 1, . . .

Example: Toss a die until a 5 appears for the third time. What’s theprobability that we’ll need exactly 7 tosses?

Let W be the number of tosses required. Clearly, W ∼ NegBin(3, 1/6).

P (W = 7) =

(7− 1

3− 1

)(1/6)3(5/6)7−3.

ISYE 6739 — Goldsman 8/5/20 20 / 108

Geometric and Negative Binomial Distributions

Just to be complete, let’s get the pmf of W . Note that W = k iff you getexactly r − 1 successes by time k − 1, and then the rth success at time k.So. . .

P (W = k) =

[(k − 1

r − 1

)pr−1qk−r

]p =

(k − 1

r − 1

)prqk−r, k = r, r + 1, . . .

Example: Toss a die until a 5 appears for the third time. What’s theprobability that we’ll need exactly 7 tosses?

Let W be the number of tosses required. Clearly, W ∼ NegBin(3, 1/6).

P (W = 7) =

(7− 1

3− 1

)(1/6)3(5/6)7−3.

ISYE 6739 — Goldsman 8/5/20 20 / 108

Geometric and Negative Binomial Distributions

Just to be complete, let’s get the pmf of W . Note that W = k iff you getexactly r − 1 successes by time k − 1, and then the rth success at time k.So. . .

P (W = k) =

[(k − 1

r − 1

)pr−1qk−r

]p =

(k − 1

r − 1

)prqk−r, k = r, r + 1, . . .

Example: Toss a die until a 5 appears for the third time.

What’s theprobability that we’ll need exactly 7 tosses?

Let W be the number of tosses required. Clearly, W ∼ NegBin(3, 1/6).

P (W = 7) =

(7− 1

3− 1

)(1/6)3(5/6)7−3.

ISYE 6739 — Goldsman 8/5/20 20 / 108

Geometric and Negative Binomial Distributions

Just to be complete, let’s get the pmf of W . Note that W = k iff you getexactly r − 1 successes by time k − 1, and then the rth success at time k.So. . .

P (W = k) =

[(k − 1

r − 1

)pr−1qk−r

]p =

(k − 1

r − 1

)prqk−r, k = r, r + 1, . . .

Example: Toss a die until a 5 appears for the third time. What’s theprobability that we’ll need exactly 7 tosses?

Let W be the number of tosses required. Clearly, W ∼ NegBin(3, 1/6).

P (W = 7) =

(7− 1

3− 1

)(1/6)3(5/6)7−3.

ISYE 6739 — Goldsman 8/5/20 20 / 108

Geometric and Negative Binomial Distributions

Just to be complete, let’s get the pmf of W . Note that W = k iff you getexactly r − 1 successes by time k − 1, and then the rth success at time k.So. . .

P (W = k) =

[(k − 1

r − 1

)pr−1qk−r

]p =

(k − 1

r − 1

)prqk−r, k = r, r + 1, . . .

Example: Toss a die until a 5 appears for the third time. What’s theprobability that we’ll need exactly 7 tosses?

Let W be the number of tosses required. Clearly, W ∼ NegBin(3, 1/6).

P (W = 7) =

(7− 1

3− 1

)(1/6)3(5/6)7−3.

ISYE 6739 — Goldsman 8/5/20 20 / 108

Geometric and Negative Binomial Distributions

Just to be complete, let’s get the pmf of W . Note that W = k iff you getexactly r − 1 successes by time k − 1, and then the rth success at time k.So. . .

P (W = k) =

[(k − 1

r − 1

)pr−1qk−r

]p =

(k − 1

r − 1

)prqk−r, k = r, r + 1, . . .

Example: Toss a die until a 5 appears for the third time. What’s theprobability that we’ll need exactly 7 tosses?

Let W be the number of tosses required. Clearly, W ∼ NegBin(3, 1/6).

P (W = 7) =

(7− 1

3− 1

)(1/6)3(5/6)7−3.

ISYE 6739 — Goldsman 8/5/20 20 / 108

Geometric and Negative Binomial Distributions

How are the Binomial and NegBin Related?

X1, . . . , Xniid∼ Bern(p) ⇒ Y ≡

∑ni=1Xi ∼ Bin(n, p).

Z1, . . . , Zriid∼ Geom(p) ⇒ W ≡

∑ri=1 Zi ∼ NegBin(r, p).

E[Y ] = np, Var(Y ) = npq.

E[W ] = r/p, Var(W ) = rq/p2.

ISYE 6739 — Goldsman 8/5/20 21 / 108

Geometric and Negative Binomial Distributions

How are the Binomial and NegBin Related?

X1, . . . , Xniid∼ Bern(p) ⇒ Y ≡

∑ni=1Xi ∼ Bin(n, p).

Z1, . . . , Zriid∼ Geom(p) ⇒ W ≡

∑ri=1 Zi ∼ NegBin(r, p).

E[Y ] = np, Var(Y ) = npq.

E[W ] = r/p, Var(W ) = rq/p2.

ISYE 6739 — Goldsman 8/5/20 21 / 108

Geometric and Negative Binomial Distributions

How are the Binomial and NegBin Related?

X1, . . . , Xniid∼ Bern(p) ⇒ Y ≡

∑ni=1Xi ∼ Bin(n, p).

Z1, . . . , Zriid∼ Geom(p) ⇒ W ≡

∑ri=1 Zi ∼ NegBin(r, p).

E[Y ] = np, Var(Y ) = npq.

E[W ] = r/p, Var(W ) = rq/p2.

ISYE 6739 — Goldsman 8/5/20 21 / 108

Geometric and Negative Binomial Distributions

How are the Binomial and NegBin Related?

X1, . . . , Xniid∼ Bern(p) ⇒ Y ≡

∑ni=1Xi ∼ Bin(n, p).

Z1, . . . , Zriid∼ Geom(p) ⇒ W ≡

∑ri=1 Zi ∼ NegBin(r, p).

E[Y ] = np, Var(Y ) = npq.

E[W ] = r/p, Var(W ) = rq/p2.

ISYE 6739 — Goldsman 8/5/20 21 / 108

Geometric and Negative Binomial Distributions

How are the Binomial and NegBin Related?

X1, . . . , Xniid∼ Bern(p) ⇒ Y ≡

∑ni=1Xi ∼ Bin(n, p).

Z1, . . . , Zriid∼ Geom(p) ⇒ W ≡

∑ri=1 Zi ∼ NegBin(r, p).

E[Y ] = np, Var(Y ) = npq.

E[W ] = r/p, Var(W ) = rq/p2.

ISYE 6739 — Goldsman 8/5/20 21 / 108

Poisson Distribution

1 Bernoulli and Binomial Distributions

2 Hypergeometric Distribution3 Geometric and Negative Binomial Distributions

4 Poisson Distribution5 Uniform, Exponential, and Friends6 Other Continuous Distributions7 Normal Distribution: Basics8 Standard Normal Distribution9 Sample Mean of Normals

10 The Central Limit Theorem + Proof

11 Central Limit Theorem Examples

12 Extensions — Multivariate Normal Distribution13 Extensions — Lognormal Distribution

14 Computer Stuff

ISYE 6739 — Goldsman 8/5/20 22 / 108

Poisson Distribution

Lesson 4.4 — Poisson Distribution

We’ll first talk about Poisson processes.

Let N(t) be a counting process. That is, N(t) is the number ofoccurrences (or arrivals, or events) of some process over the time interval[0, t]. N(t) looks like a step function.

Examples: N(t) could be any of the following.(a) Cars entering a shopping center (by time t).(b) Defects on a wire (of length t).(c) Raisins in cookie dough (of volume t).

Let λ > 0 be the average number of occurrences per unit time (or length orvolume).

In the above examples, we might have:(a) λ = 10/min. (b) λ = 0.5/ft. (c) λ = 4/in3.

ISYE 6739 — Goldsman 8/5/20 23 / 108

Poisson Distribution

Lesson 4.4 — Poisson Distribution

We’ll first talk about Poisson processes.

Let N(t) be a counting process. That is, N(t) is the number ofoccurrences (or arrivals, or events) of some process over the time interval[0, t]. N(t) looks like a step function.

Examples: N(t) could be any of the following.(a) Cars entering a shopping center (by time t).(b) Defects on a wire (of length t).(c) Raisins in cookie dough (of volume t).

Let λ > 0 be the average number of occurrences per unit time (or length orvolume).

In the above examples, we might have:(a) λ = 10/min. (b) λ = 0.5/ft. (c) λ = 4/in3.

ISYE 6739 — Goldsman 8/5/20 23 / 108

Poisson Distribution

Lesson 4.4 — Poisson Distribution

We’ll first talk about Poisson processes.

Let N(t) be a counting process. That is, N(t) is the number ofoccurrences (or arrivals, or events) of some process over the time interval[0, t]. N(t) looks like a step function.

Examples: N(t) could be any of the following.(a) Cars entering a shopping center (by time t).(b) Defects on a wire (of length t).(c) Raisins in cookie dough (of volume t).

Let λ > 0 be the average number of occurrences per unit time (or length orvolume).

In the above examples, we might have:(a) λ = 10/min. (b) λ = 0.5/ft. (c) λ = 4/in3.

ISYE 6739 — Goldsman 8/5/20 23 / 108

Poisson Distribution

Lesson 4.4 — Poisson Distribution

We’ll first talk about Poisson processes.

Let N(t) be a counting process. That is, N(t) is the number ofoccurrences (or arrivals, or events) of some process over the time interval[0, t]. N(t) looks like a step function.

Examples: N(t) could be any of the following.

(a) Cars entering a shopping center (by time t).(b) Defects on a wire (of length t).(c) Raisins in cookie dough (of volume t).

Let λ > 0 be the average number of occurrences per unit time (or length orvolume).

In the above examples, we might have:(a) λ = 10/min. (b) λ = 0.5/ft. (c) λ = 4/in3.

ISYE 6739 — Goldsman 8/5/20 23 / 108

Poisson Distribution

Lesson 4.4 — Poisson Distribution

We’ll first talk about Poisson processes.

Let N(t) be a counting process. That is, N(t) is the number ofoccurrences (or arrivals, or events) of some process over the time interval[0, t]. N(t) looks like a step function.

Examples: N(t) could be any of the following.(a) Cars entering a shopping center (by time t).

(b) Defects on a wire (of length t).(c) Raisins in cookie dough (of volume t).

Let λ > 0 be the average number of occurrences per unit time (or length orvolume).

In the above examples, we might have:(a) λ = 10/min. (b) λ = 0.5/ft. (c) λ = 4/in3.

ISYE 6739 — Goldsman 8/5/20 23 / 108

Poisson Distribution

Lesson 4.4 — Poisson Distribution

We’ll first talk about Poisson processes.

Let N(t) be a counting process. That is, N(t) is the number ofoccurrences (or arrivals, or events) of some process over the time interval[0, t]. N(t) looks like a step function.

Examples: N(t) could be any of the following.(a) Cars entering a shopping center (by time t).(b) Defects on a wire (of length t).

(c) Raisins in cookie dough (of volume t).

Let λ > 0 be the average number of occurrences per unit time (or length orvolume).

In the above examples, we might have:(a) λ = 10/min. (b) λ = 0.5/ft. (c) λ = 4/in3.

ISYE 6739 — Goldsman 8/5/20 23 / 108

Poisson Distribution

Lesson 4.4 — Poisson Distribution

We’ll first talk about Poisson processes.

Let N(t) be a counting process. That is, N(t) is the number ofoccurrences (or arrivals, or events) of some process over the time interval[0, t]. N(t) looks like a step function.

Examples: N(t) could be any of the following.(a) Cars entering a shopping center (by time t).(b) Defects on a wire (of length t).(c) Raisins in cookie dough (of volume t).

Let λ > 0 be the average number of occurrences per unit time (or length orvolume).

In the above examples, we might have:(a) λ = 10/min. (b) λ = 0.5/ft. (c) λ = 4/in3.

ISYE 6739 — Goldsman 8/5/20 23 / 108

Poisson Distribution

Lesson 4.4 — Poisson Distribution

We’ll first talk about Poisson processes.

Let N(t) be a counting process. That is, N(t) is the number ofoccurrences (or arrivals, or events) of some process over the time interval[0, t]. N(t) looks like a step function.

Examples: N(t) could be any of the following.(a) Cars entering a shopping center (by time t).(b) Defects on a wire (of length t).(c) Raisins in cookie dough (of volume t).

Let λ > 0 be the average number of occurrences per unit time (or length orvolume).

In the above examples, we might have:(a) λ = 10/min. (b) λ = 0.5/ft. (c) λ = 4/in3.

ISYE 6739 — Goldsman 8/5/20 23 / 108

Poisson Distribution

Lesson 4.4 — Poisson Distribution

We’ll first talk about Poisson processes.

Let N(t) be a counting process. That is, N(t) is the number ofoccurrences (or arrivals, or events) of some process over the time interval[0, t]. N(t) looks like a step function.

Examples: N(t) could be any of the following.(a) Cars entering a shopping center (by time t).(b) Defects on a wire (of length t).(c) Raisins in cookie dough (of volume t).

Let λ > 0 be the average number of occurrences per unit time (or length orvolume).

In the above examples, we might have:

(a) λ = 10/min. (b) λ = 0.5/ft. (c) λ = 4/in3.

ISYE 6739 — Goldsman 8/5/20 23 / 108

Poisson Distribution

Lesson 4.4 — Poisson Distribution

We’ll first talk about Poisson processes.

Let N(t) be a counting process. That is, N(t) is the number ofoccurrences (or arrivals, or events) of some process over the time interval[0, t]. N(t) looks like a step function.

Examples: N(t) could be any of the following.(a) Cars entering a shopping center (by time t).(b) Defects on a wire (of length t).(c) Raisins in cookie dough (of volume t).

Let λ > 0 be the average number of occurrences per unit time (or length orvolume).

In the above examples, we might have:(a) λ = 10/min.

(b) λ = 0.5/ft. (c) λ = 4/in3.

ISYE 6739 — Goldsman 8/5/20 23 / 108

Poisson Distribution

Lesson 4.4 — Poisson Distribution

We’ll first talk about Poisson processes.

Let N(t) be a counting process. That is, N(t) is the number ofoccurrences (or arrivals, or events) of some process over the time interval[0, t]. N(t) looks like a step function.

Examples: N(t) could be any of the following.(a) Cars entering a shopping center (by time t).(b) Defects on a wire (of length t).(c) Raisins in cookie dough (of volume t).

Let λ > 0 be the average number of occurrences per unit time (or length orvolume).

In the above examples, we might have:(a) λ = 10/min. (b) λ = 0.5/ft.

(c) λ = 4/in3.

ISYE 6739 — Goldsman 8/5/20 23 / 108

Poisson Distribution

Lesson 4.4 — Poisson Distribution

We’ll first talk about Poisson processes.

Let N(t) be a counting process. That is, N(t) is the number ofoccurrences (or arrivals, or events) of some process over the time interval[0, t]. N(t) looks like a step function.

Examples: N(t) could be any of the following.(a) Cars entering a shopping center (by time t).(b) Defects on a wire (of length t).(c) Raisins in cookie dough (of volume t).

Let λ > 0 be the average number of occurrences per unit time (or length orvolume).

In the above examples, we might have:(a) λ = 10/min. (b) λ = 0.5/ft. (c) λ = 4/in3.

ISYE 6739 — Goldsman 8/5/20 23 / 108

Poisson Distribution

A Poisson process is a specific counting process. . . .

First, some notation: o(h) is a generic function that goes to zero faster than hgoes to zero.

Definition: A Poisson process (PP) is one that satisfies the following:

(i) There is a short enough interval of time h, such that, for all t,

P (N(t+ h)−N(t) = 0) = 1− λh+ o(h)

P (N(t+ h)−N(t) = 1) = λh+ o(h)

P (N(t+ h)−N(t) ≥ 2) = o(h)

(ii) The distribution of the “increment” N(t+ h)−N(t) only depends onthe length h.

(iii) If a < b < c < d, then the two “increments” N(d)−N(c) andN(b)−N(a) are independent RVs.

ISYE 6739 — Goldsman 8/5/20 24 / 108

Poisson Distribution

A Poisson process is a specific counting process. . . .

First, some notation: o(h) is a generic function that goes to zero faster than hgoes to zero.

Definition: A Poisson process (PP) is one that satisfies the following:

(i) There is a short enough interval of time h, such that, for all t,

P (N(t+ h)−N(t) = 0) = 1− λh+ o(h)

P (N(t+ h)−N(t) = 1) = λh+ o(h)

P (N(t+ h)−N(t) ≥ 2) = o(h)

(ii) The distribution of the “increment” N(t+ h)−N(t) only depends onthe length h.

(iii) If a < b < c < d, then the two “increments” N(d)−N(c) andN(b)−N(a) are independent RVs.

ISYE 6739 — Goldsman 8/5/20 24 / 108

Poisson Distribution

A Poisson process is a specific counting process. . . .

First, some notation: o(h) is a generic function that goes to zero faster than hgoes to zero.

Definition: A Poisson process (PP) is one that satisfies the following:

(i) There is a short enough interval of time h, such that, for all t,

P (N(t+ h)−N(t) = 0) = 1− λh+ o(h)

P (N(t+ h)−N(t) = 1) = λh+ o(h)

P (N(t+ h)−N(t) ≥ 2) = o(h)

(ii) The distribution of the “increment” N(t+ h)−N(t) only depends onthe length h.

(iii) If a < b < c < d, then the two “increments” N(d)−N(c) andN(b)−N(a) are independent RVs.

ISYE 6739 — Goldsman 8/5/20 24 / 108

Poisson Distribution

A Poisson process is a specific counting process. . . .

First, some notation: o(h) is a generic function that goes to zero faster than hgoes to zero.

Definition: A Poisson process (PP) is one that satisfies the following:

(i) There is a short enough interval of time h, such that, for all t,

P (N(t+ h)−N(t) = 0) = 1− λh+ o(h)

P (N(t+ h)−N(t) = 1) = λh+ o(h)

P (N(t+ h)−N(t) ≥ 2) = o(h)

(ii) The distribution of the “increment” N(t+ h)−N(t) only depends onthe length h.

(iii) If a < b < c < d, then the two “increments” N(d)−N(c) andN(b)−N(a) are independent RVs.

ISYE 6739 — Goldsman 8/5/20 24 / 108

Poisson Distribution

A Poisson process is a specific counting process. . . .

First, some notation: o(h) is a generic function that goes to zero faster than hgoes to zero.

Definition: A Poisson process (PP) is one that satisfies the following:

(i) There is a short enough interval of time h, such that, for all t,

P (N(t+ h)−N(t) = 0) = 1− λh+ o(h)

P (N(t+ h)−N(t) = 1) = λh+ o(h)

P (N(t+ h)−N(t) ≥ 2) = o(h)

(ii) The distribution of the “increment” N(t+ h)−N(t) only depends onthe length h.

(iii) If a < b < c < d, then the two “increments” N(d)−N(c) andN(b)−N(a) are independent RVs.

ISYE 6739 — Goldsman 8/5/20 24 / 108

Poisson Distribution

A Poisson process is a specific counting process. . . .

First, some notation: o(h) is a generic function that goes to zero faster than hgoes to zero.

Definition: A Poisson process (PP) is one that satisfies the following:

(i) There is a short enough interval of time h, such that, for all t,

P (N(t+ h)−N(t) = 0) = 1− λh+ o(h)

P (N(t+ h)−N(t) = 1) = λh+ o(h)

P (N(t+ h)−N(t) ≥ 2) = o(h)

(ii) The distribution of the “increment” N(t+ h)−N(t) only depends onthe length h.

(iii) If a < b < c < d, then the two “increments” N(d)−N(c) andN(b)−N(a) are independent RVs.

ISYE 6739 — Goldsman 8/5/20 24 / 108

Poisson Distribution

A Poisson process is a specific counting process. . . .

First, some notation: o(h) is a generic function that goes to zero faster than hgoes to zero.

Definition: A Poisson process (PP) is one that satisfies the following:

(i) There is a short enough interval of time h, such that, for all t,

P (N(t+ h)−N(t) = 0) = 1− λh+ o(h)

P (N(t+ h)−N(t) = 1) = λh+ o(h)

P (N(t+ h)−N(t) ≥ 2) = o(h)

(ii) The distribution of the “increment” N(t+ h)−N(t) only depends onthe length h.

(iii) If a < b < c < d, then the two “increments” N(d)−N(c) andN(b)−N(a) are independent RVs.

ISYE 6739 — Goldsman 8/5/20 24 / 108

Poisson Distribution

A Poisson process is a specific counting process. . . .

First, some notation: o(h) is a generic function that goes to zero faster than hgoes to zero.

Definition: A Poisson process (PP) is one that satisfies the following:

(i) There is a short enough interval of time h, such that, for all t,

P (N(t+ h)−N(t) = 0) = 1− λh+ o(h)

P (N(t+ h)−N(t) = 1) = λh+ o(h)

P (N(t+ h)−N(t) ≥ 2) = o(h)

(ii) The distribution of the “increment” N(t+ h)−N(t) only depends onthe length h.

(iii) If a < b < c < d, then the two “increments” N(d)−N(c) andN(b)−N(a) are independent RVs.

ISYE 6739 — Goldsman 8/5/20 24 / 108

Poisson Distribution

A Poisson process is a specific counting process. . . .

First, some notation: o(h) is a generic function that goes to zero faster than hgoes to zero.

Definition: A Poisson process (PP) is one that satisfies the following:

(i) There is a short enough interval of time h, such that, for all t,

P (N(t+ h)−N(t) = 0) = 1− λh+ o(h)

P (N(t+ h)−N(t) = 1) = λh+ o(h)

P (N(t+ h)−N(t) ≥ 2) = o(h)

(ii) The distribution of the “increment” N(t+ h)−N(t) only depends onthe length h.

(iii) If a < b < c < d, then the two “increments” N(d)−N(c) andN(b)−N(a) are independent RVs.

ISYE 6739 — Goldsman 8/5/20 24 / 108

Poisson Distribution

English translation of Poisson process assumptions.

(i) Arrivals basically occur one-at-a-time, and then at rate λ/unit time. (Wemust make sure that λ doesn’t change over time.)

(ii) The arrival pattern is stationary — it doesn’t change over time.

(iii) The numbers of arrivals in two disjoint time intervals are independent

Poisson Process Example: Neutrinos hit a detector. Occurrences are rareenough so that they really do happen one-at-a-time. You never get arrivals ofgroups of neutrinos. Further, the rate doesn’t vary over time, and all arrivalsare independent of each other. 2

Anti-Example: Customers arrive at a restaurant. They show up in groups,not one-at-a-time. The rate varies over the day (more at dinnertime). Arrivalsmay not be independent. This ain’t a Poisson process. 2

ISYE 6739 — Goldsman 8/5/20 25 / 108

Poisson Distribution

English translation of Poisson process assumptions.

(i) Arrivals basically occur one-at-a-time, and then at rate λ/unit time. (Wemust make sure that λ doesn’t change over time.)

(ii) The arrival pattern is stationary — it doesn’t change over time.

(iii) The numbers of arrivals in two disjoint time intervals are independent

Poisson Process Example: Neutrinos hit a detector. Occurrences are rareenough so that they really do happen one-at-a-time. You never get arrivals ofgroups of neutrinos. Further, the rate doesn’t vary over time, and all arrivalsare independent of each other. 2

Anti-Example: Customers arrive at a restaurant. They show up in groups,not one-at-a-time. The rate varies over the day (more at dinnertime). Arrivalsmay not be independent. This ain’t a Poisson process. 2

ISYE 6739 — Goldsman 8/5/20 25 / 108

Poisson Distribution

English translation of Poisson process assumptions.

(i) Arrivals basically occur one-at-a-time, and then at rate λ/unit time. (Wemust make sure that λ doesn’t change over time.)

(ii) The arrival pattern is stationary — it doesn’t change over time.

(iii) The numbers of arrivals in two disjoint time intervals are independent

Poisson Process Example: Neutrinos hit a detector. Occurrences are rareenough so that they really do happen one-at-a-time. You never get arrivals ofgroups of neutrinos. Further, the rate doesn’t vary over time, and all arrivalsare independent of each other. 2

Anti-Example: Customers arrive at a restaurant. They show up in groups,not one-at-a-time. The rate varies over the day (more at dinnertime). Arrivalsmay not be independent. This ain’t a Poisson process. 2

ISYE 6739 — Goldsman 8/5/20 25 / 108

Poisson Distribution

English translation of Poisson process assumptions.

(i) Arrivals basically occur one-at-a-time, and then at rate λ/unit time. (Wemust make sure that λ doesn’t change over time.)

(ii) The arrival pattern is stationary — it doesn’t change over time.

(iii) The numbers of arrivals in two disjoint time intervals are independent

Poisson Process Example: Neutrinos hit a detector. Occurrences are rareenough so that they really do happen one-at-a-time. You never get arrivals ofgroups of neutrinos. Further, the rate doesn’t vary over time, and all arrivalsare independent of each other. 2

Anti-Example: Customers arrive at a restaurant. They show up in groups,not one-at-a-time. The rate varies over the day (more at dinnertime). Arrivalsmay not be independent. This ain’t a Poisson process. 2

ISYE 6739 — Goldsman 8/5/20 25 / 108

Poisson Distribution

English translation of Poisson process assumptions.

(i) Arrivals basically occur one-at-a-time, and then at rate λ/unit time. (Wemust make sure that λ doesn’t change over time.)

(ii) The arrival pattern is stationary — it doesn’t change over time.

(iii) The numbers of arrivals in two disjoint time intervals are independent

Poisson Process Example: Neutrinos hit a detector. Occurrences are rareenough so that they really do happen one-at-a-time. You never get arrivals ofgroups of neutrinos. Further, the rate doesn’t vary over time, and all arrivalsare independent of each other. 2

Anti-Example: Customers arrive at a restaurant. They show up in groups,not one-at-a-time. The rate varies over the day (more at dinnertime). Arrivalsmay not be independent. This ain’t a Poisson process. 2

ISYE 6739 — Goldsman 8/5/20 25 / 108

Poisson Distribution

English translation of Poisson process assumptions.

(i) Arrivals basically occur one-at-a-time, and then at rate λ/unit time. (Wemust make sure that λ doesn’t change over time.)

(ii) The arrival pattern is stationary — it doesn’t change over time.

(iii) The numbers of arrivals in two disjoint time intervals are independent

Poisson Process Example: Neutrinos hit a detector. Occurrences are rareenough so that they really do happen one-at-a-time. You never get arrivals ofgroups of neutrinos. Further, the rate doesn’t vary over time, and all arrivalsare independent of each other. 2

Anti-Example: Customers arrive at a restaurant. They show up in groups,not one-at-a-time. The rate varies over the day (more at dinnertime). Arrivalsmay not be independent. This ain’t a Poisson process. 2

ISYE 6739 — Goldsman 8/5/20 25 / 108

Poisson Distribution

Definition: Let X be the number of occurrences in a Poisson(λ) process in aunit interval of time.

Then X has the Poisson distribution with parameterλ.

Notation: X ∼ Pois(λ).

Theorem/Definition: X ∼ Pois(λ)⇒ P (X = k) = e−λλk/k!,k = 0, 1, 2, . . ..

Proof: The proof follows from the PP assumptions and involves some simpledifferential equations.

To begin with, let’s define Px(t) ≡ P (N(t) = x), i.e., the probability ofexactly x arrivals by time t.

Note that the probability that there haven’t been any arrivals by time t+ h canbe written in terms of the probability that there haven’t been any arrivals bytime t. . . .

ISYE 6739 — Goldsman 8/5/20 26 / 108

Poisson Distribution

Definition: Let X be the number of occurrences in a Poisson(λ) process in aunit interval of time. Then X has the Poisson distribution with parameterλ.

Notation: X ∼ Pois(λ).

Theorem/Definition: X ∼ Pois(λ)⇒ P (X = k) = e−λλk/k!,k = 0, 1, 2, . . ..

Proof: The proof follows from the PP assumptions and involves some simpledifferential equations.

To begin with, let’s define Px(t) ≡ P (N(t) = x), i.e., the probability ofexactly x arrivals by time t.

Note that the probability that there haven’t been any arrivals by time t+ h canbe written in terms of the probability that there haven’t been any arrivals bytime t. . . .

ISYE 6739 — Goldsman 8/5/20 26 / 108

Poisson Distribution

Definition: Let X be the number of occurrences in a Poisson(λ) process in aunit interval of time. Then X has the Poisson distribution with parameterλ.

Notation: X ∼ Pois(λ).

Theorem/Definition: X ∼ Pois(λ)⇒ P (X = k) = e−λλk/k!,k = 0, 1, 2, . . ..

Proof: The proof follows from the PP assumptions and involves some simpledifferential equations.

To begin with, let’s define Px(t) ≡ P (N(t) = x), i.e., the probability ofexactly x arrivals by time t.

Note that the probability that there haven’t been any arrivals by time t+ h canbe written in terms of the probability that there haven’t been any arrivals bytime t. . . .

ISYE 6739 — Goldsman 8/5/20 26 / 108

Poisson Distribution

Definition: Let X be the number of occurrences in a Poisson(λ) process in aunit interval of time. Then X has the Poisson distribution with parameterλ.

Notation: X ∼ Pois(λ).

Theorem/Definition: X ∼ Pois(λ)⇒ P (X = k) = e−λλk/k!,k = 0, 1, 2, . . ..

Proof: The proof follows from the PP assumptions and involves some simpledifferential equations.

To begin with, let’s define Px(t) ≡ P (N(t) = x), i.e., the probability ofexactly x arrivals by time t.

Note that the probability that there haven’t been any arrivals by time t+ h canbe written in terms of the probability that there haven’t been any arrivals bytime t. . . .

ISYE 6739 — Goldsman 8/5/20 26 / 108

Poisson Distribution

Definition: Let X be the number of occurrences in a Poisson(λ) process in aunit interval of time. Then X has the Poisson distribution with parameterλ.

Notation: X ∼ Pois(λ).

Theorem/Definition: X ∼ Pois(λ)⇒ P (X = k) = e−λλk/k!,k = 0, 1, 2, . . ..

Proof: The proof follows from the PP assumptions and involves some simpledifferential equations.

To begin with, let’s define Px(t) ≡ P (N(t) = x), i.e., the probability ofexactly x arrivals by time t.

Note that the probability that there haven’t been any arrivals by time t+ h canbe written in terms of the probability that there haven’t been any arrivals bytime t. . . .

ISYE 6739 — Goldsman 8/5/20 26 / 108

Poisson Distribution

Definition: Let X be the number of occurrences in a Poisson(λ) process in aunit interval of time. Then X has the Poisson distribution with parameterλ.

Notation: X ∼ Pois(λ).

Theorem/Definition: X ∼ Pois(λ)⇒ P (X = k) = e−λλk/k!,k = 0, 1, 2, . . ..

Proof: The proof follows from the PP assumptions and involves some simpledifferential equations.

To begin with, let’s define Px(t) ≡ P (N(t) = x), i.e., the probability ofexactly x arrivals by time t.

Note that the probability that there haven’t been any arrivals by time t+ h canbe written in terms of the probability that there haven’t been any arrivals bytime t. . . .

ISYE 6739 — Goldsman 8/5/20 26 / 108

Poisson Distribution

Definition: Let X be the number of occurrences in a Poisson(λ) process in aunit interval of time. Then X has the Poisson distribution with parameterλ.

Notation: X ∼ Pois(λ).

Theorem/Definition: X ∼ Pois(λ)⇒ P (X = k) = e−λλk/k!,k = 0, 1, 2, . . ..

Proof: The proof follows from the PP assumptions and involves some simpledifferential equations.

To begin with, let’s define Px(t) ≡ P (N(t) = x), i.e., the probability ofexactly x arrivals by time t.

Note that the probability that there haven’t been any arrivals by time t+ h canbe written in terms of the probability that there haven’t been any arrivals bytime t. . . .

ISYE 6739 — Goldsman 8/5/20 26 / 108

Poisson Distribution

P0(t+ h)

= P (N(t+ h) = 0)

= P (no arrivals by time t and then no arrivals by time t+ h)

= P({N(t) = 0} ∩ {N(t+ h)−N(t) = 0}

)= P (N(t) = 0)P (N(t+ h)−N(t) = 0) (by indep increments (iii)).= P (N(t) = 0)(1− λh) (by (i) and a little bit of (ii))

= P0(t)(1− λh).

Thus,P0(t+ h)− P0(t)

h

.= −λP0(t).

Taking the limit as h→ 0, we have

P ′0(t) = −λP0(t). (1)

ISYE 6739 — Goldsman 8/5/20 27 / 108

Poisson Distribution

P0(t+ h)

= P (N(t+ h) = 0)

= P (no arrivals by time t and then no arrivals by time t+ h)

= P({N(t) = 0} ∩ {N(t+ h)−N(t) = 0}

)= P (N(t) = 0)P (N(t+ h)−N(t) = 0) (by indep increments (iii)).= P (N(t) = 0)(1− λh) (by (i) and a little bit of (ii))

= P0(t)(1− λh).

Thus,P0(t+ h)− P0(t)

h

.= −λP0(t).

Taking the limit as h→ 0, we have

P ′0(t) = −λP0(t). (1)

ISYE 6739 — Goldsman 8/5/20 27 / 108

Poisson Distribution

P0(t+ h)

= P (N(t+ h) = 0)

= P (no arrivals by time t and then no arrivals by time t+ h)

= P({N(t) = 0} ∩ {N(t+ h)−N(t) = 0}

)

= P (N(t) = 0)P (N(t+ h)−N(t) = 0) (by indep increments (iii)).= P (N(t) = 0)(1− λh) (by (i) and a little bit of (ii))

= P0(t)(1− λh).

Thus,P0(t+ h)− P0(t)

h

.= −λP0(t).

Taking the limit as h→ 0, we have

P ′0(t) = −λP0(t). (1)

ISYE 6739 — Goldsman 8/5/20 27 / 108

Poisson Distribution

P0(t+ h)

= P (N(t+ h) = 0)

= P (no arrivals by time t and then no arrivals by time t+ h)

= P({N(t) = 0} ∩ {N(t+ h)−N(t) = 0}

)= P (N(t) = 0)P (N(t+ h)−N(t) = 0) (by indep increments (iii))

.= P (N(t) = 0)(1− λh) (by (i) and a little bit of (ii))

= P0(t)(1− λh).

Thus,P0(t+ h)− P0(t)

h

.= −λP0(t).

Taking the limit as h→ 0, we have

P ′0(t) = −λP0(t). (1)

ISYE 6739 — Goldsman 8/5/20 27 / 108

Poisson Distribution

P0(t+ h)

= P (N(t+ h) = 0)

= P (no arrivals by time t and then no arrivals by time t+ h)

= P({N(t) = 0} ∩ {N(t+ h)−N(t) = 0}

)= P (N(t) = 0)P (N(t+ h)−N(t) = 0) (by indep increments (iii)).= P (N(t) = 0)(1− λh) (by (i) and a little bit of (ii))

= P0(t)(1− λh).

Thus,P0(t+ h)− P0(t)

h

.= −λP0(t).

Taking the limit as h→ 0, we have

P ′0(t) = −λP0(t). (1)

ISYE 6739 — Goldsman 8/5/20 27 / 108

Poisson Distribution

P0(t+ h)

= P (N(t+ h) = 0)

= P (no arrivals by time t and then no arrivals by time t+ h)

= P({N(t) = 0} ∩ {N(t+ h)−N(t) = 0}

)= P (N(t) = 0)P (N(t+ h)−N(t) = 0) (by indep increments (iii)).= P (N(t) = 0)(1− λh) (by (i) and a little bit of (ii))

= P0(t)(1− λh).

Thus,P0(t+ h)− P0(t)

h

.= −λP0(t).

Taking the limit as h→ 0, we have

P ′0(t) = −λP0(t). (1)

ISYE 6739 — Goldsman 8/5/20 27 / 108

Poisson Distribution

P0(t+ h)

= P (N(t+ h) = 0)

= P (no arrivals by time t and then no arrivals by time t+ h)

= P({N(t) = 0} ∩ {N(t+ h)−N(t) = 0}

)= P (N(t) = 0)P (N(t+ h)−N(t) = 0) (by indep increments (iii)).= P (N(t) = 0)(1− λh) (by (i) and a little bit of (ii))

= P0(t)(1− λh).

Thus,P0(t+ h)− P0(t)

h

.= −λP0(t).

Taking the limit as h→ 0, we have

P ′0(t) = −λP0(t). (1)

ISYE 6739 — Goldsman 8/5/20 27 / 108

Poisson Distribution

P0(t+ h)

= P (N(t+ h) = 0)

= P (no arrivals by time t and then no arrivals by time t+ h)

= P({N(t) = 0} ∩ {N(t+ h)−N(t) = 0}

)= P (N(t) = 0)P (N(t+ h)−N(t) = 0) (by indep increments (iii)).= P (N(t) = 0)(1− λh) (by (i) and a little bit of (ii))

= P0(t)(1− λh).

Thus,P0(t+ h)− P0(t)

h

.= −λP0(t).

Taking the limit as h→ 0, we have

P ′0(t) = −λP0(t). (1)

ISYE 6739 — Goldsman 8/5/20 27 / 108

Poisson Distribution

P0(t+ h)

= P (N(t+ h) = 0)

= P (no arrivals by time t and then no arrivals by time t+ h)

= P({N(t) = 0} ∩ {N(t+ h)−N(t) = 0}

)= P (N(t) = 0)P (N(t+ h)−N(t) = 0) (by indep increments (iii)).= P (N(t) = 0)(1− λh) (by (i) and a little bit of (ii))

= P0(t)(1− λh).

Thus,P0(t+ h)− P0(t)

h

.= −λP0(t).

Taking the limit as h→ 0, we have

P ′0(t) = −λP0(t). (1)

ISYE 6739 — Goldsman 8/5/20 27 / 108

Poisson Distribution

Similarly, for x > 0, we have

Px(t+ h) = P (N(t+ h) = x)

= P (N(t+ h) = x and no arrivals during [t, t+ h])

+P (N(t+ h) = x and ≥ 1 arrival during [t, t+ h])

(Law of Total Probability).= P

({N(t) = x} ∩ {N(t+ h)−N(t) = 0}

)+P({N(t) = x− 1} ∩ {N(t+ h)−N(t) = 1}

)(by (i), only consider case of one arrival in [t, t+ h])

= P (N(t) = x)P (N(t+ h)−N(t) = 0)

+P (N(t) = x− 1)P (N(t+ h)−N(t) = 1)

(by independent increments (iii)).= Px(t)(1− λh) + Px−1(t)λh.

ISYE 6739 — Goldsman 8/5/20 28 / 108

Poisson Distribution

Similarly, for x > 0, we have

Px(t+ h) = P (N(t+ h) = x)

= P (N(t+ h) = x and no arrivals during [t, t+ h])

+P (N(t+ h) = x and ≥ 1 arrival during [t, t+ h])

(Law of Total Probability).= P

({N(t) = x} ∩ {N(t+ h)−N(t) = 0}

)+P({N(t) = x− 1} ∩ {N(t+ h)−N(t) = 1}

)(by (i), only consider case of one arrival in [t, t+ h])

= P (N(t) = x)P (N(t+ h)−N(t) = 0)

+P (N(t) = x− 1)P (N(t+ h)−N(t) = 1)

(by independent increments (iii)).= Px(t)(1− λh) + Px−1(t)λh.

ISYE 6739 — Goldsman 8/5/20 28 / 108

Poisson Distribution

Similarly, for x > 0, we have

Px(t+ h) = P (N(t+ h) = x)

= P (N(t+ h) = x and no arrivals during [t, t+ h])

+P (N(t+ h) = x and ≥ 1 arrival during [t, t+ h])

(Law of Total Probability)

.= P

({N(t) = x} ∩ {N(t+ h)−N(t) = 0}

)+P({N(t) = x− 1} ∩ {N(t+ h)−N(t) = 1}

)(by (i), only consider case of one arrival in [t, t+ h])

= P (N(t) = x)P (N(t+ h)−N(t) = 0)

+P (N(t) = x− 1)P (N(t+ h)−N(t) = 1)

(by independent increments (iii)).= Px(t)(1− λh) + Px−1(t)λh.

ISYE 6739 — Goldsman 8/5/20 28 / 108

Poisson Distribution

Similarly, for x > 0, we have

Px(t+ h) = P (N(t+ h) = x)

= P (N(t+ h) = x and no arrivals during [t, t+ h])

+P (N(t+ h) = x and ≥ 1 arrival during [t, t+ h])

(Law of Total Probability).= P

({N(t) = x} ∩ {N(t+ h)−N(t) = 0}

)+P({N(t) = x− 1} ∩ {N(t+ h)−N(t) = 1}

)(by (i), only consider case of one arrival in [t, t+ h])

= P (N(t) = x)P (N(t+ h)−N(t) = 0)

+P (N(t) = x− 1)P (N(t+ h)−N(t) = 1)

(by independent increments (iii)).= Px(t)(1− λh) + Px−1(t)λh.

ISYE 6739 — Goldsman 8/5/20 28 / 108

Poisson Distribution

Similarly, for x > 0, we have

Px(t+ h) = P (N(t+ h) = x)

= P (N(t+ h) = x and no arrivals during [t, t+ h])

+P (N(t+ h) = x and ≥ 1 arrival during [t, t+ h])

(Law of Total Probability).= P

({N(t) = x} ∩ {N(t+ h)−N(t) = 0}

)+P({N(t) = x− 1} ∩ {N(t+ h)−N(t) = 1}

)(by (i), only consider case of one arrival in [t, t+ h])

= P (N(t) = x)P (N(t+ h)−N(t) = 0)

+P (N(t) = x− 1)P (N(t+ h)−N(t) = 1)

(by independent increments (iii))

.= Px(t)(1− λh) + Px−1(t)λh.

ISYE 6739 — Goldsman 8/5/20 28 / 108

Poisson Distribution

Similarly, for x > 0, we have

Px(t+ h) = P (N(t+ h) = x)

= P (N(t+ h) = x and no arrivals during [t, t+ h])

+P (N(t+ h) = x and ≥ 1 arrival during [t, t+ h])

(Law of Total Probability).= P

({N(t) = x} ∩ {N(t+ h)−N(t) = 0}

)+P({N(t) = x− 1} ∩ {N(t+ h)−N(t) = 1}

)(by (i), only consider case of one arrival in [t, t+ h])

= P (N(t) = x)P (N(t+ h)−N(t) = 0)

+P (N(t) = x− 1)P (N(t+ h)−N(t) = 1)

(by independent increments (iii)).= Px(t)(1− λh) + Px−1(t)λh.

ISYE 6739 — Goldsman 8/5/20 28 / 108

Poisson Distribution

Taking the limits as h→ 0, we obtain

P ′x(t) = λ[Px−1(t)− Px(t)

], x = 1, 2, . . . . (2)

The solution of differential equations (1) and (2) is easily shown to be

Px(t) =(λt)xe−λt

x!, x = 0, 1, 2, . . . .

(If you don’t believe me, just plug in and see for yourself!)

Noting that t = 1 for the Pois(λ) finally completes the proof. 2

Remark: λ can be changed simply by changing the units of time.

Examples:X = # calls to a switchboard in 1 minute ∼ Pois(3 / min)Y = # calls to a switchboard in 5 minutes ∼ Pois(15 / 5 min)Z = # calls to a switchboard in 10 sec ∼ Pois(0.5 / 10 sec)

ISYE 6739 — Goldsman 8/5/20 29 / 108

Poisson Distribution

Taking the limits as h→ 0, we obtain

P ′x(t) = λ[Px−1(t)− Px(t)

], x = 1, 2, . . . . (2)

The solution of differential equations (1) and (2) is easily shown to be

Px(t) =(λt)xe−λt

x!, x = 0, 1, 2, . . . .

(If you don’t believe me, just plug in and see for yourself!)

Noting that t = 1 for the Pois(λ) finally completes the proof. 2

Remark: λ can be changed simply by changing the units of time.

Examples:X = # calls to a switchboard in 1 minute ∼ Pois(3 / min)Y = # calls to a switchboard in 5 minutes ∼ Pois(15 / 5 min)Z = # calls to a switchboard in 10 sec ∼ Pois(0.5 / 10 sec)

ISYE 6739 — Goldsman 8/5/20 29 / 108

Poisson Distribution

Taking the limits as h→ 0, we obtain

P ′x(t) = λ[Px−1(t)− Px(t)

], x = 1, 2, . . . . (2)

The solution of differential equations (1) and (2) is easily shown to be

Px(t) =(λt)xe−λt

x!, x = 0, 1, 2, . . . .

(If you don’t believe me, just plug in and see for yourself!)

Noting that t = 1 for the Pois(λ) finally completes the proof. 2

Remark: λ can be changed simply by changing the units of time.

Examples:X = # calls to a switchboard in 1 minute ∼ Pois(3 / min)Y = # calls to a switchboard in 5 minutes ∼ Pois(15 / 5 min)Z = # calls to a switchboard in 10 sec ∼ Pois(0.5 / 10 sec)

ISYE 6739 — Goldsman 8/5/20 29 / 108

Poisson Distribution

Taking the limits as h→ 0, we obtain

P ′x(t) = λ[Px−1(t)− Px(t)

], x = 1, 2, . . . . (2)

The solution of differential equations (1) and (2) is easily shown to be

Px(t) =(λt)xe−λt

x!, x = 0, 1, 2, . . . .

(If you don’t believe me, just plug in and see for yourself!)

Noting that t = 1 for the Pois(λ) finally completes the proof. 2

Remark: λ can be changed simply by changing the units of time.

Examples:X = # calls to a switchboard in 1 minute ∼ Pois(3 / min)Y = # calls to a switchboard in 5 minutes ∼ Pois(15 / 5 min)Z = # calls to a switchboard in 10 sec ∼ Pois(0.5 / 10 sec)

ISYE 6739 — Goldsman 8/5/20 29 / 108

Poisson Distribution

Taking the limits as h→ 0, we obtain

P ′x(t) = λ[Px−1(t)− Px(t)

], x = 1, 2, . . . . (2)

The solution of differential equations (1) and (2) is easily shown to be

Px(t) =(λt)xe−λt

x!, x = 0, 1, 2, . . . .

(If you don’t believe me, just plug in and see for yourself!)

Noting that t = 1 for the Pois(λ) finally completes the proof. 2

Remark: λ can be changed simply by changing the units of time.

Examples:X = # calls to a switchboard in 1 minute ∼ Pois(3 / min)Y = # calls to a switchboard in 5 minutes ∼ Pois(15 / 5 min)Z = # calls to a switchboard in 10 sec ∼ Pois(0.5 / 10 sec)

ISYE 6739 — Goldsman 8/5/20 29 / 108

Poisson Distribution

Taking the limits as h→ 0, we obtain

P ′x(t) = λ[Px−1(t)− Px(t)

], x = 1, 2, . . . . (2)

The solution of differential equations (1) and (2) is easily shown to be

Px(t) =(λt)xe−λt

x!, x = 0, 1, 2, . . . .

(If you don’t believe me, just plug in and see for yourself!)

Noting that t = 1 for the Pois(λ) finally completes the proof. 2

Remark: λ can be changed simply by changing the units of time.

Examples:X = # calls to a switchboard in 1 minute ∼ Pois(3 / min)Y = # calls to a switchboard in 5 minutes ∼ Pois(15 / 5 min)Z = # calls to a switchboard in 10 sec ∼ Pois(0.5 / 10 sec)

ISYE 6739 — Goldsman 8/5/20 29 / 108

Poisson Distribution

Taking the limits as h→ 0, we obtain

P ′x(t) = λ[Px−1(t)− Px(t)

], x = 1, 2, . . . . (2)

The solution of differential equations (1) and (2) is easily shown to be

Px(t) =(λt)xe−λt

x!, x = 0, 1, 2, . . . .

(If you don’t believe me, just plug in and see for yourself!)

Noting that t = 1 for the Pois(λ) finally completes the proof. 2

Remark: λ can be changed simply by changing the units of time.

Examples:X = # calls to a switchboard in 1 minute ∼ Pois(3 / min)

Y = # calls to a switchboard in 5 minutes ∼ Pois(15 / 5 min)Z = # calls to a switchboard in 10 sec ∼ Pois(0.5 / 10 sec)

ISYE 6739 — Goldsman 8/5/20 29 / 108

Poisson Distribution

Taking the limits as h→ 0, we obtain

P ′x(t) = λ[Px−1(t)− Px(t)

], x = 1, 2, . . . . (2)

The solution of differential equations (1) and (2) is easily shown to be

Px(t) =(λt)xe−λt

x!, x = 0, 1, 2, . . . .

(If you don’t believe me, just plug in and see for yourself!)

Noting that t = 1 for the Pois(λ) finally completes the proof. 2

Remark: λ can be changed simply by changing the units of time.

Examples:X = # calls to a switchboard in 1 minute ∼ Pois(3 / min)Y = # calls to a switchboard in 5 minutes ∼ Pois(15 / 5 min)

Z = # calls to a switchboard in 10 sec ∼ Pois(0.5 / 10 sec)

ISYE 6739 — Goldsman 8/5/20 29 / 108

Poisson Distribution

Taking the limits as h→ 0, we obtain

P ′x(t) = λ[Px−1(t)− Px(t)

], x = 1, 2, . . . . (2)

The solution of differential equations (1) and (2) is easily shown to be

Px(t) =(λt)xe−λt

x!, x = 0, 1, 2, . . . .

(If you don’t believe me, just plug in and see for yourself!)

Noting that t = 1 for the Pois(λ) finally completes the proof. 2

Remark: λ can be changed simply by changing the units of time.

Examples:X = # calls to a switchboard in 1 minute ∼ Pois(3 / min)Y = # calls to a switchboard in 5 minutes ∼ Pois(15 / 5 min)Z = # calls to a switchboard in 10 sec ∼ Pois(0.5 / 10 sec)

ISYE 6739 — Goldsman 8/5/20 29 / 108

Poisson Distribution

Theorem: X ∼ Pois(λ)⇒ mgf is MX(t) = eλ(et−1).

Proof:

MX(t) = E[etX ] =∞∑k=0

etk(e−λλk

k!

)= e−λ

∞∑k=0

(λet)k

k!= e−λeλe

t. 2

Theorem: X ∼ Pois(λ)⇒ E[X] = Var(X) = λ.

Proof (using mgf):

E[X] =d

dtMX(t)

∣∣∣t=0

=d

dteλ(e

t−1)∣∣∣t=0

= λetMX(t)∣∣∣t=0

= λ.

ISYE 6739 — Goldsman 8/5/20 30 / 108

Poisson Distribution

Theorem: X ∼ Pois(λ)⇒ mgf is MX(t) = eλ(et−1).

Proof:

MX(t) = E[etX ]

=∞∑k=0

etk(e−λλk

k!

)= e−λ

∞∑k=0

(λet)k

k!= e−λeλe

t. 2

Theorem: X ∼ Pois(λ)⇒ E[X] = Var(X) = λ.

Proof (using mgf):

E[X] =d

dtMX(t)

∣∣∣t=0

=d

dteλ(e

t−1)∣∣∣t=0

= λetMX(t)∣∣∣t=0

= λ.

ISYE 6739 — Goldsman 8/5/20 30 / 108

Poisson Distribution

Theorem: X ∼ Pois(λ)⇒ mgf is MX(t) = eλ(et−1).

Proof:

MX(t) = E[etX ] =

∞∑k=0

etk(e−λλk

k!

)

= e−λ∞∑k=0

(λet)k

k!= e−λeλe

t. 2

Theorem: X ∼ Pois(λ)⇒ E[X] = Var(X) = λ.

Proof (using mgf):

E[X] =d

dtMX(t)

∣∣∣t=0

=d

dteλ(e

t−1)∣∣∣t=0

= λetMX(t)∣∣∣t=0

= λ.

ISYE 6739 — Goldsman 8/5/20 30 / 108

Poisson Distribution

Theorem: X ∼ Pois(λ)⇒ mgf is MX(t) = eλ(et−1).

Proof:

MX(t) = E[etX ] =

∞∑k=0

etk(e−λλk

k!

)= e−λ

∞∑k=0

(λet)k

k!

= e−λeλet. 2

Theorem: X ∼ Pois(λ)⇒ E[X] = Var(X) = λ.

Proof (using mgf):

E[X] =d

dtMX(t)

∣∣∣t=0

=d

dteλ(e

t−1)∣∣∣t=0

= λetMX(t)∣∣∣t=0

= λ.

ISYE 6739 — Goldsman 8/5/20 30 / 108

Poisson Distribution

Theorem: X ∼ Pois(λ)⇒ mgf is MX(t) = eλ(et−1).

Proof:

MX(t) = E[etX ] =

∞∑k=0

etk(e−λλk

k!

)= e−λ

∞∑k=0

(λet)k

k!= e−λeλe

t. 2

Theorem: X ∼ Pois(λ)⇒ E[X] = Var(X) = λ.

Proof (using mgf):

E[X] =d

dtMX(t)

∣∣∣t=0

=d

dteλ(e

t−1)∣∣∣t=0

= λetMX(t)∣∣∣t=0

= λ.

ISYE 6739 — Goldsman 8/5/20 30 / 108

Poisson Distribution

Theorem: X ∼ Pois(λ)⇒ mgf is MX(t) = eλ(et−1).

Proof:

MX(t) = E[etX ] =

∞∑k=0

etk(e−λλk

k!

)= e−λ

∞∑k=0

(λet)k

k!= e−λeλe

t. 2

Theorem: X ∼ Pois(λ)⇒ E[X] = Var(X) = λ.

Proof (using mgf):

E[X] =d

dtMX(t)

∣∣∣t=0

=d

dteλ(e

t−1)∣∣∣t=0

= λetMX(t)∣∣∣t=0

= λ.

ISYE 6739 — Goldsman 8/5/20 30 / 108

Poisson Distribution

Theorem: X ∼ Pois(λ)⇒ mgf is MX(t) = eλ(et−1).

Proof:

MX(t) = E[etX ] =

∞∑k=0

etk(e−λλk

k!

)= e−λ

∞∑k=0

(λet)k

k!= e−λeλe

t. 2

Theorem: X ∼ Pois(λ)⇒ E[X] = Var(X) = λ.

Proof (using mgf):

E[X] =d

dtMX(t)

∣∣∣t=0

=d

dteλ(e

t−1)∣∣∣t=0

= λetMX(t)∣∣∣t=0

= λ.

ISYE 6739 — Goldsman 8/5/20 30 / 108

Poisson Distribution

Theorem: X ∼ Pois(λ)⇒ mgf is MX(t) = eλ(et−1).

Proof:

MX(t) = E[etX ] =

∞∑k=0

etk(e−λλk

k!

)= e−λ

∞∑k=0

(λet)k

k!= e−λeλe

t. 2

Theorem: X ∼ Pois(λ)⇒ E[X] = Var(X) = λ.

Proof (using mgf):

E[X] =d

dtMX(t)

∣∣∣t=0

=d

dteλ(e

t−1)∣∣∣t=0

= λetMX(t)∣∣∣t=0

= λ.

ISYE 6739 — Goldsman 8/5/20 30 / 108

Poisson Distribution

Theorem: X ∼ Pois(λ)⇒ mgf is MX(t) = eλ(et−1).

Proof:

MX(t) = E[etX ] =

∞∑k=0

etk(e−λλk

k!

)= e−λ

∞∑k=0

(λet)k

k!= e−λeλe

t. 2

Theorem: X ∼ Pois(λ)⇒ E[X] = Var(X) = λ.

Proof (using mgf):

E[X] =d

dtMX(t)

∣∣∣t=0

=d

dteλ(e

t−1)∣∣∣t=0

= λetMX(t)∣∣∣t=0

= λ.

ISYE 6739 — Goldsman 8/5/20 30 / 108

Poisson Distribution

Theorem: X ∼ Pois(λ)⇒ mgf is MX(t) = eλ(et−1).

Proof:

MX(t) = E[etX ] =

∞∑k=0

etk(e−λλk

k!

)= e−λ

∞∑k=0

(λet)k

k!= e−λeλe

t. 2

Theorem: X ∼ Pois(λ)⇒ E[X] = Var(X) = λ.

Proof (using mgf):

E[X] =d

dtMX(t)

∣∣∣t=0

=d

dteλ(e

t−1)∣∣∣t=0

= λetMX(t)∣∣∣t=0

= λ.

ISYE 6739 — Goldsman 8/5/20 30 / 108

Poisson Distribution

Similarly,

E[X2] =d2

dt2MX(t)

∣∣∣t=0

=d

dt

( ddtMX(t)

)∣∣∣t=0

= λd

dt

(etMX(t)

)∣∣∣t=0

= λ[etMX(t) + et

d

dtMX(t)

]∣∣∣t=0

= λet[MX(t) + λetMX(t)

]∣∣∣t=0

= λ(1 + λ).

Thus, Var(X) = E[X2]− (E[X])2 = λ(1 + λ)− λ2 = λ. 2

ISYE 6739 — Goldsman 8/5/20 31 / 108

Poisson Distribution

Similarly,

E[X2] =d2

dt2MX(t)

∣∣∣t=0

=d

dt

( ddtMX(t)

)∣∣∣t=0

= λd

dt

(etMX(t)

)∣∣∣t=0

= λ[etMX(t) + et

d

dtMX(t)

]∣∣∣t=0

= λet[MX(t) + λetMX(t)

]∣∣∣t=0

= λ(1 + λ).

Thus, Var(X) = E[X2]− (E[X])2 = λ(1 + λ)− λ2 = λ. 2

ISYE 6739 — Goldsman 8/5/20 31 / 108

Poisson Distribution

Similarly,

E[X2] =d2

dt2MX(t)

∣∣∣t=0

=d

dt

( ddtMX(t)

)∣∣∣t=0

= λd

dt

(etMX(t)

)∣∣∣t=0

= λ[etMX(t) + et

d

dtMX(t)

]∣∣∣t=0

= λet[MX(t) + λetMX(t)

]∣∣∣t=0

= λ(1 + λ).

Thus, Var(X) = E[X2]− (E[X])2 = λ(1 + λ)− λ2 = λ. 2

ISYE 6739 — Goldsman 8/5/20 31 / 108

Poisson Distribution

Similarly,

E[X2] =d2

dt2MX(t)

∣∣∣t=0

=d

dt

( ddtMX(t)

)∣∣∣t=0

= λd

dt

(etMX(t)

)∣∣∣t=0

= λ[etMX(t) + et

d

dtMX(t)

]∣∣∣t=0

= λet[MX(t) + λetMX(t)

]∣∣∣t=0

= λ(1 + λ).

Thus, Var(X) = E[X2]− (E[X])2 = λ(1 + λ)− λ2 = λ. 2

ISYE 6739 — Goldsman 8/5/20 31 / 108

Poisson Distribution

Similarly,

E[X2] =d2

dt2MX(t)

∣∣∣t=0

=d

dt

( ddtMX(t)

)∣∣∣t=0

= λd

dt

(etMX(t)

)∣∣∣t=0

= λ[etMX(t) + et

d

dtMX(t)

]∣∣∣t=0

= λet[MX(t) + λetMX(t)

]∣∣∣t=0

= λ(1 + λ).

Thus, Var(X) = E[X2]− (E[X])2 = λ(1 + λ)− λ2 = λ. 2

ISYE 6739 — Goldsman 8/5/20 31 / 108

Poisson Distribution

Similarly,

E[X2] =d2

dt2MX(t)

∣∣∣t=0

=d

dt

( ddtMX(t)

)∣∣∣t=0

= λd

dt

(etMX(t)

)∣∣∣t=0

= λ[etMX(t) + et

d

dtMX(t)

]∣∣∣t=0

= λet[MX(t) + λetMX(t)

]∣∣∣t=0

= λ(1 + λ).

Thus, Var(X) = E[X2]− (E[X])2 = λ(1 + λ)− λ2 = λ. 2

ISYE 6739 — Goldsman 8/5/20 31 / 108

Poisson Distribution

Similarly,

E[X2] =d2

dt2MX(t)

∣∣∣t=0

=d

dt

( ddtMX(t)

)∣∣∣t=0

= λd

dt

(etMX(t)

)∣∣∣t=0

= λ[etMX(t) + et

d

dtMX(t)

]∣∣∣t=0

= λet[MX(t) + λetMX(t)

]∣∣∣t=0

= λ(1 + λ).

Thus, Var(X) = E[X2]− (E[X])2 = λ(1 + λ)− λ2 = λ. 2

ISYE 6739 — Goldsman 8/5/20 31 / 108

Poisson Distribution

Example: Calls to a switchboard arrive as a Poisson process with rate 3calls/min.

Let X = number of calls in 1 minute.

So X ∼ Pois(3), E[X] = Var(X) = 3, P (X ≤ 4) =∑4

k=0 e−33k/k!.

Let Y = number of calls in 40 sec.

So Y ∼ Pois(2), E[Y ] = Var(Y ) = 2, P (Y ≤ 4) =∑4

k=0 e−22k/k!. 2

ISYE 6739 — Goldsman 8/5/20 32 / 108

Poisson Distribution

Example: Calls to a switchboard arrive as a Poisson process with rate 3calls/min.

Let X = number of calls in 1 minute.

So X ∼ Pois(3), E[X] = Var(X) = 3, P (X ≤ 4) =∑4

k=0 e−33k/k!.

Let Y = number of calls in 40 sec.

So Y ∼ Pois(2), E[Y ] = Var(Y ) = 2, P (Y ≤ 4) =∑4

k=0 e−22k/k!. 2

ISYE 6739 — Goldsman 8/5/20 32 / 108

Poisson Distribution

Example: Calls to a switchboard arrive as a Poisson process with rate 3calls/min.

Let X = number of calls in 1 minute.

So X ∼ Pois(3), E[X] = Var(X) = 3, P (X ≤ 4) =∑4

k=0 e−33k/k!.

Let Y = number of calls in 40 sec.

So Y ∼ Pois(2), E[Y ] = Var(Y ) = 2, P (Y ≤ 4) =∑4

k=0 e−22k/k!. 2

ISYE 6739 — Goldsman 8/5/20 32 / 108

Poisson Distribution

Example: Calls to a switchboard arrive as a Poisson process with rate 3calls/min.

Let X = number of calls in 1 minute.

So X ∼ Pois(3), E[X] = Var(X) = 3, P (X ≤ 4) =∑4

k=0 e−33k/k!.

Let Y = number of calls in 40 sec.

So Y ∼ Pois(2), E[Y ] = Var(Y ) = 2, P (Y ≤ 4) =∑4

k=0 e−22k/k!. 2

ISYE 6739 — Goldsman 8/5/20 32 / 108

Poisson Distribution

Example: Calls to a switchboard arrive as a Poisson process with rate 3calls/min.

Let X = number of calls in 1 minute.

So X ∼ Pois(3), E[X] = Var(X) = 3, P (X ≤ 4) =∑4

k=0 e−33k/k!.

Let Y = number of calls in 40 sec.

So Y ∼ Pois(2), E[Y ] = Var(Y ) = 2, P (Y ≤ 4) =∑4

k=0 e−22k/k!. 2

ISYE 6739 — Goldsman 8/5/20 32 / 108

Poisson Distribution

Theorem (Additive Property of Poissons): Suppose X1, . . . , Xn areindependent with Xi ∼ Pois(λi), i = 1, . . . , n. Then

Y ≡n∑i=1

Xi ∼ Pois( n∑i=1

λi

).

Proof: Since the Xi’s are independent, we have

MY (t) =

n∏i=1

MXi(t) =

n∏i=1

eλi(et−1) = e(

∑ni=1 λi)(e

t−1),

which is the mgf of the Pois(∑n

i=1 λi)

distribution. 2

ISYE 6739 — Goldsman 8/5/20 33 / 108

Poisson Distribution

Theorem (Additive Property of Poissons): Suppose X1, . . . , Xn areindependent with Xi ∼ Pois(λi), i = 1, . . . , n. Then

Y ≡n∑i=1

Xi ∼ Pois( n∑i=1

λi

).

Proof: Since the Xi’s are independent, we have

MY (t) =

n∏i=1

MXi(t) =

n∏i=1

eλi(et−1) = e(

∑ni=1 λi)(e

t−1),

which is the mgf of the Pois(∑n

i=1 λi)

distribution. 2

ISYE 6739 — Goldsman 8/5/20 33 / 108

Poisson Distribution

Theorem (Additive Property of Poissons): Suppose X1, . . . , Xn areindependent with Xi ∼ Pois(λi), i = 1, . . . , n. Then

Y ≡n∑i=1

Xi ∼ Pois( n∑i=1

λi

).

Proof: Since the Xi’s are independent, we have

MY (t) =

n∏i=1

MXi(t) =

n∏i=1

eλi(et−1) = e(

∑ni=1 λi)(e

t−1),

which is the mgf of the Pois(∑n

i=1 λi)

distribution. 2

ISYE 6739 — Goldsman 8/5/20 33 / 108

Poisson Distribution

Theorem (Additive Property of Poissons): Suppose X1, . . . , Xn areindependent with Xi ∼ Pois(λi), i = 1, . . . , n. Then

Y ≡n∑i=1

Xi ∼ Pois( n∑i=1

λi

).

Proof: Since the Xi’s are independent, we have

MY (t) =

n∏i=1

MXi(t)

=

n∏i=1

eλi(et−1) = e(

∑ni=1 λi)(e

t−1),

which is the mgf of the Pois(∑n

i=1 λi)

distribution. 2

ISYE 6739 — Goldsman 8/5/20 33 / 108

Poisson Distribution

Theorem (Additive Property of Poissons): Suppose X1, . . . , Xn areindependent with Xi ∼ Pois(λi), i = 1, . . . , n. Then

Y ≡n∑i=1

Xi ∼ Pois( n∑i=1

λi

).

Proof: Since the Xi’s are independent, we have

MY (t) =

n∏i=1

MXi(t) =

n∏i=1

eλi(et−1)

= e(∑n

i=1 λi)(et−1),

which is the mgf of the Pois(∑n

i=1 λi)

distribution. 2

ISYE 6739 — Goldsman 8/5/20 33 / 108

Poisson Distribution

Theorem (Additive Property of Poissons): Suppose X1, . . . , Xn areindependent with Xi ∼ Pois(λi), i = 1, . . . , n. Then

Y ≡n∑i=1

Xi ∼ Pois( n∑i=1

λi

).

Proof: Since the Xi’s are independent, we have

MY (t) =

n∏i=1

MXi(t) =

n∏i=1

eλi(et−1) = e(

∑ni=1 λi)(e

t−1),

which is the mgf of the Pois(∑n

i=1 λi)

distribution. 2

ISYE 6739 — Goldsman 8/5/20 33 / 108

Poisson Distribution

Theorem (Additive Property of Poissons): Suppose X1, . . . , Xn areindependent with Xi ∼ Pois(λi), i = 1, . . . , n. Then

Y ≡n∑i=1

Xi ∼ Pois( n∑i=1

λi

).

Proof: Since the Xi’s are independent, we have

MY (t) =

n∏i=1

MXi(t) =

n∏i=1

eλi(et−1) = e(

∑ni=1 λi)(e

t−1),

which is the mgf of the Pois(∑n

i=1 λi)

distribution. 2

ISYE 6739 — Goldsman 8/5/20 33 / 108

Poisson Distribution

Example: Cars driven by males [females] arrive at a parking lot according toa Poisson process with a rate of λ1 = 3 / hr [λ2 = 5 / hr]. All arrivals areindependent.

What’s the probability of exactly 2 arrivals in the next 30 minutes?

The total number of arrivals is Pois(λ1 + λ2 = 8/hr), and so the total in thenext 30 minutes is X ∼ Pois(4). So P (X = 2) = e−442/2!. 2

ISYE 6739 — Goldsman 8/5/20 34 / 108

Poisson Distribution

Example: Cars driven by males [females] arrive at a parking lot according toa Poisson process with a rate of λ1 = 3 / hr [λ2 = 5 / hr]. All arrivals areindependent.

What’s the probability of exactly 2 arrivals in the next 30 minutes?

The total number of arrivals is Pois(λ1 + λ2 = 8/hr), and so the total in thenext 30 minutes is X ∼ Pois(4). So P (X = 2) = e−442/2!. 2

ISYE 6739 — Goldsman 8/5/20 34 / 108

Poisson Distribution

Example: Cars driven by males [females] arrive at a parking lot according toa Poisson process with a rate of λ1 = 3 / hr [λ2 = 5 / hr]. All arrivals areindependent.

What’s the probability of exactly 2 arrivals in the next 30 minutes?

The total number of arrivals is Pois(λ1 + λ2 = 8/hr), and so the total in thenext 30 minutes is X ∼ Pois(4).

So P (X = 2) = e−442/2!. 2

ISYE 6739 — Goldsman 8/5/20 34 / 108

Poisson Distribution

Example: Cars driven by males [females] arrive at a parking lot according toa Poisson process with a rate of λ1 = 3 / hr [λ2 = 5 / hr]. All arrivals areindependent.

What’s the probability of exactly 2 arrivals in the next 30 minutes?

The total number of arrivals is Pois(λ1 + λ2 = 8/hr), and so the total in thenext 30 minutes is X ∼ Pois(4). So P (X = 2) = e−442/2!. 2

ISYE 6739 — Goldsman 8/5/20 34 / 108

Uniform, Exponential, and Friends

1 Bernoulli and Binomial Distributions

2 Hypergeometric Distribution3 Geometric and Negative Binomial Distributions

4 Poisson Distribution5 Uniform, Exponential, and Friends6 Other Continuous Distributions7 Normal Distribution: Basics8 Standard Normal Distribution9 Sample Mean of Normals

10 The Central Limit Theorem + Proof

11 Central Limit Theorem Examples

12 Extensions — Multivariate Normal Distribution13 Extensions — Lognormal Distribution

14 Computer Stuff

ISYE 6739 — Goldsman 8/5/20 35 / 108

Uniform, Exponential, and Friends

Lesson 4.5 — Uniform, Exponential, and Friends

Definition: The RV X has the Uniform distribution if it has pdf and cdf

f(x) =

{1b−a , a ≤ x ≤ b0, otherwise

and F (x) =

0, x < axb−a , a ≤ x ≤ b1, x ≥ b.

Notation: X ∼ Unif(a, b).

Previous work showed that

E[X] =a+ b

2and Var(X) =

(a− b)2

12.

We can also derive the mgf,

MX(t) = E[etX ] =

∫ b

aetx

1

b− adx =

etb − eta

t(b− a).

ISYE 6739 — Goldsman 8/5/20 36 / 108

Uniform, Exponential, and Friends

Lesson 4.5 — Uniform, Exponential, and Friends

Definition: The RV X has the Uniform distribution if it has pdf and cdf

f(x) =

{1b−a , a ≤ x ≤ b0, otherwise

and F (x) =

0, x < axb−a , a ≤ x ≤ b1, x ≥ b.

Notation: X ∼ Unif(a, b).

Previous work showed that

E[X] =a+ b

2and Var(X) =

(a− b)2

12.

We can also derive the mgf,

MX(t) = E[etX ] =

∫ b

aetx

1

b− adx =

etb − eta

t(b− a).

ISYE 6739 — Goldsman 8/5/20 36 / 108

Uniform, Exponential, and Friends

Lesson 4.5 — Uniform, Exponential, and Friends

Definition: The RV X has the Uniform distribution if it has pdf and cdf

f(x) =

{1b−a , a ≤ x ≤ b0, otherwise

and F (x) =

0, x < axb−a , a ≤ x ≤ b1, x ≥ b.

Notation: X ∼ Unif(a, b).

Previous work showed that

E[X] =a+ b

2and Var(X) =

(a− b)2

12.

We can also derive the mgf,

MX(t) = E[etX ] =

∫ b

aetx

1

b− adx =

etb − eta

t(b− a).

ISYE 6739 — Goldsman 8/5/20 36 / 108

Uniform, Exponential, and Friends

Lesson 4.5 — Uniform, Exponential, and Friends

Definition: The RV X has the Uniform distribution if it has pdf and cdf

f(x) =

{1b−a , a ≤ x ≤ b0, otherwise

and F (x) =

0, x < axb−a , a ≤ x ≤ b1, x ≥ b.

Notation: X ∼ Unif(a, b).

Previous work showed that

E[X] =a+ b

2and Var(X) =

(a− b)2

12.

We can also derive the mgf,

MX(t) = E[etX ] =

∫ b

aetx

1

b− adx =

etb − eta

t(b− a).

ISYE 6739 — Goldsman 8/5/20 36 / 108

Uniform, Exponential, and Friends

Lesson 4.5 — Uniform, Exponential, and Friends

Definition: The RV X has the Uniform distribution if it has pdf and cdf

f(x) =

{1b−a , a ≤ x ≤ b0, otherwise

and F (x) =

0, x < axb−a , a ≤ x ≤ b1, x ≥ b.

Notation: X ∼ Unif(a, b).

Previous work showed that

E[X] =a+ b

2and Var(X) =

(a− b)2

12.

We can also derive the mgf,

MX(t) = E[etX ] =

∫ b

aetx

1

b− adx =

etb − eta

t(b− a).

ISYE 6739 — Goldsman 8/5/20 36 / 108

Uniform, Exponential, and Friends

Lesson 4.5 — Uniform, Exponential, and Friends

Definition: The RV X has the Uniform distribution if it has pdf and cdf

f(x) =

{1b−a , a ≤ x ≤ b0, otherwise

and F (x) =

0, x < axb−a , a ≤ x ≤ b1, x ≥ b.

Notation: X ∼ Unif(a, b).

Previous work showed that

E[X] =a+ b

2and Var(X) =

(a− b)2

12.

We can also derive the mgf,

MX(t) = E[etX ] =

∫ b

aetx

1

b− adx =

etb − eta

t(b− a).

ISYE 6739 — Goldsman 8/5/20 36 / 108

Uniform, Exponential, and Friends

Definition: The Exponential(λ) distribution has pdf

f(x) =

{λe−λx, x > 0

0, otherwise.

Previous work showed that the cdf F (x) = 1− e−λx,

E[X] = 1/λ, and Var(X) = 1/λ2.

We also derived the mgf,

MX(t) = E[etX ] =

∫ ∞0

etxf(x) dx =λ

λ− t, t < λ.

ISYE 6739 — Goldsman 8/5/20 37 / 108

Uniform, Exponential, and Friends

Definition: The Exponential(λ) distribution has pdf

f(x) =

{λe−λx, x > 0

0, otherwise.

Previous work showed that the cdf F (x) = 1− e−λx,

E[X] = 1/λ, and Var(X) = 1/λ2.

We also derived the mgf,

MX(t) = E[etX ] =

∫ ∞0

etxf(x) dx =λ

λ− t, t < λ.

ISYE 6739 — Goldsman 8/5/20 37 / 108

Uniform, Exponential, and Friends

Definition: The Exponential(λ) distribution has pdf

f(x) =

{λe−λx, x > 0

0, otherwise.

Previous work showed that the cdf F (x) = 1− e−λx,

E[X] = 1/λ, and Var(X) = 1/λ2.

We also derived the mgf,

MX(t) = E[etX ] =

∫ ∞0

etxf(x) dx =λ

λ− t, t < λ.

ISYE 6739 — Goldsman 8/5/20 37 / 108

Uniform, Exponential, and Friends

Definition: The Exponential(λ) distribution has pdf

f(x) =

{λe−λx, x > 0

0, otherwise.

Previous work showed that the cdf F (x) = 1− e−λx,

E[X] = 1/λ, and Var(X) = 1/λ2.

We also derived the mgf,

MX(t) = E[etX ] =

∫ ∞0

etxf(x) dx =λ

λ− t, t < λ.

ISYE 6739 — Goldsman 8/5/20 37 / 108

Uniform, Exponential, and Friends

Definition: The Exponential(λ) distribution has pdf

f(x) =

{λe−λx, x > 0

0, otherwise.

Previous work showed that the cdf F (x) = 1− e−λx,

E[X] = 1/λ, and Var(X) = 1/λ2.

We also derived the mgf,

MX(t) = E[etX ] =

∫ ∞0

etxf(x) dx =λ

λ− t, t < λ.

ISYE 6739 — Goldsman 8/5/20 37 / 108

Uniform, Exponential, and Friends

Definition: The Exponential(λ) distribution has pdf

f(x) =

{λe−λx, x > 0

0, otherwise.

Previous work showed that the cdf F (x) = 1− e−λx,

E[X] = 1/λ, and Var(X) = 1/λ2.

We also derived the mgf,

MX(t) = E[etX ] =

∫ ∞0

etxf(x) dx =λ

λ− t, t < λ.

ISYE 6739 — Goldsman 8/5/20 37 / 108

Uniform, Exponential, and Friends

Memoryless Property of Exponential

Theorem: Suppose that X ∼ Exp(λ). Then for positive s, t, we have

P (X > s+ t|X > s) = P (X > t).

Similar to the discrete Geometric distribution, the probability that X willsurvive an additional t time units is the (unconditional) probability that it willsurvive at least t — it forgot that it made it past time s! It’s always “like new”!

Proof:

P (X > s+ t|X > s) =P (X > s+ t ∩X > s)

P (X > s)=

P (X > s+ t)

P (X > s)

=e−λ(s+t)

e−λs= e−λt = P (X > t). 2

ISYE 6739 — Goldsman 8/5/20 38 / 108

Uniform, Exponential, and Friends

Memoryless Property of Exponential

Theorem: Suppose that X ∼ Exp(λ). Then for positive s, t, we have

P (X > s+ t|X > s) = P (X > t).

Similar to the discrete Geometric distribution, the probability that X willsurvive an additional t time units is the (unconditional) probability that it willsurvive at least t — it forgot that it made it past time s! It’s always “like new”!

Proof:

P (X > s+ t|X > s) =P (X > s+ t ∩X > s)

P (X > s)=

P (X > s+ t)

P (X > s)

=e−λ(s+t)

e−λs= e−λt = P (X > t). 2

ISYE 6739 — Goldsman 8/5/20 38 / 108

Uniform, Exponential, and Friends

Memoryless Property of Exponential

Theorem: Suppose that X ∼ Exp(λ). Then for positive s, t, we have

P (X > s+ t|X > s) = P (X > t).

Similar to the discrete Geometric distribution, the probability that X willsurvive an additional t time units is the (unconditional) probability that it willsurvive at least t — it forgot that it made it past time s! It’s always “like new”!

Proof:

P (X > s+ t|X > s) =P (X > s+ t ∩X > s)

P (X > s)=

P (X > s+ t)

P (X > s)

=e−λ(s+t)

e−λs= e−λt = P (X > t). 2

ISYE 6739 — Goldsman 8/5/20 38 / 108

Uniform, Exponential, and Friends

Memoryless Property of Exponential

Theorem: Suppose that X ∼ Exp(λ). Then for positive s, t, we have

P (X > s+ t|X > s) = P (X > t).

Similar to the discrete Geometric distribution, the probability that X willsurvive an additional t time units is the (unconditional) probability that it willsurvive at least t

— it forgot that it made it past time s! It’s always “like new”!

Proof:

P (X > s+ t|X > s) =P (X > s+ t ∩X > s)

P (X > s)=

P (X > s+ t)

P (X > s)

=e−λ(s+t)

e−λs= e−λt = P (X > t). 2

ISYE 6739 — Goldsman 8/5/20 38 / 108

Uniform, Exponential, and Friends

Memoryless Property of Exponential

Theorem: Suppose that X ∼ Exp(λ). Then for positive s, t, we have

P (X > s+ t|X > s) = P (X > t).

Similar to the discrete Geometric distribution, the probability that X willsurvive an additional t time units is the (unconditional) probability that it willsurvive at least t — it forgot that it made it past time s! It’s always “like new”!

Proof:

P (X > s+ t|X > s) =P (X > s+ t ∩X > s)

P (X > s)=

P (X > s+ t)

P (X > s)

=e−λ(s+t)

e−λs= e−λt = P (X > t). 2

ISYE 6739 — Goldsman 8/5/20 38 / 108

Uniform, Exponential, and Friends

Memoryless Property of Exponential

Theorem: Suppose that X ∼ Exp(λ). Then for positive s, t, we have

P (X > s+ t|X > s) = P (X > t).

Similar to the discrete Geometric distribution, the probability that X willsurvive an additional t time units is the (unconditional) probability that it willsurvive at least t — it forgot that it made it past time s! It’s always “like new”!

Proof:

P (X > s+ t|X > s) =P (X > s+ t ∩X > s)

P (X > s)

=P (X > s+ t)

P (X > s)

=e−λ(s+t)

e−λs= e−λt = P (X > t). 2

ISYE 6739 — Goldsman 8/5/20 38 / 108

Uniform, Exponential, and Friends

Memoryless Property of Exponential

Theorem: Suppose that X ∼ Exp(λ). Then for positive s, t, we have

P (X > s+ t|X > s) = P (X > t).

Similar to the discrete Geometric distribution, the probability that X willsurvive an additional t time units is the (unconditional) probability that it willsurvive at least t — it forgot that it made it past time s! It’s always “like new”!

Proof:

P (X > s+ t|X > s) =P (X > s+ t ∩X > s)

P (X > s)=

P (X > s+ t)

P (X > s)

=e−λ(s+t)

e−λs= e−λt = P (X > t). 2

ISYE 6739 — Goldsman 8/5/20 38 / 108

Uniform, Exponential, and Friends

Memoryless Property of Exponential

Theorem: Suppose that X ∼ Exp(λ). Then for positive s, t, we have

P (X > s+ t|X > s) = P (X > t).

Similar to the discrete Geometric distribution, the probability that X willsurvive an additional t time units is the (unconditional) probability that it willsurvive at least t — it forgot that it made it past time s! It’s always “like new”!

Proof:

P (X > s+ t|X > s) =P (X > s+ t ∩X > s)

P (X > s)=

P (X > s+ t)

P (X > s)

=e−λ(s+t)

e−λs

= e−λt = P (X > t). 2

ISYE 6739 — Goldsman 8/5/20 38 / 108

Uniform, Exponential, and Friends

Memoryless Property of Exponential

Theorem: Suppose that X ∼ Exp(λ). Then for positive s, t, we have

P (X > s+ t|X > s) = P (X > t).

Similar to the discrete Geometric distribution, the probability that X willsurvive an additional t time units is the (unconditional) probability that it willsurvive at least t — it forgot that it made it past time s! It’s always “like new”!

Proof:

P (X > s+ t|X > s) =P (X > s+ t ∩X > s)

P (X > s)=

P (X > s+ t)

P (X > s)

=e−λ(s+t)

e−λs= e−λt = P (X > t). 2

ISYE 6739 — Goldsman 8/5/20 38 / 108

Uniform, Exponential, and Friends

Example: Suppose that the life of a lightbulb is exponential with a mean of10 months.

If the light survives 20 months, what’s the probability that it’llsurvive another 10?

P (X > 30|X > 20) = P (X > 10) = e−λx = e−(1/10)(10) = e−1. 2

Example: If the time to the next bus is exponentially distributed with a meanof 10 minutes, and you’ve already been waiting 20 minutes, you can expect towait 10 more. 2

Remark: The exponential is the only continuous distribution with theMemoryless Property.

Remark: Look at E[X] and Var(X) for the Geometric distribution and seehow they’re similar to those for the exponential. (Not a coincidence.)

ISYE 6739 — Goldsman 8/5/20 39 / 108

Uniform, Exponential, and Friends

Example: Suppose that the life of a lightbulb is exponential with a mean of10 months. If the light survives 20 months, what’s the probability that it’llsurvive another 10?

P (X > 30|X > 20) = P (X > 10) = e−λx = e−(1/10)(10) = e−1. 2

Example: If the time to the next bus is exponentially distributed with a meanof 10 minutes, and you’ve already been waiting 20 minutes, you can expect towait 10 more. 2

Remark: The exponential is the only continuous distribution with theMemoryless Property.

Remark: Look at E[X] and Var(X) for the Geometric distribution and seehow they’re similar to those for the exponential. (Not a coincidence.)

ISYE 6739 — Goldsman 8/5/20 39 / 108

Uniform, Exponential, and Friends

Example: Suppose that the life of a lightbulb is exponential with a mean of10 months. If the light survives 20 months, what’s the probability that it’llsurvive another 10?

P (X > 30|X > 20) = P (X > 10)

= e−λx = e−(1/10)(10) = e−1. 2

Example: If the time to the next bus is exponentially distributed with a meanof 10 minutes, and you’ve already been waiting 20 minutes, you can expect towait 10 more. 2

Remark: The exponential is the only continuous distribution with theMemoryless Property.

Remark: Look at E[X] and Var(X) for the Geometric distribution and seehow they’re similar to those for the exponential. (Not a coincidence.)

ISYE 6739 — Goldsman 8/5/20 39 / 108

Uniform, Exponential, and Friends

Example: Suppose that the life of a lightbulb is exponential with a mean of10 months. If the light survives 20 months, what’s the probability that it’llsurvive another 10?

P (X > 30|X > 20) = P (X > 10) = e−λx

= e−(1/10)(10) = e−1. 2

Example: If the time to the next bus is exponentially distributed with a meanof 10 minutes, and you’ve already been waiting 20 minutes, you can expect towait 10 more. 2

Remark: The exponential is the only continuous distribution with theMemoryless Property.

Remark: Look at E[X] and Var(X) for the Geometric distribution and seehow they’re similar to those for the exponential. (Not a coincidence.)

ISYE 6739 — Goldsman 8/5/20 39 / 108

Uniform, Exponential, and Friends

Example: Suppose that the life of a lightbulb is exponential with a mean of10 months. If the light survives 20 months, what’s the probability that it’llsurvive another 10?

P (X > 30|X > 20) = P (X > 10) = e−λx = e−(1/10)(10)

= e−1. 2

Example: If the time to the next bus is exponentially distributed with a meanof 10 minutes, and you’ve already been waiting 20 minutes, you can expect towait 10 more. 2

Remark: The exponential is the only continuous distribution with theMemoryless Property.

Remark: Look at E[X] and Var(X) for the Geometric distribution and seehow they’re similar to those for the exponential. (Not a coincidence.)

ISYE 6739 — Goldsman 8/5/20 39 / 108

Uniform, Exponential, and Friends

Example: Suppose that the life of a lightbulb is exponential with a mean of10 months. If the light survives 20 months, what’s the probability that it’llsurvive another 10?

P (X > 30|X > 20) = P (X > 10) = e−λx = e−(1/10)(10) = e−1. 2

Example: If the time to the next bus is exponentially distributed with a meanof 10 minutes, and you’ve already been waiting 20 minutes, you can expect towait 10 more. 2

Remark: The exponential is the only continuous distribution with theMemoryless Property.

Remark: Look at E[X] and Var(X) for the Geometric distribution and seehow they’re similar to those for the exponential. (Not a coincidence.)

ISYE 6739 — Goldsman 8/5/20 39 / 108

Uniform, Exponential, and Friends

Example: Suppose that the life of a lightbulb is exponential with a mean of10 months. If the light survives 20 months, what’s the probability that it’llsurvive another 10?

P (X > 30|X > 20) = P (X > 10) = e−λx = e−(1/10)(10) = e−1. 2

Example: If the time to the next bus is exponentially distributed with a meanof 10 minutes, and you’ve already been waiting 20 minutes, you can expect towait 10 more. 2

Remark: The exponential is the only continuous distribution with theMemoryless Property.

Remark: Look at E[X] and Var(X) for the Geometric distribution and seehow they’re similar to those for the exponential. (Not a coincidence.)

ISYE 6739 — Goldsman 8/5/20 39 / 108

Uniform, Exponential, and Friends

Example: Suppose that the life of a lightbulb is exponential with a mean of10 months. If the light survives 20 months, what’s the probability that it’llsurvive another 10?

P (X > 30|X > 20) = P (X > 10) = e−λx = e−(1/10)(10) = e−1. 2

Example: If the time to the next bus is exponentially distributed with a meanof 10 minutes, and you’ve already been waiting 20 minutes, you can expect towait 10 more. 2

Remark: The exponential is the only continuous distribution with theMemoryless Property.

Remark: Look at E[X] and Var(X) for the Geometric distribution and seehow they’re similar to those for the exponential. (Not a coincidence.)

ISYE 6739 — Goldsman 8/5/20 39 / 108

Uniform, Exponential, and Friends

Example: Suppose that the life of a lightbulb is exponential with a mean of10 months. If the light survives 20 months, what’s the probability that it’llsurvive another 10?

P (X > 30|X > 20) = P (X > 10) = e−λx = e−(1/10)(10) = e−1. 2

Example: If the time to the next bus is exponentially distributed with a meanof 10 minutes, and you’ve already been waiting 20 minutes, you can expect towait 10 more. 2

Remark: The exponential is the only continuous distribution with theMemoryless Property.

Remark: Look at E[X] and Var(X) for the Geometric distribution and seehow they’re similar to those for the exponential. (Not a coincidence.)

ISYE 6739 — Goldsman 8/5/20 39 / 108

Uniform, Exponential, and Friends

Definition: If X is a cts RV with pdf f(x) and cdf F (x), then its failurerate function is

S(t) ≡ f(t)

P (X > t)=

f(t)

1− F (t),

which can loosely be regarded as X’s instantaneous rate of death, given that ithas so far survived to time t.

Example: If X ∼ Exp(λ), then S(t) = λe−λt/e−λt = λ. So if X is theexponential lifetime of a lightbulb, then its instantaneous burn-out rate isalways λ — always good as new! This is clearly a result of the MemorylessProperty. 2

ISYE 6739 — Goldsman 8/5/20 40 / 108

Uniform, Exponential, and Friends

Definition: If X is a cts RV with pdf f(x) and cdf F (x), then its failurerate function is

S(t) ≡ f(t)

P (X > t)=

f(t)

1− F (t),

which can loosely be regarded as X’s instantaneous rate of death, given that ithas so far survived to time t.

Example: If X ∼ Exp(λ), then S(t) = λe−λt/e−λt = λ. So if X is theexponential lifetime of a lightbulb, then its instantaneous burn-out rate isalways λ — always good as new! This is clearly a result of the MemorylessProperty. 2

ISYE 6739 — Goldsman 8/5/20 40 / 108

Uniform, Exponential, and Friends

Definition: If X is a cts RV with pdf f(x) and cdf F (x), then its failurerate function is

S(t) ≡ f(t)

P (X > t)=

f(t)

1− F (t),

which can loosely be regarded as X’s instantaneous rate of death, given that ithas so far survived to time t.

Example: If X ∼ Exp(λ), then S(t) = λe−λt/e−λt = λ. So if X is theexponential lifetime of a lightbulb, then its instantaneous burn-out rate isalways λ — always good as new! This is clearly a result of the MemorylessProperty. 2

ISYE 6739 — Goldsman 8/5/20 40 / 108

Uniform, Exponential, and Friends

Definition: If X is a cts RV with pdf f(x) and cdf F (x), then its failurerate function is

S(t) ≡ f(t)

P (X > t)=

f(t)

1− F (t),

which can loosely be regarded as X’s instantaneous rate of death, given that ithas so far survived to time t.

Example: If X ∼ Exp(λ), then S(t) = λe−λt/e−λt = λ. So if X is theexponential lifetime of a lightbulb, then its instantaneous burn-out rate isalways λ — always good as new! This is clearly a result of the MemorylessProperty. 2

ISYE 6739 — Goldsman 8/5/20 40 / 108

Uniform, Exponential, and Friends

The Exponential is also related to the Poisson!

Theorem: Let X be the amount of time until the first arrival in a Poissonprocess with rate λ. Then X ∼ Exp(λ).

Proof:

F (x) = P (X ≤ x) = 1− P (no arrivals in [0, x])

= 1− e−λx(λx)0

0!(since # arrivals in [0, x] is Pois(λx))

= 1− e−λx. 2

Theorem: Amazingly, it can be shown (after a lot of work) that theinterarrival times of a PP are all iid Exp(λ)! See for yourself when you take astochastic processes course.

ISYE 6739 — Goldsman 8/5/20 41 / 108

Uniform, Exponential, and Friends

The Exponential is also related to the Poisson!

Theorem: Let X be the amount of time until the first arrival in a Poissonprocess with rate λ. Then X ∼ Exp(λ).

Proof:

F (x) = P (X ≤ x) = 1− P (no arrivals in [0, x])

= 1− e−λx(λx)0

0!(since # arrivals in [0, x] is Pois(λx))

= 1− e−λx. 2

Theorem: Amazingly, it can be shown (after a lot of work) that theinterarrival times of a PP are all iid Exp(λ)! See for yourself when you take astochastic processes course.

ISYE 6739 — Goldsman 8/5/20 41 / 108

Uniform, Exponential, and Friends

The Exponential is also related to the Poisson!

Theorem: Let X be the amount of time until the first arrival in a Poissonprocess with rate λ. Then X ∼ Exp(λ).

Proof:

F (x) = P (X ≤ x) = 1− P (no arrivals in [0, x])

= 1− e−λx(λx)0

0!(since # arrivals in [0, x] is Pois(λx))

= 1− e−λx. 2

Theorem: Amazingly, it can be shown (after a lot of work) that theinterarrival times of a PP are all iid Exp(λ)! See for yourself when you take astochastic processes course.

ISYE 6739 — Goldsman 8/5/20 41 / 108

Uniform, Exponential, and Friends

The Exponential is also related to the Poisson!

Theorem: Let X be the amount of time until the first arrival in a Poissonprocess with rate λ. Then X ∼ Exp(λ).

Proof:

F (x) = P (X ≤ x) = 1− P (no arrivals in [0, x])

= 1− e−λx(λx)0

0!(since # arrivals in [0, x] is Pois(λx))

= 1− e−λx. 2

Theorem: Amazingly, it can be shown (after a lot of work) that theinterarrival times of a PP are all iid Exp(λ)! See for yourself when you take astochastic processes course.

ISYE 6739 — Goldsman 8/5/20 41 / 108

Uniform, Exponential, and Friends

The Exponential is also related to the Poisson!

Theorem: Let X be the amount of time until the first arrival in a Poissonprocess with rate λ. Then X ∼ Exp(λ).

Proof:

F (x) = P (X ≤ x) = 1− P (no arrivals in [0, x])

= 1− e−λx(λx)0

0!(since # arrivals in [0, x] is Pois(λx))

= 1− e−λx. 2

Theorem: Amazingly, it can be shown (after a lot of work) that theinterarrival times of a PP are all iid Exp(λ)! See for yourself when you take astochastic processes course.

ISYE 6739 — Goldsman 8/5/20 41 / 108

Uniform, Exponential, and Friends

The Exponential is also related to the Poisson!

Theorem: Let X be the amount of time until the first arrival in a Poissonprocess with rate λ. Then X ∼ Exp(λ).

Proof:

F (x) = P (X ≤ x) = 1− P (no arrivals in [0, x])

= 1− e−λx(λx)0

0!(since # arrivals in [0, x] is Pois(λx))

= 1− e−λx. 2

Theorem: Amazingly, it can be shown (after a lot of work) that theinterarrival times of a PP are all iid Exp(λ)! See for yourself when you take astochastic processes course.

ISYE 6739 — Goldsman 8/5/20 41 / 108

Uniform, Exponential, and Friends

Example: Suppose that arrivals to a shopping center are from a PP with rateλ = 20/hr.

What’s the probability that the time between the 13th and 14thcustomers will be at least 4 minutes?

Let the time between customers 13 and 14 be X . Since we have a PP, theinterarrivals are iid Exp(λ = 20/hr), so

P (X > 4 min) = P (X > 1/15 hr) = e−λt = e−20/15. 2

ISYE 6739 — Goldsman 8/5/20 42 / 108

Uniform, Exponential, and Friends

Example: Suppose that arrivals to a shopping center are from a PP with rateλ = 20/hr. What’s the probability that the time between the 13th and 14thcustomers will be at least 4 minutes?

Let the time between customers 13 and 14 be X . Since we have a PP, theinterarrivals are iid Exp(λ = 20/hr), so

P (X > 4 min) = P (X > 1/15 hr) = e−λt = e−20/15. 2

ISYE 6739 — Goldsman 8/5/20 42 / 108

Uniform, Exponential, and Friends

Example: Suppose that arrivals to a shopping center are from a PP with rateλ = 20/hr. What’s the probability that the time between the 13th and 14thcustomers will be at least 4 minutes?

Let the time between customers 13 and 14 be X . Since we have a PP, theinterarrivals are iid Exp(λ = 20/hr), so

P (X > 4 min) = P (X > 1/15 hr) = e−λt = e−20/15. 2

ISYE 6739 — Goldsman 8/5/20 42 / 108

Uniform, Exponential, and Friends

Example: Suppose that arrivals to a shopping center are from a PP with rateλ = 20/hr. What’s the probability that the time between the 13th and 14thcustomers will be at least 4 minutes?

Let the time between customers 13 and 14 be X . Since we have a PP, theinterarrivals are iid Exp(λ = 20/hr), so

P (X > 4 min) = P (X > 1/15 hr) = e−λt = e−20/15. 2

ISYE 6739 — Goldsman 8/5/20 42 / 108

Uniform, Exponential, and Friends

Definition/Theorem: Suppose X1, . . . , Xkiid∼ Exp(λ), and let

S =∑k

i=1Xi. Then S has the Erlang distribution with parameters k andλ, denoted S ∼ Erlangk(λ).

The Erlang is simply the sum of iid exponentials.

Special Case: Erlang1(λ) ∼ Exp(λ).

The pdf and cdf of the Erlang are

f(s) =λke−λssk−1

(k − 1)!, s ≥ 0, and F (s) = 1−

k−1∑i=0

e−λs(λs)i

i!.

Notice that the cdf is the sum of a bunch of Poisson probabilities. (Won’t do ithere, but this observation helps in the derivation of the cdf.)

ISYE 6739 — Goldsman 8/5/20 43 / 108

Uniform, Exponential, and Friends

Definition/Theorem: Suppose X1, . . . , Xkiid∼ Exp(λ), and let

S =∑k

i=1Xi. Then S has the Erlang distribution with parameters k andλ, denoted S ∼ Erlangk(λ).

The Erlang is simply the sum of iid exponentials.

Special Case: Erlang1(λ) ∼ Exp(λ).

The pdf and cdf of the Erlang are

f(s) =λke−λssk−1

(k − 1)!, s ≥ 0, and F (s) = 1−

k−1∑i=0

e−λs(λs)i

i!.

Notice that the cdf is the sum of a bunch of Poisson probabilities. (Won’t do ithere, but this observation helps in the derivation of the cdf.)

ISYE 6739 — Goldsman 8/5/20 43 / 108

Uniform, Exponential, and Friends

Definition/Theorem: Suppose X1, . . . , Xkiid∼ Exp(λ), and let

S =∑k

i=1Xi. Then S has the Erlang distribution with parameters k andλ, denoted S ∼ Erlangk(λ).

The Erlang is simply the sum of iid exponentials.

Special Case: Erlang1(λ) ∼ Exp(λ).

The pdf and cdf of the Erlang are

f(s) =λke−λssk−1

(k − 1)!, s ≥ 0, and F (s) = 1−

k−1∑i=0

e−λs(λs)i

i!.

Notice that the cdf is the sum of a bunch of Poisson probabilities. (Won’t do ithere, but this observation helps in the derivation of the cdf.)

ISYE 6739 — Goldsman 8/5/20 43 / 108

Uniform, Exponential, and Friends

Definition/Theorem: Suppose X1, . . . , Xkiid∼ Exp(λ), and let

S =∑k

i=1Xi. Then S has the Erlang distribution with parameters k andλ, denoted S ∼ Erlangk(λ).

The Erlang is simply the sum of iid exponentials.

Special Case: Erlang1(λ) ∼ Exp(λ).

The pdf and cdf of the Erlang are

f(s) =λke−λssk−1

(k − 1)!, s ≥ 0, and F (s) = 1−

k−1∑i=0

e−λs(λs)i

i!.

Notice that the cdf is the sum of a bunch of Poisson probabilities. (Won’t do ithere, but this observation helps in the derivation of the cdf.)

ISYE 6739 — Goldsman 8/5/20 43 / 108

Uniform, Exponential, and Friends

Definition/Theorem: Suppose X1, . . . , Xkiid∼ Exp(λ), and let

S =∑k

i=1Xi. Then S has the Erlang distribution with parameters k andλ, denoted S ∼ Erlangk(λ).

The Erlang is simply the sum of iid exponentials.

Special Case: Erlang1(λ) ∼ Exp(λ).

The pdf and cdf of the Erlang are

f(s) =λke−λssk−1

(k − 1)!, s ≥ 0, and

F (s) = 1−k−1∑i=0

e−λs(λs)i

i!.

Notice that the cdf is the sum of a bunch of Poisson probabilities. (Won’t do ithere, but this observation helps in the derivation of the cdf.)

ISYE 6739 — Goldsman 8/5/20 43 / 108

Uniform, Exponential, and Friends

Definition/Theorem: Suppose X1, . . . , Xkiid∼ Exp(λ), and let

S =∑k

i=1Xi. Then S has the Erlang distribution with parameters k andλ, denoted S ∼ Erlangk(λ).

The Erlang is simply the sum of iid exponentials.

Special Case: Erlang1(λ) ∼ Exp(λ).

The pdf and cdf of the Erlang are

f(s) =λke−λssk−1

(k − 1)!, s ≥ 0, and F (s) = 1−

k−1∑i=0

e−λs(λs)i

i!.

Notice that the cdf is the sum of a bunch of Poisson probabilities. (Won’t do ithere, but this observation helps in the derivation of the cdf.)

ISYE 6739 — Goldsman 8/5/20 43 / 108

Uniform, Exponential, and Friends

Definition/Theorem: Suppose X1, . . . , Xkiid∼ Exp(λ), and let

S =∑k

i=1Xi. Then S has the Erlang distribution with parameters k andλ, denoted S ∼ Erlangk(λ).

The Erlang is simply the sum of iid exponentials.

Special Case: Erlang1(λ) ∼ Exp(λ).

The pdf and cdf of the Erlang are

f(s) =λke−λssk−1

(k − 1)!, s ≥ 0, and F (s) = 1−

k−1∑i=0

e−λs(λs)i

i!.

Notice that the cdf is the sum of a bunch of Poisson probabilities. (Won’t do ithere, but this observation helps in the derivation of the cdf.)

ISYE 6739 — Goldsman 8/5/20 43 / 108

Uniform, Exponential, and Friends

Expected value, variance, and mgf:

E[S] = E[ k∑i=1

Xi

]=

k∑i=1

E[Xi] = k/λ

Var(S) = k/λ2

MS(t) =

λ− t

)k.

Example: Suppose X and Y are iid Exp(2). Find P (X + Y < 1).

P (X + Y < 1) = 1−k−1∑i=0

e−λs(λs)i

i!= 1−

2−1∑i=0

e−(2·1)(2 · 1)i

i!= 0.594. 2

ISYE 6739 — Goldsman 8/5/20 44 / 108

Uniform, Exponential, and Friends

Expected value, variance, and mgf:

E[S] = E[ k∑i=1

Xi

]=

k∑i=1

E[Xi] = k/λ

Var(S) = k/λ2

MS(t) =

λ− t

)k.

Example: Suppose X and Y are iid Exp(2). Find P (X + Y < 1).

P (X + Y < 1) = 1−k−1∑i=0

e−λs(λs)i

i!= 1−

2−1∑i=0

e−(2·1)(2 · 1)i

i!= 0.594. 2

ISYE 6739 — Goldsman 8/5/20 44 / 108

Uniform, Exponential, and Friends

Expected value, variance, and mgf:

E[S] = E[ k∑i=1

Xi

]=

k∑i=1

E[Xi] = k/λ

Var(S) = k/λ2

MS(t) =

λ− t

)k.

Example: Suppose X and Y are iid Exp(2). Find P (X + Y < 1).

P (X + Y < 1) = 1−k−1∑i=0

e−λs(λs)i

i!= 1−

2−1∑i=0

e−(2·1)(2 · 1)i

i!= 0.594. 2

ISYE 6739 — Goldsman 8/5/20 44 / 108

Uniform, Exponential, and Friends

Expected value, variance, and mgf:

E[S] = E[ k∑i=1

Xi

]=

k∑i=1

E[Xi] = k/λ

Var(S) = k/λ2

MS(t) =

λ− t

)k.

Example: Suppose X and Y are iid Exp(2). Find P (X + Y < 1).

P (X + Y < 1) = 1−k−1∑i=0

e−λs(λs)i

i!= 1−

2−1∑i=0

e−(2·1)(2 · 1)i

i!= 0.594. 2

ISYE 6739 — Goldsman 8/5/20 44 / 108

Uniform, Exponential, and Friends

Expected value, variance, and mgf:

E[S] = E[ k∑i=1

Xi

]=

k∑i=1

E[Xi] = k/λ

Var(S) = k/λ2

MS(t) =

λ− t

)k.

Example: Suppose X and Y are iid Exp(2). Find P (X + Y < 1).

P (X + Y < 1) = 1−k−1∑i=0

e−λs(λs)i

i!

= 1−2−1∑i=0

e−(2·1)(2 · 1)i

i!= 0.594. 2

ISYE 6739 — Goldsman 8/5/20 44 / 108

Uniform, Exponential, and Friends

Expected value, variance, and mgf:

E[S] = E[ k∑i=1

Xi

]=

k∑i=1

E[Xi] = k/λ

Var(S) = k/λ2

MS(t) =

λ− t

)k.

Example: Suppose X and Y are iid Exp(2). Find P (X + Y < 1).

P (X + Y < 1) = 1−k−1∑i=0

e−λs(λs)i

i!= 1−

2−1∑i=0

e−(2·1)(2 · 1)i

i!= 0.594. 2

ISYE 6739 — Goldsman 8/5/20 44 / 108

Uniform, Exponential, and Friends

Definition: X has the Gamma distribution with parameters α > 0 andλ > 0 if it has pdf

f(x) =λαe−λxxα−1

Γ(α), x ≥ 0,

whereΓ(α) ≡

∫ ∞0

tα−1e−t dt

is the gamma function.

Remark: The Gamma distribution generalizes the Erlang distribution (whereα has to be a positive integer). It has the same expected value and variance asthe Erlang, with α in place of k.

Remark: If α is a positive integer, then Γ(α) = (α− 1)!. Party trick:Γ(1/2) =

√π.

ISYE 6739 — Goldsman 8/5/20 45 / 108

Uniform, Exponential, and Friends

Definition: X has the Gamma distribution with parameters α > 0 andλ > 0 if it has pdf

f(x) =λαe−λxxα−1

Γ(α), x ≥ 0,

whereΓ(α) ≡

∫ ∞0

tα−1e−t dt

is the gamma function.

Remark: The Gamma distribution generalizes the Erlang distribution (whereα has to be a positive integer). It has the same expected value and variance asthe Erlang, with α in place of k.

Remark: If α is a positive integer, then Γ(α) = (α− 1)!. Party trick:Γ(1/2) =

√π.

ISYE 6739 — Goldsman 8/5/20 45 / 108

Uniform, Exponential, and Friends

Definition: X has the Gamma distribution with parameters α > 0 andλ > 0 if it has pdf

f(x) =λαe−λxxα−1

Γ(α), x ≥ 0,

whereΓ(α) ≡

∫ ∞0

tα−1e−t dt

is the gamma function.

Remark: The Gamma distribution generalizes the Erlang distribution (whereα has to be a positive integer). It has the same expected value and variance asthe Erlang, with α in place of k.

Remark: If α is a positive integer, then Γ(α) = (α− 1)!. Party trick:Γ(1/2) =

√π.

ISYE 6739 — Goldsman 8/5/20 45 / 108

Uniform, Exponential, and Friends

Definition: X has the Gamma distribution with parameters α > 0 andλ > 0 if it has pdf

f(x) =λαe−λxxα−1

Γ(α), x ≥ 0,

whereΓ(α) ≡

∫ ∞0

tα−1e−t dt

is the gamma function.

Remark: The Gamma distribution generalizes the Erlang distribution (whereα has to be a positive integer). It has the same expected value and variance asthe Erlang, with α in place of k.

Remark: If α is a positive integer, then Γ(α) = (α− 1)!. Party trick:Γ(1/2) =

√π.

ISYE 6739 — Goldsman 8/5/20 45 / 108

Uniform, Exponential, and Friends

Definition: X has the Gamma distribution with parameters α > 0 andλ > 0 if it has pdf

f(x) =λαe−λxxα−1

Γ(α), x ≥ 0,

whereΓ(α) ≡

∫ ∞0

tα−1e−t dt

is the gamma function.

Remark: The Gamma distribution generalizes the Erlang distribution (whereα has to be a positive integer). It has the same expected value and variance asthe Erlang, with α in place of k.

Remark: If α is a positive integer, then Γ(α) = (α− 1)!. Party trick:Γ(1/2) =

√π.

ISYE 6739 — Goldsman 8/5/20 45 / 108

Other Continuous Distributions

1 Bernoulli and Binomial Distributions

2 Hypergeometric Distribution3 Geometric and Negative Binomial Distributions

4 Poisson Distribution5 Uniform, Exponential, and Friends6 Other Continuous Distributions7 Normal Distribution: Basics8 Standard Normal Distribution9 Sample Mean of Normals

10 The Central Limit Theorem + Proof

11 Central Limit Theorem Examples

12 Extensions — Multivariate Normal Distribution13 Extensions — Lognormal Distribution

14 Computer Stuff

ISYE 6739 — Goldsman 8/5/20 46 / 108

Other Continuous Distributions

Lesson 4.6 — Other Continuous Distributions

Triangular(a, b, c) Distribution — good for modeling RVs on the basis oflimited data (minimum, mode, maximum).

f(x) =

2(x−a)

(b−a)(c−a) , a < x ≤ b2(c−x)

(c−b)(c−a) , b < x < c

0, otherwise.

E[X] =a+ b+ c

3and Var(X) = mess

ISYE 6739 — Goldsman 8/5/20 47 / 108

Other Continuous Distributions

Lesson 4.6 — Other Continuous Distributions

Triangular(a, b, c) Distribution — good for modeling RVs on the basis oflimited data (minimum, mode, maximum).

f(x) =

2(x−a)

(b−a)(c−a) , a < x ≤ b2(c−x)

(c−b)(c−a) , b < x < c

0, otherwise.

E[X] =a+ b+ c

3and Var(X) = mess

ISYE 6739 — Goldsman 8/5/20 47 / 108

Other Continuous Distributions

Lesson 4.6 — Other Continuous Distributions

Triangular(a, b, c) Distribution — good for modeling RVs on the basis oflimited data (minimum, mode, maximum).

f(x) =

2(x−a)

(b−a)(c−a) , a < x ≤ b2(c−x)

(c−b)(c−a) , b < x < c

0, otherwise.

E[X] =a+ b+ c

3and Var(X) = mess

ISYE 6739 — Goldsman 8/5/20 47 / 108

Other Continuous Distributions

Lesson 4.6 — Other Continuous Distributions

Triangular(a, b, c) Distribution — good for modeling RVs on the basis oflimited data (minimum, mode, maximum).

f(x) =

2(x−a)

(b−a)(c−a) , a < x ≤ b2(c−x)

(c−b)(c−a) , b < x < c

0, otherwise.

E[X] =a+ b+ c

3and

Var(X) = mess

ISYE 6739 — Goldsman 8/5/20 47 / 108

Other Continuous Distributions

Lesson 4.6 — Other Continuous Distributions

Triangular(a, b, c) Distribution — good for modeling RVs on the basis oflimited data (minimum, mode, maximum).

f(x) =

2(x−a)

(b−a)(c−a) , a < x ≤ b2(c−x)

(c−b)(c−a) , b < x < c

0, otherwise.

E[X] =a+ b+ c

3and Var(X) = mess

ISYE 6739 — Goldsman 8/5/20 47 / 108

Other Continuous Distributions

Beta(a, b) Distribution — good for modeling RVs that are restricted to aninterval.

f(x) =Γ(a+ b)

Γ(a)Γ(b)xa−1(1− x)b−1, 0 < x < 1.

E[X] =a

a+ band Var(X) =

ab

(a+ b)2(a+ b+ 1).

This distribution gets its name from the beta function, which is defined as

β(a, b) ≡ Γ(a)Γ(b)

Γ(a+ b)=

∫ 1

0xa−1(1− x)b−1 dx.

ISYE 6739 — Goldsman 8/5/20 48 / 108

Other Continuous Distributions

Beta(a, b) Distribution — good for modeling RVs that are restricted to aninterval.

f(x) =Γ(a+ b)

Γ(a)Γ(b)xa−1(1− x)b−1, 0 < x < 1.

E[X] =a

a+ band Var(X) =

ab

(a+ b)2(a+ b+ 1).

This distribution gets its name from the beta function, which is defined as

β(a, b) ≡ Γ(a)Γ(b)

Γ(a+ b)=

∫ 1

0xa−1(1− x)b−1 dx.

ISYE 6739 — Goldsman 8/5/20 48 / 108

Other Continuous Distributions

Beta(a, b) Distribution — good for modeling RVs that are restricted to aninterval.

f(x) =Γ(a+ b)

Γ(a)Γ(b)xa−1(1− x)b−1, 0 < x < 1.

E[X] =a

a+ band

Var(X) =ab

(a+ b)2(a+ b+ 1).

This distribution gets its name from the beta function, which is defined as

β(a, b) ≡ Γ(a)Γ(b)

Γ(a+ b)=

∫ 1

0xa−1(1− x)b−1 dx.

ISYE 6739 — Goldsman 8/5/20 48 / 108

Other Continuous Distributions

Beta(a, b) Distribution — good for modeling RVs that are restricted to aninterval.

f(x) =Γ(a+ b)

Γ(a)Γ(b)xa−1(1− x)b−1, 0 < x < 1.

E[X] =a

a+ band Var(X) =

ab

(a+ b)2(a+ b+ 1).

This distribution gets its name from the beta function, which is defined as

β(a, b) ≡ Γ(a)Γ(b)

Γ(a+ b)=

∫ 1

0xa−1(1− x)b−1 dx.

ISYE 6739 — Goldsman 8/5/20 48 / 108

Other Continuous Distributions

Beta(a, b) Distribution — good for modeling RVs that are restricted to aninterval.

f(x) =Γ(a+ b)

Γ(a)Γ(b)xa−1(1− x)b−1, 0 < x < 1.

E[X] =a

a+ band Var(X) =

ab

(a+ b)2(a+ b+ 1).

This distribution gets its name from the beta function, which is defined as

β(a, b) ≡ Γ(a)Γ(b)

Γ(a+ b)=

∫ 1

0xa−1(1− x)b−1 dx.

ISYE 6739 — Goldsman 8/5/20 48 / 108

Other Continuous Distributions

Beta(a, b) Distribution — good for modeling RVs that are restricted to aninterval.

f(x) =Γ(a+ b)

Γ(a)Γ(b)xa−1(1− x)b−1, 0 < x < 1.

E[X] =a

a+ band Var(X) =

ab

(a+ b)2(a+ b+ 1).

This distribution gets its name from the beta function, which is defined as

β(a, b) ≡ Γ(a)Γ(b)

Γ(a+ b)=

∫ 1

0xa−1(1− x)b−1 dx.

ISYE 6739 — Goldsman 8/5/20 48 / 108

Other Continuous Distributions

The Beta distribution is very flexible. Here are a few family portraits.

Remark: Certain versions of the Uniform (a = b = 1) and Triangulardistributions are special cases of the Beta.

ISYE 6739 — Goldsman 8/5/20 49 / 108

Other Continuous Distributions

The Beta distribution is very flexible. Here are a few family portraits.

Remark: Certain versions of the Uniform (a = b = 1) and Triangulardistributions are special cases of the Beta.

ISYE 6739 — Goldsman 8/5/20 49 / 108

Other Continuous Distributions

The Beta distribution is very flexible. Here are a few family portraits.

Remark: Certain versions of the Uniform (a = b = 1) and Triangulardistributions are special cases of the Beta.

ISYE 6739 — Goldsman 8/5/20 49 / 108

Other Continuous Distributions

Weibull(a, b) Distribution — good for modeling reliability models. a isthe “scale” parameter, and b is the “shape” parameter.

f(x) = ab(ax)b−1e−(ax)b, x > 0.

F (x) = 1− exp[−(ax)b], x > 0.

E[X] = (1/a)Γ(1 + (1/b)) and Var(X) = slight mess

Remark: The Exponential is a special case of the Weibull.

Example: Time-to-failure T for a transmitter has a Weibull distribution withparameters a = 1/(200 hrs) and b = 1/3. Then

E[T ] = 200Γ(1 + 3) = 1200 hrs.

The probability that it fails before 2000 hrs is

F (2000) = 1− exp[−(2000/200)1/3] = 0.884. 2

ISYE 6739 — Goldsman 8/5/20 50 / 108

Other Continuous Distributions

Weibull(a, b) Distribution — good for modeling reliability models. a isthe “scale” parameter, and b is the “shape” parameter.

f(x) = ab(ax)b−1e−(ax)b, x > 0.

F (x) = 1− exp[−(ax)b], x > 0.

E[X] = (1/a)Γ(1 + (1/b)) and Var(X) = slight mess

Remark: The Exponential is a special case of the Weibull.

Example: Time-to-failure T for a transmitter has a Weibull distribution withparameters a = 1/(200 hrs) and b = 1/3. Then

E[T ] = 200Γ(1 + 3) = 1200 hrs.

The probability that it fails before 2000 hrs is

F (2000) = 1− exp[−(2000/200)1/3] = 0.884. 2

ISYE 6739 — Goldsman 8/5/20 50 / 108

Other Continuous Distributions

Weibull(a, b) Distribution — good for modeling reliability models. a isthe “scale” parameter, and b is the “shape” parameter.

f(x) = ab(ax)b−1e−(ax)b, x > 0.

F (x) = 1− exp[−(ax)b], x > 0.

E[X] = (1/a)Γ(1 + (1/b)) and Var(X) = slight mess

Remark: The Exponential is a special case of the Weibull.

Example: Time-to-failure T for a transmitter has a Weibull distribution withparameters a = 1/(200 hrs) and b = 1/3. Then

E[T ] = 200Γ(1 + 3) = 1200 hrs.

The probability that it fails before 2000 hrs is

F (2000) = 1− exp[−(2000/200)1/3] = 0.884. 2

ISYE 6739 — Goldsman 8/5/20 50 / 108

Other Continuous Distributions

Weibull(a, b) Distribution — good for modeling reliability models. a isthe “scale” parameter, and b is the “shape” parameter.

f(x) = ab(ax)b−1e−(ax)b, x > 0.

F (x) = 1− exp[−(ax)b], x > 0.

E[X] = (1/a)Γ(1 + (1/b)) and Var(X) = slight mess

Remark: The Exponential is a special case of the Weibull.

Example: Time-to-failure T for a transmitter has a Weibull distribution withparameters a = 1/(200 hrs) and b = 1/3. Then

E[T ] = 200Γ(1 + 3) = 1200 hrs.

The probability that it fails before 2000 hrs is

F (2000) = 1− exp[−(2000/200)1/3] = 0.884. 2

ISYE 6739 — Goldsman 8/5/20 50 / 108

Other Continuous Distributions

Weibull(a, b) Distribution — good for modeling reliability models. a isthe “scale” parameter, and b is the “shape” parameter.

f(x) = ab(ax)b−1e−(ax)b, x > 0.

F (x) = 1− exp[−(ax)b], x > 0.

E[X] = (1/a)Γ(1 + (1/b)) and Var(X) = slight mess

Remark: The Exponential is a special case of the Weibull.

Example: Time-to-failure T for a transmitter has a Weibull distribution withparameters a = 1/(200 hrs) and b = 1/3. Then

E[T ] = 200Γ(1 + 3) = 1200 hrs.

The probability that it fails before 2000 hrs is

F (2000) = 1− exp[−(2000/200)1/3] = 0.884. 2

ISYE 6739 — Goldsman 8/5/20 50 / 108

Other Continuous Distributions

Weibull(a, b) Distribution — good for modeling reliability models. a isthe “scale” parameter, and b is the “shape” parameter.

f(x) = ab(ax)b−1e−(ax)b, x > 0.

F (x) = 1− exp[−(ax)b], x > 0.

E[X] = (1/a)Γ(1 + (1/b)) and Var(X) = slight mess

Remark: The Exponential is a special case of the Weibull.

Example: Time-to-failure T for a transmitter has a Weibull distribution withparameters a = 1/(200 hrs) and b = 1/3. Then

E[T ] = 200Γ(1 + 3) = 1200 hrs.

The probability that it fails before 2000 hrs is

F (2000) = 1− exp[−(2000/200)1/3] = 0.884. 2

ISYE 6739 — Goldsman 8/5/20 50 / 108

Other Continuous Distributions

Weibull(a, b) Distribution — good for modeling reliability models. a isthe “scale” parameter, and b is the “shape” parameter.

f(x) = ab(ax)b−1e−(ax)b, x > 0.

F (x) = 1− exp[−(ax)b], x > 0.

E[X] = (1/a)Γ(1 + (1/b)) and Var(X) = slight mess

Remark: The Exponential is a special case of the Weibull.

Example: Time-to-failure T for a transmitter has a Weibull distribution withparameters a = 1/(200 hrs) and b = 1/3. Then

E[T ] = 200Γ(1 + 3) = 1200 hrs.

The probability that it fails before 2000 hrs is

F (2000) = 1− exp[−(2000/200)1/3] = 0.884. 2

ISYE 6739 — Goldsman 8/5/20 50 / 108

Other Continuous Distributions

Weibull(a, b) Distribution — good for modeling reliability models. a isthe “scale” parameter, and b is the “shape” parameter.

f(x) = ab(ax)b−1e−(ax)b, x > 0.

F (x) = 1− exp[−(ax)b], x > 0.

E[X] = (1/a)Γ(1 + (1/b)) and Var(X) = slight mess

Remark: The Exponential is a special case of the Weibull.

Example: Time-to-failure T for a transmitter has a Weibull distribution withparameters a = 1/(200 hrs) and b = 1/3. Then

E[T ] = 200Γ(1 + 3) = 1200 hrs.

The probability that it fails before 2000 hrs is

F (2000) = 1− exp[−(2000/200)1/3] = 0.884. 2

ISYE 6739 — Goldsman 8/5/20 50 / 108

Other Continuous Distributions

Cauchy Distribution — A “fat-tailed” distribution good for disprovingthings!

f(x) =1

π(1 + x2)and F (x) =

1

2+

arctan(x)

π, x ∈ R.

Theorem: The Cauchy distribution has an undefined mean and infinitevariance!

Weird Fact: X1, . . . , Xniid∼ Cauchy⇒

∑ni=1Xi/n ∼ Cauchy. Even if you

take the average of a bunch of Cauchys, you’re right back where you started!

ISYE 6739 — Goldsman 8/5/20 51 / 108

Other Continuous Distributions

Cauchy Distribution — A “fat-tailed” distribution good for disprovingthings!

f(x) =1

π(1 + x2)and

F (x) =1

2+

arctan(x)

π, x ∈ R.

Theorem: The Cauchy distribution has an undefined mean and infinitevariance!

Weird Fact: X1, . . . , Xniid∼ Cauchy⇒

∑ni=1Xi/n ∼ Cauchy. Even if you

take the average of a bunch of Cauchys, you’re right back where you started!

ISYE 6739 — Goldsman 8/5/20 51 / 108

Other Continuous Distributions

Cauchy Distribution — A “fat-tailed” distribution good for disprovingthings!

f(x) =1

π(1 + x2)and F (x) =

1

2+

arctan(x)

π, x ∈ R.

Theorem: The Cauchy distribution has an undefined mean and infinitevariance!

Weird Fact: X1, . . . , Xniid∼ Cauchy⇒

∑ni=1Xi/n ∼ Cauchy. Even if you

take the average of a bunch of Cauchys, you’re right back where you started!

ISYE 6739 — Goldsman 8/5/20 51 / 108

Other Continuous Distributions

Cauchy Distribution — A “fat-tailed” distribution good for disprovingthings!

f(x) =1

π(1 + x2)and F (x) =

1

2+

arctan(x)

π, x ∈ R.

Theorem: The Cauchy distribution has an undefined mean and infinitevariance!

Weird Fact: X1, . . . , Xniid∼ Cauchy⇒

∑ni=1Xi/n ∼ Cauchy. Even if you

take the average of a bunch of Cauchys, you’re right back where you started!

ISYE 6739 — Goldsman 8/5/20 51 / 108

Other Continuous Distributions

Cauchy Distribution — A “fat-tailed” distribution good for disprovingthings!

f(x) =1

π(1 + x2)and F (x) =

1

2+

arctan(x)

π, x ∈ R.

Theorem: The Cauchy distribution has an undefined mean and infinitevariance!

Weird Fact: X1, . . . , Xniid∼ Cauchy⇒

∑ni=1Xi/n ∼ Cauchy. Even if you

take the average of a bunch of Cauchys, you’re right back where you started!

ISYE 6739 — Goldsman 8/5/20 51 / 108

Other Continuous Distributions

Alphabet Soup of Other Distributions

χ2 distribution — coming up when we talk Statistics

t distribution — coming up

F distribution — coming up

Pareto, LaPlace, Rayleigh, Gumbel, Johnson distributions

Etc. . . .

ISYE 6739 — Goldsman 8/5/20 52 / 108

Other Continuous Distributions

Alphabet Soup of Other Distributions

χ2 distribution — coming up when we talk Statistics

t distribution — coming up

F distribution — coming up

Pareto, LaPlace, Rayleigh, Gumbel, Johnson distributions

Etc. . . .

ISYE 6739 — Goldsman 8/5/20 52 / 108

Other Continuous Distributions

Alphabet Soup of Other Distributions

χ2 distribution — coming up when we talk Statistics

t distribution — coming up

F distribution — coming up

Pareto, LaPlace, Rayleigh, Gumbel, Johnson distributions

Etc. . . .

ISYE 6739 — Goldsman 8/5/20 52 / 108

Other Continuous Distributions

Alphabet Soup of Other Distributions

χ2 distribution — coming up when we talk Statistics

t distribution — coming up

F distribution — coming up

Pareto, LaPlace, Rayleigh, Gumbel, Johnson distributions

Etc. . . .

ISYE 6739 — Goldsman 8/5/20 52 / 108

Other Continuous Distributions

Alphabet Soup of Other Distributions

χ2 distribution — coming up when we talk Statistics

t distribution — coming up

F distribution — coming up

Pareto, LaPlace, Rayleigh, Gumbel, Johnson distributions

Etc. . . .

ISYE 6739 — Goldsman 8/5/20 52 / 108

Other Continuous Distributions

Alphabet Soup of Other Distributions

χ2 distribution — coming up when we talk Statistics

t distribution — coming up

F distribution — coming up

Pareto, LaPlace, Rayleigh, Gumbel, Johnson distributions

Etc. . . .

ISYE 6739 — Goldsman 8/5/20 52 / 108

Normal Distribution: Basics

1 Bernoulli and Binomial Distributions

2 Hypergeometric Distribution3 Geometric and Negative Binomial Distributions

4 Poisson Distribution5 Uniform, Exponential, and Friends6 Other Continuous Distributions7 Normal Distribution: Basics8 Standard Normal Distribution9 Sample Mean of Normals

10 The Central Limit Theorem + Proof

11 Central Limit Theorem Examples

12 Extensions — Multivariate Normal Distribution13 Extensions — Lognormal Distribution

14 Computer Stuff

ISYE 6739 — Goldsman 8/5/20 53 / 108

Normal Distribution: Basics

Lesson 4.7 — Normal Distribution: Basics

The Normal Distribution is so important that we’re giving it an entire section.

Definition: X ∼ Nor(µ, σ2) if it has pdf

f(x) =1√

2πσ2exp

[−(x− µ)2

2σ2

], ∀x ∈ R.

Remark: The Normal distribution is also called the Gaussian distribution.

Examples: Heights, weights, SAT scores, crop yields, and averages ofthings tend to be normal.

ISYE 6739 — Goldsman 8/5/20 54 / 108

Normal Distribution: Basics

Lesson 4.7 — Normal Distribution: Basics

The Normal Distribution is so important that we’re giving it an entire section.

Definition: X ∼ Nor(µ, σ2) if it has pdf

f(x) =1√

2πσ2exp

[−(x− µ)2

2σ2

], ∀x ∈ R.

Remark: The Normal distribution is also called the Gaussian distribution.

Examples: Heights, weights, SAT scores, crop yields, and averages ofthings tend to be normal.

ISYE 6739 — Goldsman 8/5/20 54 / 108

Normal Distribution: Basics

Lesson 4.7 — Normal Distribution: Basics

The Normal Distribution is so important that we’re giving it an entire section.

Definition: X ∼ Nor(µ, σ2) if it has pdf

f(x) =1√

2πσ2exp

[−(x− µ)2

2σ2

], ∀x ∈ R.

Remark: The Normal distribution is also called the Gaussian distribution.

Examples: Heights, weights, SAT scores, crop yields, and averages ofthings tend to be normal.

ISYE 6739 — Goldsman 8/5/20 54 / 108

Normal Distribution: Basics

Lesson 4.7 — Normal Distribution: Basics

The Normal Distribution is so important that we’re giving it an entire section.

Definition: X ∼ Nor(µ, σ2) if it has pdf

f(x) =1√

2πσ2exp

[−(x− µ)2

2σ2

], ∀x ∈ R.

Remark: The Normal distribution is also called the Gaussian distribution.

Examples: Heights, weights, SAT scores, crop yields, and averages ofthings tend to be normal.

ISYE 6739 — Goldsman 8/5/20 54 / 108

Normal Distribution: Basics

Lesson 4.7 — Normal Distribution: Basics

The Normal Distribution is so important that we’re giving it an entire section.

Definition: X ∼ Nor(µ, σ2) if it has pdf

f(x) =1√

2πσ2exp

[−(x− µ)2

2σ2

], ∀x ∈ R.

Remark: The Normal distribution is also called the Gaussian distribution.

Examples: Heights, weights, SAT scores, crop yields, and averages ofthings tend to be normal.

ISYE 6739 — Goldsman 8/5/20 54 / 108

Normal Distribution: Basics

The pdf f(x) is “bell-shaped” and symmetric around x = µ, with tails fallingoff quickly as you move away from µ.

Small σ2 corresponds to a “tall, skinny” bell curve; large σ2 gives a “short,fat” bell curve.

ISYE 6739 — Goldsman 8/5/20 55 / 108

Normal Distribution: Basics

The pdf f(x) is “bell-shaped” and symmetric around x = µ, with tails fallingoff quickly as you move away from µ.

Small σ2 corresponds to a “tall, skinny” bell curve; large σ2 gives a “short,fat” bell curve.

ISYE 6739 — Goldsman 8/5/20 55 / 108

Normal Distribution: Basics

The pdf f(x) is “bell-shaped” and symmetric around x = µ, with tails fallingoff quickly as you move away from µ.

Small σ2 corresponds to a “tall, skinny” bell curve; large σ2 gives a “short,fat” bell curve.

ISYE 6739 — Goldsman 8/5/20 55 / 108

Normal Distribution: Basics

Fun Fact (1):∫R f(x) dx = 1.

Proof: Transform to polar coordinates. Good luck. 2

Fun Fact (2): The cdf is

F (x) =

∫ x

−∞

1√2πσ2

exp

[−(t− µ)2

2σ2

]dt = ??

Remark: No closed-form solution for this. Stay tuned.

Fun Facts (3) and (4): E[X] = µ and Var(X) = σ2.

Proof: Integration by parts or mgf (below).

ISYE 6739 — Goldsman 8/5/20 56 / 108

Normal Distribution: Basics

Fun Fact (1):∫R f(x) dx = 1.

Proof: Transform to polar coordinates. Good luck. 2

Fun Fact (2): The cdf is

F (x) =

∫ x

−∞

1√2πσ2

exp

[−(t− µ)2

2σ2

]dt = ??

Remark: No closed-form solution for this. Stay tuned.

Fun Facts (3) and (4): E[X] = µ and Var(X) = σ2.

Proof: Integration by parts or mgf (below).

ISYE 6739 — Goldsman 8/5/20 56 / 108

Normal Distribution: Basics

Fun Fact (1):∫R f(x) dx = 1.

Proof: Transform to polar coordinates. Good luck. 2

Fun Fact (2): The cdf is

F (x) =

∫ x

−∞

1√2πσ2

exp

[−(t− µ)2

2σ2

]dt = ??

Remark: No closed-form solution for this. Stay tuned.

Fun Facts (3) and (4): E[X] = µ and Var(X) = σ2.

Proof: Integration by parts or mgf (below).

ISYE 6739 — Goldsman 8/5/20 56 / 108

Normal Distribution: Basics

Fun Fact (1):∫R f(x) dx = 1.

Proof: Transform to polar coordinates. Good luck. 2

Fun Fact (2): The cdf is

F (x) =

∫ x

−∞

1√2πσ2

exp

[−(t− µ)2

2σ2

]dt = ??

Remark: No closed-form solution for this. Stay tuned.

Fun Facts (3) and (4): E[X] = µ and Var(X) = σ2.

Proof: Integration by parts or mgf (below).

ISYE 6739 — Goldsman 8/5/20 56 / 108

Normal Distribution: Basics

Fun Fact (1):∫R f(x) dx = 1.

Proof: Transform to polar coordinates. Good luck. 2

Fun Fact (2): The cdf is

F (x) =

∫ x

−∞

1√2πσ2

exp

[−(t− µ)2

2σ2

]dt = ??

Remark: No closed-form solution for this. Stay tuned.

Fun Facts (3) and (4): E[X] = µ and Var(X) = σ2.

Proof: Integration by parts or mgf (below).

ISYE 6739 — Goldsman 8/5/20 56 / 108

Normal Distribution: Basics

Fun Fact (1):∫R f(x) dx = 1.

Proof: Transform to polar coordinates. Good luck. 2

Fun Fact (2): The cdf is

F (x) =

∫ x

−∞

1√2πσ2

exp

[−(t− µ)2

2σ2

]dt = ??

Remark: No closed-form solution for this. Stay tuned.

Fun Facts (3) and (4): E[X] = µ and Var(X) = σ2.

Proof: Integration by parts or mgf (below).

ISYE 6739 — Goldsman 8/5/20 56 / 108

Normal Distribution: Basics

Fun Fact (5): If X is any Normal RV, then

P (µ− σ < X < µ+ σ) = 0.6827

P (µ− 2σ < X < µ+ 2σ) = 0.9545

P (µ− 3σ < X < µ+ 3σ) = 0.9973.

So almost all of the probability is contained within 3 standard deviations ofthe mean. (This is sort of what Toyota is referring to when it brags about“six-sigma” quality.)

Fun Fact (6): The mgf is MX(t) = exp(µt+ 12σ

2t2).

Proof: Calculus (or look it up in a table of integrals).

ISYE 6739 — Goldsman 8/5/20 57 / 108

Normal Distribution: Basics

Fun Fact (5): If X is any Normal RV, then

P (µ− σ < X < µ+ σ) = 0.6827

P (µ− 2σ < X < µ+ 2σ) = 0.9545

P (µ− 3σ < X < µ+ 3σ) = 0.9973.

So almost all of the probability is contained within 3 standard deviations ofthe mean. (This is sort of what Toyota is referring to when it brags about“six-sigma” quality.)

Fun Fact (6): The mgf is MX(t) = exp(µt+ 12σ

2t2).

Proof: Calculus (or look it up in a table of integrals).

ISYE 6739 — Goldsman 8/5/20 57 / 108

Normal Distribution: Basics

Fun Fact (5): If X is any Normal RV, then

P (µ− σ < X < µ+ σ) = 0.6827

P (µ− 2σ < X < µ+ 2σ) = 0.9545

P (µ− 3σ < X < µ+ 3σ) = 0.9973.

So almost all of the probability is contained within 3 standard deviations ofthe mean. (This is sort of what Toyota is referring to when it brags about“six-sigma” quality.)

Fun Fact (6): The mgf is MX(t) = exp(µt+ 12σ

2t2).

Proof: Calculus (or look it up in a table of integrals).

ISYE 6739 — Goldsman 8/5/20 57 / 108

Normal Distribution: Basics

Fun Fact (5): If X is any Normal RV, then

P (µ− σ < X < µ+ σ) = 0.6827

P (µ− 2σ < X < µ+ 2σ) = 0.9545

P (µ− 3σ < X < µ+ 3σ) = 0.9973.

So almost all of the probability is contained within 3 standard deviations ofthe mean. (This is sort of what Toyota is referring to when it brags about“six-sigma” quality.)

Fun Fact (6): The mgf is MX(t) = exp(µt+ 12σ

2t2).

Proof: Calculus (or look it up in a table of integrals).

ISYE 6739 — Goldsman 8/5/20 57 / 108

Normal Distribution: Basics

Fun Fact (5): If X is any Normal RV, then

P (µ− σ < X < µ+ σ) = 0.6827

P (µ− 2σ < X < µ+ 2σ) = 0.9545

P (µ− 3σ < X < µ+ 3σ) = 0.9973.

So almost all of the probability is contained within 3 standard deviations ofthe mean. (This is sort of what Toyota is referring to when it brags about“six-sigma” quality.)

Fun Fact (6): The mgf is MX(t) = exp(µt+ 12σ

2t2).

Proof: Calculus (or look it up in a table of integrals).

ISYE 6739 — Goldsman 8/5/20 57 / 108

Normal Distribution: Basics

Fun Fact (5): If X is any Normal RV, then

P (µ− σ < X < µ+ σ) = 0.6827

P (µ− 2σ < X < µ+ 2σ) = 0.9545

P (µ− 3σ < X < µ+ 3σ) = 0.9973.

So almost all of the probability is contained within 3 standard deviations ofthe mean. (This is sort of what Toyota is referring to when it brags about“six-sigma” quality.)

Fun Fact (6): The mgf is MX(t) = exp(µt+ 12σ

2t2).

Proof: Calculus (or look it up in a table of integrals).

ISYE 6739 — Goldsman 8/5/20 57 / 108

Normal Distribution: Basics

Theorem (Additive Property of Normals): If X1, . . . , Xn are independentwith Xi ∼ Nor(µi, σ

2i ), i = 1, . . . , n, then

Y ≡n∑i=1

aiXi + b ∼ Nor( n∑i=1

aiµi + b,

n∑i=1

a2iσ2i

).

So a linear combination of independent normals is itself normal.

ISYE 6739 — Goldsman 8/5/20 58 / 108

Normal Distribution: Basics

Theorem (Additive Property of Normals): If X1, . . . , Xn are independentwith Xi ∼ Nor(µi, σ

2i ), i = 1, . . . , n, then

Y ≡n∑i=1

aiXi + b ∼ Nor( n∑i=1

aiµi + b,

n∑i=1

a2iσ2i

).

So a linear combination of independent normals is itself normal.

ISYE 6739 — Goldsman 8/5/20 58 / 108

Normal Distribution: Basics

Theorem (Additive Property of Normals): If X1, . . . , Xn are independentwith Xi ∼ Nor(µi, σ

2i ), i = 1, . . . , n, then

Y ≡n∑i=1

aiXi + b ∼ Nor( n∑i=1

aiµi + b,

n∑i=1

a2iσ2i

).

So a linear combination of independent normals is itself normal.

ISYE 6739 — Goldsman 8/5/20 58 / 108

Normal Distribution: Basics

Proof: Since Y is a linear function,

MY (t) = M∑i aiXi+b(t)

= etbM∑i aiXi

(t)

= etbn∏i=1

MaiXi(t) (Xi’s independent)

= etbn∏i=1

MXi(ait) (mgf of linear function)

= etbn∏i=1

exp[µi(ait) +

1

2σ2i (ait)

2]

(normal mgf)

= exp

[( n∑i=1

µiai + b)t+

1

2

( n∑i=1

a2iσ2i

)t2],

and we are done by mgf uniqueness. 2

ISYE 6739 — Goldsman 8/5/20 59 / 108

Normal Distribution: Basics

Proof: Since Y is a linear function,

MY (t) = M∑i aiXi+b(t) = etbM∑

i aiXi(t)

= etbn∏i=1

MaiXi(t) (Xi’s independent)

= etbn∏i=1

MXi(ait) (mgf of linear function)

= etbn∏i=1

exp[µi(ait) +

1

2σ2i (ait)

2]

(normal mgf)

= exp

[( n∑i=1

µiai + b)t+

1

2

( n∑i=1

a2iσ2i

)t2],

and we are done by mgf uniqueness. 2

ISYE 6739 — Goldsman 8/5/20 59 / 108

Normal Distribution: Basics

Proof: Since Y is a linear function,

MY (t) = M∑i aiXi+b(t) = etbM∑

i aiXi(t)

= etbn∏i=1

MaiXi(t) (Xi’s independent)

= etbn∏i=1

MXi(ait) (mgf of linear function)

= etbn∏i=1

exp[µi(ait) +

1

2σ2i (ait)

2]

(normal mgf)

= exp

[( n∑i=1

µiai + b)t+

1

2

( n∑i=1

a2iσ2i

)t2],

and we are done by mgf uniqueness. 2

ISYE 6739 — Goldsman 8/5/20 59 / 108

Normal Distribution: Basics

Proof: Since Y is a linear function,

MY (t) = M∑i aiXi+b(t) = etbM∑

i aiXi(t)

= etbn∏i=1

MaiXi(t) (Xi’s independent)

= etbn∏i=1

MXi(ait) (mgf of linear function)

= etbn∏i=1

exp[µi(ait) +

1

2σ2i (ait)

2]

(normal mgf)

= exp

[( n∑i=1

µiai + b)t+

1

2

( n∑i=1

a2iσ2i

)t2],

and we are done by mgf uniqueness. 2

ISYE 6739 — Goldsman 8/5/20 59 / 108

Normal Distribution: Basics

Proof: Since Y is a linear function,

MY (t) = M∑i aiXi+b(t) = etbM∑

i aiXi(t)

= etbn∏i=1

MaiXi(t) (Xi’s independent)

= etbn∏i=1

MXi(ait) (mgf of linear function)

= etbn∏i=1

exp[µi(ait) +

1

2σ2i (ait)

2]

(normal mgf)

= exp

[( n∑i=1

µiai + b)t+

1

2

( n∑i=1

a2iσ2i

)t2],

and we are done by mgf uniqueness. 2

ISYE 6739 — Goldsman 8/5/20 59 / 108

Normal Distribution: Basics

Proof: Since Y is a linear function,

MY (t) = M∑i aiXi+b(t) = etbM∑

i aiXi(t)

= etbn∏i=1

MaiXi(t) (Xi’s independent)

= etbn∏i=1

MXi(ait) (mgf of linear function)

= etbn∏i=1

exp[µi(ait) +

1

2σ2i (ait)

2]

(normal mgf)

= exp

[( n∑i=1

µiai + b)t+

1

2

( n∑i=1

a2iσ2i

)t2],

and we are done by mgf uniqueness. 2

ISYE 6739 — Goldsman 8/5/20 59 / 108

Normal Distribution: Basics

Proof: Since Y is a linear function,

MY (t) = M∑i aiXi+b(t) = etbM∑

i aiXi(t)

= etbn∏i=1

MaiXi(t) (Xi’s independent)

= etbn∏i=1

MXi(ait) (mgf of linear function)

= etbn∏i=1

exp[µi(ait) +

1

2σ2i (ait)

2]

(normal mgf)

= exp

[( n∑i=1

µiai + b)t+

1

2

( n∑i=1

a2iσ2i

)t2],

and we are done by mgf uniqueness. 2

ISYE 6739 — Goldsman 8/5/20 59 / 108

Normal Distribution: Basics

Remark: A normal distribution is completely characterized by its mean andvariance.

By the above, we know that a linear combination of independent normals isstill normal. Therefore, when we add up independent normals, all we have todo is figure out the mean and variance — the normality of the sum comes forfree.

Example: X ∼ Nor(3, 4), Y ∼ Nor(4, 6), and X,Y are independent. Findthe distribution of 2X − 3Y .

Solution: This is normal with

E[2X − 3Y ] = 2E[X]− 3E[Y ] = 2(3)− 3(4) = −6

andVar(2X − 3Y ) = 4Var(X) + 9Var(Y ) = 70.

Thus, 2X − 3Y ∼ Nor(−6, 70). 2

ISYE 6739 — Goldsman 8/5/20 60 / 108

Normal Distribution: Basics

Remark: A normal distribution is completely characterized by its mean andvariance.

By the above, we know that a linear combination of independent normals isstill normal.

Therefore, when we add up independent normals, all we have todo is figure out the mean and variance — the normality of the sum comes forfree.

Example: X ∼ Nor(3, 4), Y ∼ Nor(4, 6), and X,Y are independent. Findthe distribution of 2X − 3Y .

Solution: This is normal with

E[2X − 3Y ] = 2E[X]− 3E[Y ] = 2(3)− 3(4) = −6

andVar(2X − 3Y ) = 4Var(X) + 9Var(Y ) = 70.

Thus, 2X − 3Y ∼ Nor(−6, 70). 2

ISYE 6739 — Goldsman 8/5/20 60 / 108

Normal Distribution: Basics

Remark: A normal distribution is completely characterized by its mean andvariance.

By the above, we know that a linear combination of independent normals isstill normal. Therefore, when we add up independent normals, all we have todo is figure out the mean and variance — the normality of the sum comes forfree.

Example: X ∼ Nor(3, 4), Y ∼ Nor(4, 6), and X,Y are independent. Findthe distribution of 2X − 3Y .

Solution: This is normal with

E[2X − 3Y ] = 2E[X]− 3E[Y ] = 2(3)− 3(4) = −6

andVar(2X − 3Y ) = 4Var(X) + 9Var(Y ) = 70.

Thus, 2X − 3Y ∼ Nor(−6, 70). 2

ISYE 6739 — Goldsman 8/5/20 60 / 108

Normal Distribution: Basics

Remark: A normal distribution is completely characterized by its mean andvariance.

By the above, we know that a linear combination of independent normals isstill normal. Therefore, when we add up independent normals, all we have todo is figure out the mean and variance — the normality of the sum comes forfree.

Example: X ∼ Nor(3, 4), Y ∼ Nor(4, 6), and X,Y are independent. Findthe distribution of 2X − 3Y .

Solution: This is normal with

E[2X − 3Y ] = 2E[X]− 3E[Y ] = 2(3)− 3(4) = −6

andVar(2X − 3Y ) = 4Var(X) + 9Var(Y ) = 70.

Thus, 2X − 3Y ∼ Nor(−6, 70). 2

ISYE 6739 — Goldsman 8/5/20 60 / 108

Normal Distribution: Basics

Remark: A normal distribution is completely characterized by its mean andvariance.

By the above, we know that a linear combination of independent normals isstill normal. Therefore, when we add up independent normals, all we have todo is figure out the mean and variance — the normality of the sum comes forfree.

Example: X ∼ Nor(3, 4), Y ∼ Nor(4, 6), and X,Y are independent. Findthe distribution of 2X − 3Y .

Solution: This is normal with

E[2X − 3Y ] = 2E[X]− 3E[Y ] = 2(3)− 3(4) = −6

andVar(2X − 3Y ) = 4Var(X) + 9Var(Y ) = 70.

Thus, 2X − 3Y ∼ Nor(−6, 70). 2

ISYE 6739 — Goldsman 8/5/20 60 / 108

Normal Distribution: Basics

Remark: A normal distribution is completely characterized by its mean andvariance.

By the above, we know that a linear combination of independent normals isstill normal. Therefore, when we add up independent normals, all we have todo is figure out the mean and variance — the normality of the sum comes forfree.

Example: X ∼ Nor(3, 4), Y ∼ Nor(4, 6), and X,Y are independent. Findthe distribution of 2X − 3Y .

Solution: This is normal with

E[2X − 3Y ] = 2E[X]− 3E[Y ] = 2(3)− 3(4) = −6

andVar(2X − 3Y ) = 4Var(X) + 9Var(Y ) = 70.

Thus, 2X − 3Y ∼ Nor(−6, 70). 2

ISYE 6739 — Goldsman 8/5/20 60 / 108

Normal Distribution: Basics

Remark: A normal distribution is completely characterized by its mean andvariance.

By the above, we know that a linear combination of independent normals isstill normal. Therefore, when we add up independent normals, all we have todo is figure out the mean and variance — the normality of the sum comes forfree.

Example: X ∼ Nor(3, 4), Y ∼ Nor(4, 6), and X,Y are independent. Findthe distribution of 2X − 3Y .

Solution: This is normal with

E[2X − 3Y ] = 2E[X]− 3E[Y ] = 2(3)− 3(4) = −6

andVar(2X − 3Y ) = 4Var(X) + 9Var(Y ) = 70.

Thus, 2X − 3Y ∼ Nor(−6, 70). 2

ISYE 6739 — Goldsman 8/5/20 60 / 108

Normal Distribution: Basics

Corollary (of Additive Property Theorem):

X ∼ Nor(µ, σ2) ⇒ aX + b ∼ Nor(aµ+ b, a2σ2).

Proof: Immediate from the Additive Property after noting thatE[aX + b] = aµ+ b and Var(aX + b) = a2σ2. 2

Corollary (of Corollary):

X ∼ Nor(µ, σ2) ⇒ Z ≡ X − µσ

∼ Nor(0, 1).

Proof: Use above Corollary with a = 1/σ and b = −µ/σ. 2

The manipulation described in this corollary is referred to asstandardization.

ISYE 6739 — Goldsman 8/5/20 61 / 108

Normal Distribution: Basics

Corollary (of Additive Property Theorem):

X ∼ Nor(µ, σ2) ⇒ aX + b ∼ Nor(aµ+ b, a2σ2).

Proof: Immediate from the Additive Property after noting thatE[aX + b] = aµ+ b and Var(aX + b) = a2σ2. 2

Corollary (of Corollary):

X ∼ Nor(µ, σ2) ⇒ Z ≡ X − µσ

∼ Nor(0, 1).

Proof: Use above Corollary with a = 1/σ and b = −µ/σ. 2

The manipulation described in this corollary is referred to asstandardization.

ISYE 6739 — Goldsman 8/5/20 61 / 108

Normal Distribution: Basics

Corollary (of Additive Property Theorem):

X ∼ Nor(µ, σ2) ⇒ aX + b ∼ Nor(aµ+ b, a2σ2).

Proof: Immediate from the Additive Property after noting thatE[aX + b] = aµ+ b and Var(aX + b) = a2σ2. 2

Corollary (of Corollary):

X ∼ Nor(µ, σ2) ⇒ Z ≡ X − µσ

∼ Nor(0, 1).

Proof: Use above Corollary with a = 1/σ and b = −µ/σ. 2

The manipulation described in this corollary is referred to asstandardization.

ISYE 6739 — Goldsman 8/5/20 61 / 108

Normal Distribution: Basics

Corollary (of Additive Property Theorem):

X ∼ Nor(µ, σ2) ⇒ aX + b ∼ Nor(aµ+ b, a2σ2).

Proof: Immediate from the Additive Property after noting thatE[aX + b] = aµ+ b and Var(aX + b) = a2σ2. 2

Corollary (of Corollary):

X ∼ Nor(µ, σ2) ⇒ Z ≡ X − µσ

∼ Nor(0, 1).

Proof: Use above Corollary with a = 1/σ and b = −µ/σ. 2

The manipulation described in this corollary is referred to asstandardization.

ISYE 6739 — Goldsman 8/5/20 61 / 108

Normal Distribution: Basics

Corollary (of Additive Property Theorem):

X ∼ Nor(µ, σ2) ⇒ aX + b ∼ Nor(aµ+ b, a2σ2).

Proof: Immediate from the Additive Property after noting thatE[aX + b] = aµ+ b and Var(aX + b) = a2σ2. 2

Corollary (of Corollary):

X ∼ Nor(µ, σ2) ⇒ Z ≡ X − µσ

∼ Nor(0, 1).

Proof: Use above Corollary with a = 1/σ and b = −µ/σ. 2

The manipulation described in this corollary is referred to asstandardization.

ISYE 6739 — Goldsman 8/5/20 61 / 108

Normal Distribution: Basics

Corollary (of Additive Property Theorem):

X ∼ Nor(µ, σ2) ⇒ aX + b ∼ Nor(aµ+ b, a2σ2).

Proof: Immediate from the Additive Property after noting thatE[aX + b] = aµ+ b and Var(aX + b) = a2σ2. 2

Corollary (of Corollary):

X ∼ Nor(µ, σ2) ⇒ Z ≡ X − µσ

∼ Nor(0, 1).

Proof: Use above Corollary with a = 1/σ and b = −µ/σ. 2

The manipulation described in this corollary is referred to asstandardization.

ISYE 6739 — Goldsman 8/5/20 61 / 108

Normal Distribution: Basics

Corollary (of Additive Property Theorem):

X ∼ Nor(µ, σ2) ⇒ aX + b ∼ Nor(aµ+ b, a2σ2).

Proof: Immediate from the Additive Property after noting thatE[aX + b] = aµ+ b and Var(aX + b) = a2σ2. 2

Corollary (of Corollary):

X ∼ Nor(µ, σ2) ⇒ Z ≡ X − µσ

∼ Nor(0, 1).

Proof: Use above Corollary with a = 1/σ and b = −µ/σ. 2

The manipulation described in this corollary is referred to asstandardization.

ISYE 6739 — Goldsman 8/5/20 61 / 108

Standard Normal Distribution

1 Bernoulli and Binomial Distributions

2 Hypergeometric Distribution3 Geometric and Negative Binomial Distributions

4 Poisson Distribution5 Uniform, Exponential, and Friends6 Other Continuous Distributions7 Normal Distribution: Basics8 Standard Normal Distribution9 Sample Mean of Normals

10 The Central Limit Theorem + Proof

11 Central Limit Theorem Examples

12 Extensions — Multivariate Normal Distribution13 Extensions — Lognormal Distribution

14 Computer Stuff

ISYE 6739 — Goldsman 8/5/20 62 / 108

Standard Normal Distribution

Lesson 4.8 — Standard Normal Distribution

Definition: The Nor(0, 1) is called the standard normal distribution,and is often denoted by Z.

The Nor(0, 1) is nice because there are tables available for its cdf.

You can standardize any normal RV X into a standard normal by applying thetransformation Z = (X − µ)/σ. Then you can use the cdf tables.

The pdf of the Nor(0, 1) is

φ(z) ≡ 1√2πe−z

2/2, z ∈ R.

The cdf isΦ(z) ≡

∫ z

−∞φ(t) dt, z ∈ R.

ISYE 6739 — Goldsman 8/5/20 63 / 108

Standard Normal Distribution

Lesson 4.8 — Standard Normal Distribution

Definition: The Nor(0, 1) is called the standard normal distribution,and is often denoted by Z.

The Nor(0, 1) is nice because there are tables available for its cdf.

You can standardize any normal RV X into a standard normal by applying thetransformation Z = (X − µ)/σ. Then you can use the cdf tables.

The pdf of the Nor(0, 1) is

φ(z) ≡ 1√2πe−z

2/2, z ∈ R.

The cdf isΦ(z) ≡

∫ z

−∞φ(t) dt, z ∈ R.

ISYE 6739 — Goldsman 8/5/20 63 / 108

Standard Normal Distribution

Lesson 4.8 — Standard Normal Distribution

Definition: The Nor(0, 1) is called the standard normal distribution,and is often denoted by Z.

The Nor(0, 1) is nice because there are tables available for its cdf.

You can standardize any normal RV X into a standard normal by applying thetransformation Z = (X − µ)/σ. Then you can use the cdf tables.

The pdf of the Nor(0, 1) is

φ(z) ≡ 1√2πe−z

2/2, z ∈ R.

The cdf isΦ(z) ≡

∫ z

−∞φ(t) dt, z ∈ R.

ISYE 6739 — Goldsman 8/5/20 63 / 108

Standard Normal Distribution

Lesson 4.8 — Standard Normal Distribution

Definition: The Nor(0, 1) is called the standard normal distribution,and is often denoted by Z.

The Nor(0, 1) is nice because there are tables available for its cdf.

You can standardize any normal RV X into a standard normal by applying thetransformation Z = (X − µ)/σ. Then you can use the cdf tables.

The pdf of the Nor(0, 1) is

φ(z) ≡ 1√2πe−z

2/2, z ∈ R.

The cdf isΦ(z) ≡

∫ z

−∞φ(t) dt, z ∈ R.

ISYE 6739 — Goldsman 8/5/20 63 / 108

Standard Normal Distribution

Lesson 4.8 — Standard Normal Distribution

Definition: The Nor(0, 1) is called the standard normal distribution,and is often denoted by Z.

The Nor(0, 1) is nice because there are tables available for its cdf.

You can standardize any normal RV X into a standard normal by applying thetransformation Z = (X − µ)/σ. Then you can use the cdf tables.

The pdf of the Nor(0, 1) is

φ(z) ≡ 1√2πe−z

2/2, z ∈ R.

The cdf isΦ(z) ≡

∫ z

−∞φ(t) dt, z ∈ R.

ISYE 6739 — Goldsman 8/5/20 63 / 108

Standard Normal Distribution

Lesson 4.8 — Standard Normal Distribution

Definition: The Nor(0, 1) is called the standard normal distribution,and is often denoted by Z.

The Nor(0, 1) is nice because there are tables available for its cdf.

You can standardize any normal RV X into a standard normal by applying thetransformation Z = (X − µ)/σ. Then you can use the cdf tables.

The pdf of the Nor(0, 1) is

φ(z) ≡ 1√2πe−z

2/2, z ∈ R.

The cdf isΦ(z) ≡

∫ z

−∞φ(t) dt, z ∈ R.

ISYE 6739 — Goldsman 8/5/20 63 / 108

Standard Normal Distribution

Remarks: The following results are easy to derive, usually via symmetryarguments.

P (Z ≤ a) = Φ(a)

P (Z ≥ b) = 1− Φ(b)P (a ≤ Z ≤ b) = Φ(b)− Φ(a)

Φ(0) = 1/2Φ(−b) = P (Z ≤ −b) = P (Z ≥ b) = 1− Φ(b)P (−b ≤ Z ≤ b) = Φ(b)− Φ(−b) = 2Φ(b)− 1

Then

P (µ− kσ ≤ X ≤ µ+ kσ) = P (−k ≤ Z ≤ k) = 2Φ(k)− 1.

So the probability that any normal RV is within k standard deviations of itsmean doesn’t depend on the mean or variance.

ISYE 6739 — Goldsman 8/5/20 64 / 108

Standard Normal Distribution

Remarks: The following results are easy to derive, usually via symmetryarguments.

P (Z ≤ a) = Φ(a)P (Z ≥ b) = 1− Φ(b)

P (a ≤ Z ≤ b) = Φ(b)− Φ(a)

Φ(0) = 1/2Φ(−b) = P (Z ≤ −b) = P (Z ≥ b) = 1− Φ(b)P (−b ≤ Z ≤ b) = Φ(b)− Φ(−b) = 2Φ(b)− 1

Then

P (µ− kσ ≤ X ≤ µ+ kσ) = P (−k ≤ Z ≤ k) = 2Φ(k)− 1.

So the probability that any normal RV is within k standard deviations of itsmean doesn’t depend on the mean or variance.

ISYE 6739 — Goldsman 8/5/20 64 / 108

Standard Normal Distribution

Remarks: The following results are easy to derive, usually via symmetryarguments.

P (Z ≤ a) = Φ(a)P (Z ≥ b) = 1− Φ(b)P (a ≤ Z ≤ b) = Φ(b)− Φ(a)

Φ(0) = 1/2Φ(−b) = P (Z ≤ −b) = P (Z ≥ b) = 1− Φ(b)P (−b ≤ Z ≤ b) = Φ(b)− Φ(−b) = 2Φ(b)− 1

Then

P (µ− kσ ≤ X ≤ µ+ kσ) = P (−k ≤ Z ≤ k) = 2Φ(k)− 1.

So the probability that any normal RV is within k standard deviations of itsmean doesn’t depend on the mean or variance.

ISYE 6739 — Goldsman 8/5/20 64 / 108

Standard Normal Distribution

Remarks: The following results are easy to derive, usually via symmetryarguments.

P (Z ≤ a) = Φ(a)P (Z ≥ b) = 1− Φ(b)P (a ≤ Z ≤ b) = Φ(b)− Φ(a)

Φ(0) = 1/2

Φ(−b) = P (Z ≤ −b) = P (Z ≥ b) = 1− Φ(b)P (−b ≤ Z ≤ b) = Φ(b)− Φ(−b) = 2Φ(b)− 1

Then

P (µ− kσ ≤ X ≤ µ+ kσ) = P (−k ≤ Z ≤ k) = 2Φ(k)− 1.

So the probability that any normal RV is within k standard deviations of itsmean doesn’t depend on the mean or variance.

ISYE 6739 — Goldsman 8/5/20 64 / 108

Standard Normal Distribution

Remarks: The following results are easy to derive, usually via symmetryarguments.

P (Z ≤ a) = Φ(a)P (Z ≥ b) = 1− Φ(b)P (a ≤ Z ≤ b) = Φ(b)− Φ(a)

Φ(0) = 1/2Φ(−b) = P (Z ≤ −b) = P (Z ≥ b) = 1− Φ(b)

P (−b ≤ Z ≤ b) = Φ(b)− Φ(−b) = 2Φ(b)− 1

Then

P (µ− kσ ≤ X ≤ µ+ kσ) = P (−k ≤ Z ≤ k) = 2Φ(k)− 1.

So the probability that any normal RV is within k standard deviations of itsmean doesn’t depend on the mean or variance.

ISYE 6739 — Goldsman 8/5/20 64 / 108

Standard Normal Distribution

Remarks: The following results are easy to derive, usually via symmetryarguments.

P (Z ≤ a) = Φ(a)P (Z ≥ b) = 1− Φ(b)P (a ≤ Z ≤ b) = Φ(b)− Φ(a)

Φ(0) = 1/2Φ(−b) = P (Z ≤ −b) = P (Z ≥ b) = 1− Φ(b)P (−b ≤ Z ≤ b) = Φ(b)− Φ(−b) = 2Φ(b)− 1

Then

P (µ− kσ ≤ X ≤ µ+ kσ) = P (−k ≤ Z ≤ k) = 2Φ(k)− 1.

So the probability that any normal RV is within k standard deviations of itsmean doesn’t depend on the mean or variance.

ISYE 6739 — Goldsman 8/5/20 64 / 108

Standard Normal Distribution

Remarks: The following results are easy to derive, usually via symmetryarguments.

P (Z ≤ a) = Φ(a)P (Z ≥ b) = 1− Φ(b)P (a ≤ Z ≤ b) = Φ(b)− Φ(a)

Φ(0) = 1/2Φ(−b) = P (Z ≤ −b) = P (Z ≥ b) = 1− Φ(b)P (−b ≤ Z ≤ b) = Φ(b)− Φ(−b) = 2Φ(b)− 1

Then

P (µ− kσ ≤ X ≤ µ+ kσ)

= P (−k ≤ Z ≤ k) = 2Φ(k)− 1.

So the probability that any normal RV is within k standard deviations of itsmean doesn’t depend on the mean or variance.

ISYE 6739 — Goldsman 8/5/20 64 / 108

Standard Normal Distribution

Remarks: The following results are easy to derive, usually via symmetryarguments.

P (Z ≤ a) = Φ(a)P (Z ≥ b) = 1− Φ(b)P (a ≤ Z ≤ b) = Φ(b)− Φ(a)

Φ(0) = 1/2Φ(−b) = P (Z ≤ −b) = P (Z ≥ b) = 1− Φ(b)P (−b ≤ Z ≤ b) = Φ(b)− Φ(−b) = 2Φ(b)− 1

Then

P (µ− kσ ≤ X ≤ µ+ kσ) = P (−k ≤ Z ≤ k)

= 2Φ(k)− 1.

So the probability that any normal RV is within k standard deviations of itsmean doesn’t depend on the mean or variance.

ISYE 6739 — Goldsman 8/5/20 64 / 108

Standard Normal Distribution

Remarks: The following results are easy to derive, usually via symmetryarguments.

P (Z ≤ a) = Φ(a)P (Z ≥ b) = 1− Φ(b)P (a ≤ Z ≤ b) = Φ(b)− Φ(a)

Φ(0) = 1/2Φ(−b) = P (Z ≤ −b) = P (Z ≥ b) = 1− Φ(b)P (−b ≤ Z ≤ b) = Φ(b)− Φ(−b) = 2Φ(b)− 1

Then

P (µ− kσ ≤ X ≤ µ+ kσ) = P (−k ≤ Z ≤ k) = 2Φ(k)− 1.

So the probability that any normal RV is within k standard deviations of itsmean doesn’t depend on the mean or variance.

ISYE 6739 — Goldsman 8/5/20 64 / 108

Standard Normal Distribution

Remarks: The following results are easy to derive, usually via symmetryarguments.

P (Z ≤ a) = Φ(a)P (Z ≥ b) = 1− Φ(b)P (a ≤ Z ≤ b) = Φ(b)− Φ(a)

Φ(0) = 1/2Φ(−b) = P (Z ≤ −b) = P (Z ≥ b) = 1− Φ(b)P (−b ≤ Z ≤ b) = Φ(b)− Φ(−b) = 2Φ(b)− 1

Then

P (µ− kσ ≤ X ≤ µ+ kσ) = P (−k ≤ Z ≤ k) = 2Φ(k)− 1.

So the probability that any normal RV is within k standard deviations of itsmean doesn’t depend on the mean or variance.

ISYE 6739 — Goldsman 8/5/20 64 / 108

Standard Normal Distribution

Famous Nor(0, 1) table values. You can memorize these.

Or you can usesoftware calls, like NORMDIST in Excel (which calculates the cdf for anynormal distribution.)

z Φ(z) = P (Z ≤ z)0.00 0.5000

1.00 0.8413

1.28 0.8997 ≈ 0.90

1.645 0.9500

1.96 0.9750

2.33 0.9901 ≈ 0.99

3.00 0.9987

4.00 ≈ 1.0000

ISYE 6739 — Goldsman 8/5/20 65 / 108

Standard Normal Distribution

Famous Nor(0, 1) table values. You can memorize these. Or you can usesoftware calls, like NORMDIST in Excel (which calculates the cdf for anynormal distribution.)

z Φ(z) = P (Z ≤ z)0.00 0.5000

1.00 0.8413

1.28 0.8997 ≈ 0.90

1.645 0.9500

1.96 0.9750

2.33 0.9901 ≈ 0.99

3.00 0.9987

4.00 ≈ 1.0000

ISYE 6739 — Goldsman 8/5/20 65 / 108

Standard Normal Distribution

Famous Nor(0, 1) table values. You can memorize these. Or you can usesoftware calls, like NORMDIST in Excel (which calculates the cdf for anynormal distribution.)

z Φ(z) = P (Z ≤ z)

0.00 0.5000

1.00 0.8413

1.28 0.8997 ≈ 0.90

1.645 0.9500

1.96 0.9750

2.33 0.9901 ≈ 0.99

3.00 0.9987

4.00 ≈ 1.0000

ISYE 6739 — Goldsman 8/5/20 65 / 108

Standard Normal Distribution

Famous Nor(0, 1) table values. You can memorize these. Or you can usesoftware calls, like NORMDIST in Excel (which calculates the cdf for anynormal distribution.)

z Φ(z) = P (Z ≤ z)0.00 0.5000

1.00 0.8413

1.28 0.8997 ≈ 0.90

1.645 0.9500

1.96 0.9750

2.33 0.9901 ≈ 0.99

3.00 0.9987

4.00 ≈ 1.0000

ISYE 6739 — Goldsman 8/5/20 65 / 108

Standard Normal Distribution

Famous Nor(0, 1) table values. You can memorize these. Or you can usesoftware calls, like NORMDIST in Excel (which calculates the cdf for anynormal distribution.)

z Φ(z) = P (Z ≤ z)0.00 0.5000

1.00 0.8413

1.28 0.8997 ≈ 0.90

1.645 0.9500

1.96 0.9750

2.33 0.9901 ≈ 0.99

3.00 0.9987

4.00 ≈ 1.0000

ISYE 6739 — Goldsman 8/5/20 65 / 108

Standard Normal Distribution

Famous Nor(0, 1) table values. You can memorize these. Or you can usesoftware calls, like NORMDIST in Excel (which calculates the cdf for anynormal distribution.)

z Φ(z) = P (Z ≤ z)0.00 0.5000

1.00 0.8413

1.28 0.8997 ≈ 0.90

1.645 0.9500

1.96 0.9750

2.33 0.9901 ≈ 0.99

3.00 0.9987

4.00 ≈ 1.0000

ISYE 6739 — Goldsman 8/5/20 65 / 108

Standard Normal Distribution

Famous Nor(0, 1) table values. You can memorize these. Or you can usesoftware calls, like NORMDIST in Excel (which calculates the cdf for anynormal distribution.)

z Φ(z) = P (Z ≤ z)0.00 0.5000

1.00 0.8413

1.28 0.8997 ≈ 0.90

1.645 0.9500

1.96 0.9750

2.33 0.9901 ≈ 0.99

3.00 0.9987

4.00 ≈ 1.0000

ISYE 6739 — Goldsman 8/5/20 65 / 108

Standard Normal Distribution

Famous Nor(0, 1) table values. You can memorize these. Or you can usesoftware calls, like NORMDIST in Excel (which calculates the cdf for anynormal distribution.)

z Φ(z) = P (Z ≤ z)0.00 0.5000

1.00 0.8413

1.28 0.8997 ≈ 0.90

1.645 0.9500

1.96 0.9750

2.33 0.9901 ≈ 0.99

3.00 0.9987

4.00 ≈ 1.0000

ISYE 6739 — Goldsman 8/5/20 65 / 108

Standard Normal Distribution

Famous Nor(0, 1) table values. You can memorize these. Or you can usesoftware calls, like NORMDIST in Excel (which calculates the cdf for anynormal distribution.)

z Φ(z) = P (Z ≤ z)0.00 0.5000

1.00 0.8413

1.28 0.8997 ≈ 0.90

1.645 0.9500

1.96 0.9750

2.33 0.9901 ≈ 0.99

3.00 0.9987

4.00 ≈ 1.0000

ISYE 6739 — Goldsman 8/5/20 65 / 108

Standard Normal Distribution

Famous Nor(0, 1) table values. You can memorize these. Or you can usesoftware calls, like NORMDIST in Excel (which calculates the cdf for anynormal distribution.)

z Φ(z) = P (Z ≤ z)0.00 0.5000

1.00 0.8413

1.28 0.8997 ≈ 0.90

1.645 0.9500

1.96 0.9750

2.33 0.9901 ≈ 0.99

3.00 0.9987

4.00 ≈ 1.0000

ISYE 6739 — Goldsman 8/5/20 65 / 108

Standard Normal Distribution

Famous Nor(0, 1) table values. You can memorize these. Or you can usesoftware calls, like NORMDIST in Excel (which calculates the cdf for anynormal distribution.)

z Φ(z) = P (Z ≤ z)0.00 0.5000

1.00 0.8413

1.28 0.8997 ≈ 0.90

1.645 0.9500

1.96 0.9750

2.33 0.9901 ≈ 0.99

3.00 0.9987

4.00 ≈ 1.0000

ISYE 6739 — Goldsman 8/5/20 65 / 108

Standard Normal Distribution

By the earlier “Fun Facts” and then the discussion on the last two pages, theprobability that any normal RV is within k standard deviations of its mean is

P (µ− kσ ≤ X ≤ µ+ kσ) = 2Φ(k)− 1.

For k = 1, this probability is 2(0.8413)− 1 = 0.6827.

There is a 95% chance that a normal observation will be within 2 s.d.’s of itsmean.

99.7% of all observations are within 3 standard deviations of the mean!

Finally, note that Toyota’s six-sigma corresponds to 2Φ(6)− 1.= 1.0000.

ISYE 6739 — Goldsman 8/5/20 66 / 108

Standard Normal Distribution

By the earlier “Fun Facts” and then the discussion on the last two pages, theprobability that any normal RV is within k standard deviations of its mean is

P (µ− kσ ≤ X ≤ µ+ kσ) = 2Φ(k)− 1.

For k = 1, this probability is 2(0.8413)− 1 = 0.6827.

There is a 95% chance that a normal observation will be within 2 s.d.’s of itsmean.

99.7% of all observations are within 3 standard deviations of the mean!

Finally, note that Toyota’s six-sigma corresponds to 2Φ(6)− 1.= 1.0000.

ISYE 6739 — Goldsman 8/5/20 66 / 108

Standard Normal Distribution

By the earlier “Fun Facts” and then the discussion on the last two pages, theprobability that any normal RV is within k standard deviations of its mean is

P (µ− kσ ≤ X ≤ µ+ kσ) = 2Φ(k)− 1.

For k = 1, this probability is 2(0.8413)− 1 = 0.6827.

There is a 95% chance that a normal observation will be within 2 s.d.’s of itsmean.

99.7% of all observations are within 3 standard deviations of the mean!

Finally, note that Toyota’s six-sigma corresponds to 2Φ(6)− 1.= 1.0000.

ISYE 6739 — Goldsman 8/5/20 66 / 108

Standard Normal Distribution

By the earlier “Fun Facts” and then the discussion on the last two pages, theprobability that any normal RV is within k standard deviations of its mean is

P (µ− kσ ≤ X ≤ µ+ kσ) = 2Φ(k)− 1.

For k = 1, this probability is 2(0.8413)− 1 = 0.6827.

There is a 95% chance that a normal observation will be within 2 s.d.’s of itsmean.

99.7% of all observations are within 3 standard deviations of the mean!

Finally, note that Toyota’s six-sigma corresponds to 2Φ(6)− 1.= 1.0000.

ISYE 6739 — Goldsman 8/5/20 66 / 108

Standard Normal Distribution

By the earlier “Fun Facts” and then the discussion on the last two pages, theprobability that any normal RV is within k standard deviations of its mean is

P (µ− kσ ≤ X ≤ µ+ kσ) = 2Φ(k)− 1.

For k = 1, this probability is 2(0.8413)− 1 = 0.6827.

There is a 95% chance that a normal observation will be within 2 s.d.’s of itsmean.

99.7% of all observations are within 3 standard deviations of the mean!

Finally, note that Toyota’s six-sigma corresponds to 2Φ(6)− 1.= 1.0000.

ISYE 6739 — Goldsman 8/5/20 66 / 108

Standard Normal Distribution

By the earlier “Fun Facts” and then the discussion on the last two pages, theprobability that any normal RV is within k standard deviations of its mean is

P (µ− kσ ≤ X ≤ µ+ kσ) = 2Φ(k)− 1.

For k = 1, this probability is 2(0.8413)− 1 = 0.6827.

There is a 95% chance that a normal observation will be within 2 s.d.’s of itsmean.

99.7% of all observations are within 3 standard deviations of the mean!

Finally, note that Toyota’s six-sigma corresponds to 2Φ(6)− 1.= 1.0000.

ISYE 6739 — Goldsman 8/5/20 66 / 108

Standard Normal Distribution

Famous Inverse Nor(0, 1) table values. Can also use software, such asExcel’s NORMINV function, which actually calculates inverses for any normaldistribution, not just standard normal.

Φ−1(p) is the value of z such that Φ(z) = p. Φ−1(p) is called the pthquantile of Z.

p Φ−1(p)

0.90 1.28

0.95 1.645

0.975 1.96

0.99 2.33

0.995 2.58

ISYE 6739 — Goldsman 8/5/20 67 / 108

Standard Normal Distribution

Famous Inverse Nor(0, 1) table values. Can also use software, such asExcel’s NORMINV function, which actually calculates inverses for any normaldistribution, not just standard normal.

Φ−1(p) is the value of z such that Φ(z) = p. Φ−1(p) is called the pthquantile of Z.

p Φ−1(p)

0.90 1.28

0.95 1.645

0.975 1.96

0.99 2.33

0.995 2.58

ISYE 6739 — Goldsman 8/5/20 67 / 108

Standard Normal Distribution

Famous Inverse Nor(0, 1) table values. Can also use software, such asExcel’s NORMINV function, which actually calculates inverses for any normaldistribution, not just standard normal.

Φ−1(p) is the value of z such that Φ(z) = p. Φ−1(p) is called the pthquantile of Z.

p Φ−1(p)

0.90 1.28

0.95 1.645

0.975 1.96

0.99 2.33

0.995 2.58

ISYE 6739 — Goldsman 8/5/20 67 / 108

Standard Normal Distribution

Famous Inverse Nor(0, 1) table values. Can also use software, such asExcel’s NORMINV function, which actually calculates inverses for any normaldistribution, not just standard normal.

Φ−1(p) is the value of z such that Φ(z) = p. Φ−1(p) is called the pthquantile of Z.

p Φ−1(p)

0.90 1.28

0.95 1.645

0.975 1.96

0.99 2.33

0.995 2.58

ISYE 6739 — Goldsman 8/5/20 67 / 108

Standard Normal Distribution

Famous Inverse Nor(0, 1) table values. Can also use software, such asExcel’s NORMINV function, which actually calculates inverses for any normaldistribution, not just standard normal.

Φ−1(p) is the value of z such that Φ(z) = p. Φ−1(p) is called the pthquantile of Z.

p Φ−1(p)

0.90 1.28

0.95 1.645

0.975 1.96

0.99 2.33

0.995 2.58

ISYE 6739 — Goldsman 8/5/20 67 / 108

Standard Normal Distribution

Famous Inverse Nor(0, 1) table values. Can also use software, such asExcel’s NORMINV function, which actually calculates inverses for any normaldistribution, not just standard normal.

Φ−1(p) is the value of z such that Φ(z) = p. Φ−1(p) is called the pthquantile of Z.

p Φ−1(p)

0.90 1.28

0.95 1.645

0.975 1.96

0.99 2.33

0.995 2.58

ISYE 6739 — Goldsman 8/5/20 67 / 108

Standard Normal Distribution

Famous Inverse Nor(0, 1) table values. Can also use software, such asExcel’s NORMINV function, which actually calculates inverses for any normaldistribution, not just standard normal.

Φ−1(p) is the value of z such that Φ(z) = p. Φ−1(p) is called the pthquantile of Z.

p Φ−1(p)

0.90 1.28

0.95 1.645

0.975 1.96

0.99 2.33

0.995 2.58

ISYE 6739 — Goldsman 8/5/20 67 / 108

Standard Normal Distribution

Famous Inverse Nor(0, 1) table values. Can also use software, such asExcel’s NORMINV function, which actually calculates inverses for any normaldistribution, not just standard normal.

Φ−1(p) is the value of z such that Φ(z) = p. Φ−1(p) is called the pthquantile of Z.

p Φ−1(p)

0.90 1.28

0.95 1.645

0.975 1.96

0.99 2.33

0.995 2.58

ISYE 6739 — Goldsman 8/5/20 67 / 108

Standard Normal Distribution

Example: X ∼ Nor(21, 4). Find P (19 < X < 22.5). Standardizing, weget

P (19 < X < 22.5)

= P

(19− µσ

<X − µσ

<22.5− µ

σ

)= P

(19− 21

2< Z <

22.5− 21

2

)= P (−1 < Z < 0.75)

= Φ(0.75)− Φ(−1)

= Φ(0.75)− [1− Φ(1)]

= 0.7734− [1− 0.8413] = 0.6147. 2

ISYE 6739 — Goldsman 8/5/20 68 / 108

Standard Normal Distribution

Example: X ∼ Nor(21, 4). Find P (19 < X < 22.5). Standardizing, weget

P (19 < X < 22.5)

= P

(19− µσ

<X − µσ

<22.5− µ

σ

)

= P

(19− 21

2< Z <

22.5− 21

2

)= P (−1 < Z < 0.75)

= Φ(0.75)− Φ(−1)

= Φ(0.75)− [1− Φ(1)]

= 0.7734− [1− 0.8413] = 0.6147. 2

ISYE 6739 — Goldsman 8/5/20 68 / 108

Standard Normal Distribution

Example: X ∼ Nor(21, 4). Find P (19 < X < 22.5). Standardizing, weget

P (19 < X < 22.5)

= P

(19− µσ

<X − µσ

<22.5− µ

σ

)= P

(19− 21

2< Z <

22.5− 21

2

)

= P (−1 < Z < 0.75)

= Φ(0.75)− Φ(−1)

= Φ(0.75)− [1− Φ(1)]

= 0.7734− [1− 0.8413] = 0.6147. 2

ISYE 6739 — Goldsman 8/5/20 68 / 108

Standard Normal Distribution

Example: X ∼ Nor(21, 4). Find P (19 < X < 22.5). Standardizing, weget

P (19 < X < 22.5)

= P

(19− µσ

<X − µσ

<22.5− µ

σ

)= P

(19− 21

2< Z <

22.5− 21

2

)= P (−1 < Z < 0.75)

= Φ(0.75)− Φ(−1)

= Φ(0.75)− [1− Φ(1)]

= 0.7734− [1− 0.8413] = 0.6147. 2

ISYE 6739 — Goldsman 8/5/20 68 / 108

Standard Normal Distribution

Example: X ∼ Nor(21, 4). Find P (19 < X < 22.5). Standardizing, weget

P (19 < X < 22.5)

= P

(19− µσ

<X − µσ

<22.5− µ

σ

)= P

(19− 21

2< Z <

22.5− 21

2

)= P (−1 < Z < 0.75)

= Φ(0.75)− Φ(−1)

= Φ(0.75)− [1− Φ(1)]

= 0.7734− [1− 0.8413] = 0.6147. 2

ISYE 6739 — Goldsman 8/5/20 68 / 108

Standard Normal Distribution

Example: X ∼ Nor(21, 4). Find P (19 < X < 22.5). Standardizing, weget

P (19 < X < 22.5)

= P

(19− µσ

<X − µσ

<22.5− µ

σ

)= P

(19− 21

2< Z <

22.5− 21

2

)= P (−1 < Z < 0.75)

= Φ(0.75)− Φ(−1)

= Φ(0.75)− [1− Φ(1)]

= 0.7734− [1− 0.8413] = 0.6147. 2

ISYE 6739 — Goldsman 8/5/20 68 / 108

Standard Normal Distribution

Example: X ∼ Nor(21, 4). Find P (19 < X < 22.5). Standardizing, weget

P (19 < X < 22.5)

= P

(19− µσ

<X − µσ

<22.5− µ

σ

)= P

(19− 21

2< Z <

22.5− 21

2

)= P (−1 < Z < 0.75)

= Φ(0.75)− Φ(−1)

= Φ(0.75)− [1− Φ(1)]

= 0.7734− [1− 0.8413] = 0.6147. 2

ISYE 6739 — Goldsman 8/5/20 68 / 108

Standard Normal Distribution

Example: Suppose that heights of men are M ∼ Nor(68, 4) and heights ofwomen are W ∼ Nor(65, 1).

Select a man and woman independently at random.

Find the probability that the woman is taller than the man.

Answer: Note that

W −M ∼ Nor(E[W −M ],Var(W −M))

∼ Nor(65− 68, 1 + 4) ∼ Nor(−3, 5).

Then

P (W > M) = P (W −M > 0)

= P

(Z >

0 + 3√5

)= 1− Φ(3/

√5)

.= 1− 0.910 = 0.090. 2

ISYE 6739 — Goldsman 8/5/20 69 / 108

Standard Normal Distribution

Example: Suppose that heights of men are M ∼ Nor(68, 4) and heights ofwomen are W ∼ Nor(65, 1).

Select a man and woman independently at random.

Find the probability that the woman is taller than the man.

Answer: Note that

W −M ∼ Nor(E[W −M ],Var(W −M))

∼ Nor(65− 68, 1 + 4) ∼ Nor(−3, 5).

Then

P (W > M) = P (W −M > 0)

= P

(Z >

0 + 3√5

)= 1− Φ(3/

√5)

.= 1− 0.910 = 0.090. 2

ISYE 6739 — Goldsman 8/5/20 69 / 108

Standard Normal Distribution

Example: Suppose that heights of men are M ∼ Nor(68, 4) and heights ofwomen are W ∼ Nor(65, 1).

Select a man and woman independently at random.

Find the probability that the woman is taller than the man.

Answer: Note that

W −M ∼ Nor(E[W −M ],Var(W −M))

∼ Nor(65− 68, 1 + 4) ∼ Nor(−3, 5).

Then

P (W > M) = P (W −M > 0)

= P

(Z >

0 + 3√5

)= 1− Φ(3/

√5)

.= 1− 0.910 = 0.090. 2

ISYE 6739 — Goldsman 8/5/20 69 / 108

Standard Normal Distribution

Example: Suppose that heights of men are M ∼ Nor(68, 4) and heights ofwomen are W ∼ Nor(65, 1).

Select a man and woman independently at random.

Find the probability that the woman is taller than the man.

Answer: Note that

W −M ∼ Nor(E[W −M ],Var(W −M))

∼ Nor(65− 68, 1 + 4) ∼ Nor(−3, 5).

Then

P (W > M) = P (W −M > 0)

= P

(Z >

0 + 3√5

)= 1− Φ(3/

√5)

.= 1− 0.910 = 0.090. 2

ISYE 6739 — Goldsman 8/5/20 69 / 108

Standard Normal Distribution

Example: Suppose that heights of men are M ∼ Nor(68, 4) and heights ofwomen are W ∼ Nor(65, 1).

Select a man and woman independently at random.

Find the probability that the woman is taller than the man.

Answer: Note that

W −M ∼ Nor(E[W −M ],Var(W −M))

∼ Nor(65− 68, 1 + 4) ∼ Nor(−3, 5).

Then

P (W > M) = P (W −M > 0)

= P

(Z >

0 + 3√5

)= 1− Φ(3/

√5)

.= 1− 0.910 = 0.090. 2

ISYE 6739 — Goldsman 8/5/20 69 / 108

Standard Normal Distribution

Example: Suppose that heights of men are M ∼ Nor(68, 4) and heights ofwomen are W ∼ Nor(65, 1).

Select a man and woman independently at random.

Find the probability that the woman is taller than the man.

Answer: Note that

W −M ∼ Nor(E[W −M ],Var(W −M))

∼ Nor(65− 68, 1 + 4) ∼ Nor(−3, 5).

Then

P (W > M) = P (W −M > 0)

= P

(Z >

0 + 3√5

)= 1− Φ(3/

√5)

.= 1− 0.910 = 0.090. 2

ISYE 6739 — Goldsman 8/5/20 69 / 108

Standard Normal Distribution

Example: Suppose that heights of men are M ∼ Nor(68, 4) and heights ofwomen are W ∼ Nor(65, 1).

Select a man and woman independently at random.

Find the probability that the woman is taller than the man.

Answer: Note that

W −M ∼ Nor(E[W −M ],Var(W −M))

∼ Nor(65− 68, 1 + 4) ∼ Nor(−3, 5).

Then

P (W > M) = P (W −M > 0)

= P

(Z >

0 + 3√5

)

= 1− Φ(3/√

5).= 1− 0.910 = 0.090. 2

ISYE 6739 — Goldsman 8/5/20 69 / 108

Standard Normal Distribution

Example: Suppose that heights of men are M ∼ Nor(68, 4) and heights ofwomen are W ∼ Nor(65, 1).

Select a man and woman independently at random.

Find the probability that the woman is taller than the man.

Answer: Note that

W −M ∼ Nor(E[W −M ],Var(W −M))

∼ Nor(65− 68, 1 + 4) ∼ Nor(−3, 5).

Then

P (W > M) = P (W −M > 0)

= P

(Z >

0 + 3√5

)= 1− Φ(3/

√5)

.= 1− 0.910 = 0.090. 2

ISYE 6739 — Goldsman 8/5/20 69 / 108

Standard Normal Distribution

Example: Suppose that heights of men are M ∼ Nor(68, 4) and heights ofwomen are W ∼ Nor(65, 1).

Select a man and woman independently at random.

Find the probability that the woman is taller than the man.

Answer: Note that

W −M ∼ Nor(E[W −M ],Var(W −M))

∼ Nor(65− 68, 1 + 4) ∼ Nor(−3, 5).

Then

P (W > M) = P (W −M > 0)

= P

(Z >

0 + 3√5

)= 1− Φ(3/

√5)

.= 1− 0.910 = 0.090. 2

ISYE 6739 — Goldsman 8/5/20 69 / 108

Sample Mean of Normals

1 Bernoulli and Binomial Distributions

2 Hypergeometric Distribution3 Geometric and Negative Binomial Distributions

4 Poisson Distribution5 Uniform, Exponential, and Friends6 Other Continuous Distributions7 Normal Distribution: Basics8 Standard Normal Distribution9 Sample Mean of Normals

10 The Central Limit Theorem + Proof

11 Central Limit Theorem Examples

12 Extensions — Multivariate Normal Distribution13 Extensions — Lognormal Distribution

14 Computer Stuff

ISYE 6739 — Goldsman 8/5/20 70 / 108

Sample Mean of Normals

Lesson 4.9 — Sample Mean of Normals

The sample mean of X1, . . . , Xn is X̄ ≡∑n

i=1Xi/n.

Corollary (of old theorem):X1, . . . , Xn

iid∼ Nor(µ, σ2)⇒ X̄ ∼ Nor(µ, σ2/n).

Proof: By previous work, as long as X1, . . . , Xn are iid something, we haveE[X̄] = µ and Var(X̄) = σ2/n. Since X̄ is a linear combination ofindependent normals, it’s also normal. Done. 2

Remark: This result is very significant! As the number of observationsincreases, Var(X̄) gets smaller (while E[X̄] remains constant). In fact, it’scalled the Law of Large Numbers.

In the upcoming statistics portion of the course, we’ll learn that this makes X̄an excellent estimator for the mean µ, which is typically unknown inpractice.

ISYE 6739 — Goldsman 8/5/20 71 / 108

Sample Mean of Normals

Lesson 4.9 — Sample Mean of Normals

The sample mean of X1, . . . , Xn is X̄ ≡∑n

i=1Xi/n.

Corollary (of old theorem):X1, . . . , Xn

iid∼ Nor(µ, σ2)⇒ X̄ ∼ Nor(µ, σ2/n).

Proof: By previous work, as long as X1, . . . , Xn are iid something, we haveE[X̄] = µ and Var(X̄) = σ2/n. Since X̄ is a linear combination ofindependent normals, it’s also normal. Done. 2

Remark: This result is very significant! As the number of observationsincreases, Var(X̄) gets smaller (while E[X̄] remains constant). In fact, it’scalled the Law of Large Numbers.

In the upcoming statistics portion of the course, we’ll learn that this makes X̄an excellent estimator for the mean µ, which is typically unknown inpractice.

ISYE 6739 — Goldsman 8/5/20 71 / 108

Sample Mean of Normals

Lesson 4.9 — Sample Mean of Normals

The sample mean of X1, . . . , Xn is X̄ ≡∑n

i=1Xi/n.

Corollary (of old theorem):X1, . . . , Xn

iid∼ Nor(µ, σ2)⇒ X̄ ∼ Nor(µ, σ2/n).

Proof: By previous work, as long as X1, . . . , Xn are iid something, we haveE[X̄] = µ and Var(X̄) = σ2/n. Since X̄ is a linear combination ofindependent normals, it’s also normal. Done. 2

Remark: This result is very significant! As the number of observationsincreases, Var(X̄) gets smaller (while E[X̄] remains constant). In fact, it’scalled the Law of Large Numbers.

In the upcoming statistics portion of the course, we’ll learn that this makes X̄an excellent estimator for the mean µ, which is typically unknown inpractice.

ISYE 6739 — Goldsman 8/5/20 71 / 108

Sample Mean of Normals

Lesson 4.9 — Sample Mean of Normals

The sample mean of X1, . . . , Xn is X̄ ≡∑n

i=1Xi/n.

Corollary (of old theorem):X1, . . . , Xn

iid∼ Nor(µ, σ2)⇒ X̄ ∼ Nor(µ, σ2/n).

Proof: By previous work, as long as X1, . . . , Xn are iid something, we have

E[X̄] = µ and Var(X̄) = σ2/n. Since X̄ is a linear combination ofindependent normals, it’s also normal. Done. 2

Remark: This result is very significant! As the number of observationsincreases, Var(X̄) gets smaller (while E[X̄] remains constant). In fact, it’scalled the Law of Large Numbers.

In the upcoming statistics portion of the course, we’ll learn that this makes X̄an excellent estimator for the mean µ, which is typically unknown inpractice.

ISYE 6739 — Goldsman 8/5/20 71 / 108

Sample Mean of Normals

Lesson 4.9 — Sample Mean of Normals

The sample mean of X1, . . . , Xn is X̄ ≡∑n

i=1Xi/n.

Corollary (of old theorem):X1, . . . , Xn

iid∼ Nor(µ, σ2)⇒ X̄ ∼ Nor(µ, σ2/n).

Proof: By previous work, as long as X1, . . . , Xn are iid something, we haveE[X̄] = µ and Var(X̄) = σ2/n.

Since X̄ is a linear combination ofindependent normals, it’s also normal. Done. 2

Remark: This result is very significant! As the number of observationsincreases, Var(X̄) gets smaller (while E[X̄] remains constant). In fact, it’scalled the Law of Large Numbers.

In the upcoming statistics portion of the course, we’ll learn that this makes X̄an excellent estimator for the mean µ, which is typically unknown inpractice.

ISYE 6739 — Goldsman 8/5/20 71 / 108

Sample Mean of Normals

Lesson 4.9 — Sample Mean of Normals

The sample mean of X1, . . . , Xn is X̄ ≡∑n

i=1Xi/n.

Corollary (of old theorem):X1, . . . , Xn

iid∼ Nor(µ, σ2)⇒ X̄ ∼ Nor(µ, σ2/n).

Proof: By previous work, as long as X1, . . . , Xn are iid something, we haveE[X̄] = µ and Var(X̄) = σ2/n. Since X̄ is a linear combination ofindependent normals, it’s also normal. Done. 2

Remark: This result is very significant! As the number of observationsincreases, Var(X̄) gets smaller (while E[X̄] remains constant). In fact, it’scalled the Law of Large Numbers.

In the upcoming statistics portion of the course, we’ll learn that this makes X̄an excellent estimator for the mean µ, which is typically unknown inpractice.

ISYE 6739 — Goldsman 8/5/20 71 / 108

Sample Mean of Normals

Lesson 4.9 — Sample Mean of Normals

The sample mean of X1, . . . , Xn is X̄ ≡∑n

i=1Xi/n.

Corollary (of old theorem):X1, . . . , Xn

iid∼ Nor(µ, σ2)⇒ X̄ ∼ Nor(µ, σ2/n).

Proof: By previous work, as long as X1, . . . , Xn are iid something, we haveE[X̄] = µ and Var(X̄) = σ2/n. Since X̄ is a linear combination ofindependent normals, it’s also normal. Done. 2

Remark: This result is very significant! As the number of observationsincreases, Var(X̄) gets smaller (while E[X̄] remains constant). In fact, it’scalled the Law of Large Numbers.

In the upcoming statistics portion of the course, we’ll learn that this makes X̄an excellent estimator for the mean µ, which is typically unknown inpractice.

ISYE 6739 — Goldsman 8/5/20 71 / 108

Sample Mean of Normals

Lesson 4.9 — Sample Mean of Normals

The sample mean of X1, . . . , Xn is X̄ ≡∑n

i=1Xi/n.

Corollary (of old theorem):X1, . . . , Xn

iid∼ Nor(µ, σ2)⇒ X̄ ∼ Nor(µ, σ2/n).

Proof: By previous work, as long as X1, . . . , Xn are iid something, we haveE[X̄] = µ and Var(X̄) = σ2/n. Since X̄ is a linear combination ofindependent normals, it’s also normal. Done. 2

Remark: This result is very significant! As the number of observationsincreases, Var(X̄) gets smaller (while E[X̄] remains constant). In fact, it’scalled the Law of Large Numbers.

In the upcoming statistics portion of the course, we’ll learn that this makes X̄an excellent estimator for the mean µ, which is typically unknown inpractice.

ISYE 6739 — Goldsman 8/5/20 71 / 108

Sample Mean of Normals

Example: Suppose that X1, . . . , Xniid∼ Nor(µ, 16). Find the sample size n

such thatP (|X̄ − µ| ≤ 1) ≥ 0.95.

How many observations should you take so that X̄ will have a good chance ofbeing close to µ?

Solution: Note that X̄ ∼ Nor(µ, 16/n). Then

P (|X̄ − µ| ≤ 1) = P (−1 ≤ X̄ − µ ≤ 1)

= P

(−1

4/√n≤ X̄ − µ

4/√n≤ 1

4/√n

)= P

(−√n

4≤ Z ≤

√n

4

)= 2Φ(

√n/4)− 1.

ISYE 6739 — Goldsman 8/5/20 72 / 108

Sample Mean of Normals

Example: Suppose that X1, . . . , Xniid∼ Nor(µ, 16). Find the sample size n

such thatP (|X̄ − µ| ≤ 1) ≥ 0.95.

How many observations should you take so that X̄ will have a good chance ofbeing close to µ?

Solution: Note that X̄ ∼ Nor(µ, 16/n). Then

P (|X̄ − µ| ≤ 1) = P (−1 ≤ X̄ − µ ≤ 1)

= P

(−1

4/√n≤ X̄ − µ

4/√n≤ 1

4/√n

)= P

(−√n

4≤ Z ≤

√n

4

)= 2Φ(

√n/4)− 1.

ISYE 6739 — Goldsman 8/5/20 72 / 108

Sample Mean of Normals

Example: Suppose that X1, . . . , Xniid∼ Nor(µ, 16). Find the sample size n

such thatP (|X̄ − µ| ≤ 1) ≥ 0.95.

How many observations should you take so that X̄ will have a good chance ofbeing close to µ?

Solution: Note that X̄ ∼ Nor(µ, 16/n). Then

P (|X̄ − µ| ≤ 1) = P (−1 ≤ X̄ − µ ≤ 1)

= P

(−1

4/√n≤ X̄ − µ

4/√n≤ 1

4/√n

)= P

(−√n

4≤ Z ≤

√n

4

)= 2Φ(

√n/4)− 1.

ISYE 6739 — Goldsman 8/5/20 72 / 108

Sample Mean of Normals

Example: Suppose that X1, . . . , Xniid∼ Nor(µ, 16). Find the sample size n

such thatP (|X̄ − µ| ≤ 1) ≥ 0.95.

How many observations should you take so that X̄ will have a good chance ofbeing close to µ?

Solution: Note that X̄ ∼ Nor(µ, 16/n). Then

P (|X̄ − µ| ≤ 1) = P (−1 ≤ X̄ − µ ≤ 1)

= P

(−1

4/√n≤ X̄ − µ

4/√n≤ 1

4/√n

)= P

(−√n

4≤ Z ≤

√n

4

)= 2Φ(

√n/4)− 1.

ISYE 6739 — Goldsman 8/5/20 72 / 108

Sample Mean of Normals

Example: Suppose that X1, . . . , Xniid∼ Nor(µ, 16). Find the sample size n

such thatP (|X̄ − µ| ≤ 1) ≥ 0.95.

How many observations should you take so that X̄ will have a good chance ofbeing close to µ?

Solution: Note that X̄ ∼ Nor(µ, 16/n). Then

P (|X̄ − µ| ≤ 1) = P (−1 ≤ X̄ − µ ≤ 1)

= P

(−1

4/√n≤ X̄ − µ

4/√n≤ 1

4/√n

)

= P

(−√n

4≤ Z ≤

√n

4

)= 2Φ(

√n/4)− 1.

ISYE 6739 — Goldsman 8/5/20 72 / 108

Sample Mean of Normals

Example: Suppose that X1, . . . , Xniid∼ Nor(µ, 16). Find the sample size n

such thatP (|X̄ − µ| ≤ 1) ≥ 0.95.

How many observations should you take so that X̄ will have a good chance ofbeing close to µ?

Solution: Note that X̄ ∼ Nor(µ, 16/n). Then

P (|X̄ − µ| ≤ 1) = P (−1 ≤ X̄ − µ ≤ 1)

= P

(−1

4/√n≤ X̄ − µ

4/√n≤ 1

4/√n

)= P

(−√n

4≤ Z ≤

√n

4

)

= 2Φ(√n/4)− 1.

ISYE 6739 — Goldsman 8/5/20 72 / 108

Sample Mean of Normals

Example: Suppose that X1, . . . , Xniid∼ Nor(µ, 16). Find the sample size n

such thatP (|X̄ − µ| ≤ 1) ≥ 0.95.

How many observations should you take so that X̄ will have a good chance ofbeing close to µ?

Solution: Note that X̄ ∼ Nor(µ, 16/n). Then

P (|X̄ − µ| ≤ 1) = P (−1 ≤ X̄ − µ ≤ 1)

= P

(−1

4/√n≤ X̄ − µ

4/√n≤ 1

4/√n

)= P

(−√n

4≤ Z ≤

√n

4

)= 2Φ(

√n/4)− 1.

ISYE 6739 — Goldsman 8/5/20 72 / 108

Sample Mean of Normals

Now we have to find n such that this probability is at least 0.95. . . .

2Φ(√n/4)− 1 ≥ 0.95 iff

Φ(√n/4) ≥ 0.975 iff

√n

4≥ Φ−1(0.975) = 1.96

iff n ≥ 61.47 or 62.

So if you take the average of 62 observations, then X̄ has a 95% chance ofbeing within 1 of µ. 2

ISYE 6739 — Goldsman 8/5/20 73 / 108

Sample Mean of Normals

Now we have to find n such that this probability is at least 0.95. . . .

2Φ(√n/4)− 1 ≥ 0.95 iff

Φ(√n/4) ≥ 0.975 iff

√n

4≥ Φ−1(0.975) = 1.96

iff n ≥ 61.47 or 62.

So if you take the average of 62 observations, then X̄ has a 95% chance ofbeing within 1 of µ. 2

ISYE 6739 — Goldsman 8/5/20 73 / 108

Sample Mean of Normals

Now we have to find n such that this probability is at least 0.95. . . .

2Φ(√n/4)− 1 ≥ 0.95 iff

Φ(√n/4) ≥ 0.975 iff

√n

4≥ Φ−1(0.975) = 1.96

iff n ≥ 61.47 or 62.

So if you take the average of 62 observations, then X̄ has a 95% chance ofbeing within 1 of µ. 2

ISYE 6739 — Goldsman 8/5/20 73 / 108

Sample Mean of Normals

Now we have to find n such that this probability is at least 0.95. . . .

2Φ(√n/4)− 1 ≥ 0.95 iff

Φ(√n/4) ≥ 0.975 iff

√n

4≥ Φ−1(0.975) = 1.96

iff n ≥ 61.47 or 62.

So if you take the average of 62 observations, then X̄ has a 95% chance ofbeing within 1 of µ. 2

ISYE 6739 — Goldsman 8/5/20 73 / 108

Sample Mean of Normals

Now we have to find n such that this probability is at least 0.95. . . .

2Φ(√n/4)− 1 ≥ 0.95 iff

Φ(√n/4) ≥ 0.975 iff

√n

4≥ Φ−1(0.975) = 1.96

iff n ≥ 61.47 or 62.

So if you take the average of 62 observations, then X̄ has a 95% chance ofbeing within 1 of µ. 2

ISYE 6739 — Goldsman 8/5/20 73 / 108

Sample Mean of Normals

Now we have to find n such that this probability is at least 0.95. . . .

2Φ(√n/4)− 1 ≥ 0.95 iff

Φ(√n/4) ≥ 0.975 iff

√n

4≥ Φ−1(0.975) = 1.96

iff n ≥ 61.47 or 62.

So if you take the average of 62 observations, then X̄ has a 95% chance ofbeing within 1 of µ. 2

ISYE 6739 — Goldsman 8/5/20 73 / 108

The Central Limit Theorem + Proof

1 Bernoulli and Binomial Distributions

2 Hypergeometric Distribution3 Geometric and Negative Binomial Distributions

4 Poisson Distribution5 Uniform, Exponential, and Friends6 Other Continuous Distributions7 Normal Distribution: Basics8 Standard Normal Distribution9 Sample Mean of Normals

10 The Central Limit Theorem + Proof

11 Central Limit Theorem Examples

12 Extensions — Multivariate Normal Distribution13 Extensions — Lognormal Distribution

14 Computer Stuff

ISYE 6739 — Goldsman 8/5/20 74 / 108

The Central Limit Theorem + Proof

Lesson 4.10 — The Central Limit Theorem + Proof

The Central Limit Theorem is the most-important theorem in probability andstatistics!

CLT: Suppose X1, . . . , Xn are iid with E[Xi] = µ and Var(Xi) = σ2. Thenas n→∞,

Zn =

∑ni=1Xi − nµσ√n

=X̄ − µσ/√n

=X̄ − E[X̄]√

Var(X̄)

d→ Nor(0, 1),

where “ d→ ” means that the cdf of Zn→ the Nor(0, 1) cdf.

ISYE 6739 — Goldsman 8/5/20 75 / 108

The Central Limit Theorem + Proof

Lesson 4.10 — The Central Limit Theorem + Proof

The Central Limit Theorem is the most-important theorem in probability andstatistics!

CLT: Suppose X1, . . . , Xn are iid with E[Xi] = µ and Var(Xi) = σ2. Thenas n→∞,

Zn =

∑ni=1Xi − nµσ√n

=X̄ − µσ/√n

=X̄ − E[X̄]√

Var(X̄)

d→ Nor(0, 1),

where “ d→ ” means that the cdf of Zn→ the Nor(0, 1) cdf.

ISYE 6739 — Goldsman 8/5/20 75 / 108

The Central Limit Theorem + Proof

Lesson 4.10 — The Central Limit Theorem + Proof

The Central Limit Theorem is the most-important theorem in probability andstatistics!

CLT: Suppose X1, . . . , Xn are iid with E[Xi] = µ and Var(Xi) = σ2. Thenas n→∞,

Zn =

∑ni=1Xi − nµσ√n

=X̄ − µσ/√n

=X̄ − E[X̄]√

Var(X̄)

d→ Nor(0, 1),

where “ d→ ” means that the cdf of Zn→ the Nor(0, 1) cdf.

ISYE 6739 — Goldsman 8/5/20 75 / 108

The Central Limit Theorem + Proof

Lesson 4.10 — The Central Limit Theorem + Proof

The Central Limit Theorem is the most-important theorem in probability andstatistics!

CLT: Suppose X1, . . . , Xn are iid with E[Xi] = µ and Var(Xi) = σ2. Thenas n→∞,

Zn =

∑ni=1Xi − nµσ√n

=X̄ − µσ/√n

=X̄ − E[X̄]√

Var(X̄)

d→ Nor(0, 1),

where “ d→ ” means that the cdf of Zn→ the Nor(0, 1) cdf.

ISYE 6739 — Goldsman 8/5/20 75 / 108

The Central Limit Theorem + Proof

Lesson 4.10 — The Central Limit Theorem + Proof

The Central Limit Theorem is the most-important theorem in probability andstatistics!

CLT: Suppose X1, . . . , Xn are iid with E[Xi] = µ and Var(Xi) = σ2. Thenas n→∞,

Zn =

∑ni=1Xi − nµσ√n

=X̄ − µσ/√n

=X̄ − E[X̄]√

Var(X̄)

d→ Nor(0, 1),

where “ d→ ” means that the cdf of Zn→ the Nor(0, 1) cdf.

ISYE 6739 — Goldsman 8/5/20 75 / 108

The Central Limit Theorem + Proof

Lesson 4.10 — The Central Limit Theorem + Proof

The Central Limit Theorem is the most-important theorem in probability andstatistics!

CLT: Suppose X1, . . . , Xn are iid with E[Xi] = µ and Var(Xi) = σ2. Thenas n→∞,

Zn =

∑ni=1Xi − nµσ√n

=X̄ − µσ/√n

=X̄ − E[X̄]√

Var(X̄)

d→ Nor(0, 1),

where “ d→ ” means that the cdf of Zn→ the Nor(0, 1) cdf.

ISYE 6739 — Goldsman 8/5/20 75 / 108

The Central Limit Theorem + Proof

Lesson 4.10 — The Central Limit Theorem + Proof

The Central Limit Theorem is the most-important theorem in probability andstatistics!

CLT: Suppose X1, . . . , Xn are iid with E[Xi] = µ and Var(Xi) = σ2. Thenas n→∞,

Zn =

∑ni=1Xi − nµσ√n

=X̄ − µσ/√n

=X̄ − E[X̄]√

Var(X̄)

d→ Nor(0, 1),

where “ d→ ” means that the cdf of Zn→ the Nor(0, 1) cdf.

ISYE 6739 — Goldsman 8/5/20 75 / 108

The Central Limit Theorem + Proof

Slightly Honors Informal Proof: Suppose that the mgf MX(t) of the Xi’sexists and satisfies certain technical conditions that you don’t need to knowabout. (OK, MX(t) has to exist around t = 0, among other things.)

Moreover, without loss of generality (since we’re standardizing anyway) andfor notational convenience, we’ll assume that µ = 0 and σ2 = 1.

We will be done if we can show that the mgf of Zn converges to the mgf ofZ ∼ Nor(0, 1), i.e., we need to show that MZn(t)→ et

2/2 as n→∞.

To get things going, the mgf of Zn is

MZn(t) = M∑ni=1Xi/

√n (t)

= M∑ni=1Xi

(t/√n) (mgf of a linear function of a RV)

=[MX(t/

√n)]n (Xi’s are iid).

ISYE 6739 — Goldsman 8/5/20 76 / 108

The Central Limit Theorem + Proof

Slightly Honors Informal Proof: Suppose that the mgf MX(t) of the Xi’sexists and satisfies certain technical conditions that you don’t need to knowabout. (OK, MX(t) has to exist around t = 0, among other things.)

Moreover, without loss of generality (since we’re standardizing anyway) andfor notational convenience, we’ll assume that µ = 0 and σ2 = 1.

We will be done if we can show that the mgf of Zn converges to the mgf ofZ ∼ Nor(0, 1), i.e., we need to show that MZn(t)→ et

2/2 as n→∞.

To get things going, the mgf of Zn is

MZn(t) = M∑ni=1Xi/

√n (t)

= M∑ni=1Xi

(t/√n) (mgf of a linear function of a RV)

=[MX(t/

√n)]n (Xi’s are iid).

ISYE 6739 — Goldsman 8/5/20 76 / 108

The Central Limit Theorem + Proof

Slightly Honors Informal Proof: Suppose that the mgf MX(t) of the Xi’sexists and satisfies certain technical conditions that you don’t need to knowabout. (OK, MX(t) has to exist around t = 0, among other things.)

Moreover, without loss of generality (since we’re standardizing anyway) andfor notational convenience, we’ll assume that µ = 0 and σ2 = 1.

We will be done if we can show that the mgf of Zn converges to the mgf ofZ ∼ Nor(0, 1), i.e., we need to show that MZn(t)→ et

2/2 as n→∞.

To get things going, the mgf of Zn is

MZn(t) = M∑ni=1Xi/

√n (t)

= M∑ni=1Xi

(t/√n) (mgf of a linear function of a RV)

=[MX(t/

√n)]n (Xi’s are iid).

ISYE 6739 — Goldsman 8/5/20 76 / 108

The Central Limit Theorem + Proof

Slightly Honors Informal Proof: Suppose that the mgf MX(t) of the Xi’sexists and satisfies certain technical conditions that you don’t need to knowabout. (OK, MX(t) has to exist around t = 0, among other things.)

Moreover, without loss of generality (since we’re standardizing anyway) andfor notational convenience, we’ll assume that µ = 0 and σ2 = 1.

We will be done if we can show that the mgf of Zn converges to the mgf ofZ ∼ Nor(0, 1), i.e., we need to show that MZn(t)→ et

2/2 as n→∞.

To get things going, the mgf of Zn is

MZn(t) = M∑ni=1Xi/

√n (t)

= M∑ni=1Xi

(t/√n) (mgf of a linear function of a RV)

=[MX(t/

√n)]n (Xi’s are iid).

ISYE 6739 — Goldsman 8/5/20 76 / 108

The Central Limit Theorem + Proof

Slightly Honors Informal Proof: Suppose that the mgf MX(t) of the Xi’sexists and satisfies certain technical conditions that you don’t need to knowabout. (OK, MX(t) has to exist around t = 0, among other things.)

Moreover, without loss of generality (since we’re standardizing anyway) andfor notational convenience, we’ll assume that µ = 0 and σ2 = 1.

We will be done if we can show that the mgf of Zn converges to the mgf ofZ ∼ Nor(0, 1), i.e., we need to show that MZn(t)→ et

2/2 as n→∞.

To get things going, the mgf of Zn is

MZn(t) = M∑ni=1Xi/

√n (t)

= M∑ni=1Xi

(t/√n) (mgf of a linear function of a RV)

=[MX(t/

√n)]n (Xi’s are iid).

ISYE 6739 — Goldsman 8/5/20 76 / 108

The Central Limit Theorem + Proof

Slightly Honors Informal Proof: Suppose that the mgf MX(t) of the Xi’sexists and satisfies certain technical conditions that you don’t need to knowabout. (OK, MX(t) has to exist around t = 0, among other things.)

Moreover, without loss of generality (since we’re standardizing anyway) andfor notational convenience, we’ll assume that µ = 0 and σ2 = 1.

We will be done if we can show that the mgf of Zn converges to the mgf ofZ ∼ Nor(0, 1), i.e., we need to show that MZn(t)→ et

2/2 as n→∞.

To get things going, the mgf of Zn is

MZn(t) = M∑ni=1Xi/

√n (t)

= M∑ni=1Xi

(t/√n) (mgf of a linear function of a RV)

=[MX(t/

√n)]n (Xi’s are iid).

ISYE 6739 — Goldsman 8/5/20 76 / 108

The Central Limit Theorem + Proof

Thus, taking logs, our goal is to show that

limn→∞

n `n(MX(t/

√n))

= t2/2.

If we let y = 1/√n, our revised goal is to show that

limy→0

`n(MX(ty)

)y2

= t2/2.

Before proceeding further, note that

limy→0

`n(MX(ty)

)= `n

(MX(0)

)= `n(1) = 0 (3)

andlimy→0

M ′X(ty) = M ′X(0) = E[X] = µ = 0, (4)

where the last equality is from our standardization assumption.

ISYE 6739 — Goldsman 8/5/20 77 / 108

The Central Limit Theorem + Proof

Thus, taking logs, our goal is to show that

limn→∞

n `n(MX(t/

√n))

= t2/2.

If we let y = 1/√n, our revised goal is to show that

limy→0

`n(MX(ty)

)y2

= t2/2.

Before proceeding further, note that

limy→0

`n(MX(ty)

)= `n

(MX(0)

)= `n(1) = 0 (3)

andlimy→0

M ′X(ty) = M ′X(0) = E[X] = µ = 0, (4)

where the last equality is from our standardization assumption.

ISYE 6739 — Goldsman 8/5/20 77 / 108

The Central Limit Theorem + Proof

Thus, taking logs, our goal is to show that

limn→∞

n `n(MX(t/

√n))

= t2/2.

If we let y = 1/√n, our revised goal is to show that

limy→0

`n(MX(ty)

)y2

= t2/2.

Before proceeding further, note that

limy→0

`n(MX(ty)

)= `n

(MX(0)

)= `n(1) = 0 (3)

andlimy→0

M ′X(ty) = M ′X(0) = E[X] = µ = 0, (4)

where the last equality is from our standardization assumption.

ISYE 6739 — Goldsman 8/5/20 77 / 108

The Central Limit Theorem + Proof

Thus, taking logs, our goal is to show that

limn→∞

n `n(MX(t/

√n))

= t2/2.

If we let y = 1/√n, our revised goal is to show that

limy→0

`n(MX(ty)

)y2

= t2/2.

Before proceeding further, note that

limy→0

`n(MX(ty)

)= `n

(MX(0)

)= `n(1) = 0 (3)

andlimy→0

M ′X(ty) = M ′X(0) = E[X] = µ = 0, (4)

where the last equality is from our standardization assumption.

ISYE 6739 — Goldsman 8/5/20 77 / 108

The Central Limit Theorem + Proof

Thus, taking logs, our goal is to show that

limn→∞

n `n(MX(t/

√n))

= t2/2.

If we let y = 1/√n, our revised goal is to show that

limy→0

`n(MX(ty)

)y2

= t2/2.

Before proceeding further, note that

limy→0

`n(MX(ty)

)= `n

(MX(0)

)= `n(1) = 0 (3)

andlimy→0

M ′X(ty) = M ′X(0) = E[X] = µ = 0, (4)

where the last equality is from our standardization assumption.

ISYE 6739 — Goldsman 8/5/20 77 / 108

The Central Limit Theorem + Proof

Thus, taking logs, our goal is to show that

limn→∞

n `n(MX(t/

√n))

= t2/2.

If we let y = 1/√n, our revised goal is to show that

limy→0

`n(MX(ty)

)y2

= t2/2.

Before proceeding further, note that

limy→0

`n(MX(ty)

)= `n

(MX(0)

)= `n(1) = 0 (3)

andlimy→0

M ′X(ty) = M ′X(0) = E[X] = µ = 0, (4)

where the last equality is from our standardization assumption.

ISYE 6739 — Goldsman 8/5/20 77 / 108

The Central Limit Theorem + Proof

So after all of this build-up, we have

limy→0

`n(MX(ty)

)y2

= limy→0

tM ′X(ty)

2yMX(ty)(by (3) et L’Hôspital to deal with 0 / 0)

= limy→0

t2M ′′X(ty)

2MX(ty) + 2ytM ′X(ty)(by (4) et L’Hôspital encore)

=t2M ′′X(0)

2MX(0) + 0=

t2 E[X2]

2=

t2

2(E[X2] = σ2 − µ2 = 1).

ISYE 6739 — Goldsman 8/5/20 78 / 108

The Central Limit Theorem + Proof

So after all of this build-up, we have

limy→0

`n(MX(ty)

)y2

= limy→0

tM ′X(ty)

2yMX(ty)(by (3) et L’Hôspital to deal with 0 / 0)

= limy→0

t2M ′′X(ty)

2MX(ty) + 2ytM ′X(ty)(by (4) et L’Hôspital encore)

=t2M ′′X(0)

2MX(0) + 0=

t2 E[X2]

2=

t2

2(E[X2] = σ2 − µ2 = 1).

ISYE 6739 — Goldsman 8/5/20 78 / 108

The Central Limit Theorem + Proof

So after all of this build-up, we have

limy→0

`n(MX(ty)

)y2

= limy→0

tM ′X(ty)

2yMX(ty)(by (3) et L’Hôspital to deal with 0 / 0)

= limy→0

t2M ′′X(ty)

2MX(ty) + 2ytM ′X(ty)(by (4) et L’Hôspital encore)

=t2M ′′X(0)

2MX(0) + 0

=t2 E[X2]

2=

t2

2(E[X2] = σ2 − µ2 = 1).

ISYE 6739 — Goldsman 8/5/20 78 / 108

The Central Limit Theorem + Proof

So after all of this build-up, we have

limy→0

`n(MX(ty)

)y2

= limy→0

tM ′X(ty)

2yMX(ty)(by (3) et L’Hôspital to deal with 0 / 0)

= limy→0

t2M ′′X(ty)

2MX(ty) + 2ytM ′X(ty)(by (4) et L’Hôspital encore)

=t2M ′′X(0)

2MX(0) + 0=

t2 E[X2]

2

=t2

2(E[X2] = σ2 − µ2 = 1).

ISYE 6739 — Goldsman 8/5/20 78 / 108

The Central Limit Theorem + Proof

So after all of this build-up, we have

limy→0

`n(MX(ty)

)y2

= limy→0

tM ′X(ty)

2yMX(ty)(by (3) et L’Hôspital to deal with 0 / 0)

= limy→0

t2M ′′X(ty)

2MX(ty) + 2ytM ′X(ty)(by (4) et L’Hôspital encore)

=t2M ′′X(0)

2MX(0) + 0=

t2 E[X2]

2=

t2

2(E[X2] = σ2 − µ2 = 1).

ISYE 6739 — Goldsman 8/5/20 78 / 108

The Central Limit Theorem + Proof

Remarks: We have a lot to say about such an important theorem.

(1) If n is large, then X̄ ≈ Nor(µ, σ2/n).

(2) The Xi’s don’t have to be normal for the CLT to work! It even workson discrete distributions!

(3) You usually need n ≥ 30 observations for the approximation to work well.(Need fewer observations if the Xi’s come from a symmetric distribution.)

(4) You can almost always use the CLT if the observations are iid.

(5) In fact, there are versions of the CLT that are actually a lot more generalthan the theorem presented here!

ISYE 6739 — Goldsman 8/5/20 79 / 108

The Central Limit Theorem + Proof

Remarks: We have a lot to say about such an important theorem.

(1) If n is large, then X̄ ≈ Nor(µ, σ2/n).

(2) The Xi’s don’t have to be normal for the CLT to work! It even workson discrete distributions!

(3) You usually need n ≥ 30 observations for the approximation to work well.(Need fewer observations if the Xi’s come from a symmetric distribution.)

(4) You can almost always use the CLT if the observations are iid.

(5) In fact, there are versions of the CLT that are actually a lot more generalthan the theorem presented here!

ISYE 6739 — Goldsman 8/5/20 79 / 108

The Central Limit Theorem + Proof

Remarks: We have a lot to say about such an important theorem.

(1) If n is large, then X̄ ≈ Nor(µ, σ2/n).

(2) The Xi’s don’t have to be normal for the CLT to work! It even workson discrete distributions!

(3) You usually need n ≥ 30 observations for the approximation to work well.(Need fewer observations if the Xi’s come from a symmetric distribution.)

(4) You can almost always use the CLT if the observations are iid.

(5) In fact, there are versions of the CLT that are actually a lot more generalthan the theorem presented here!

ISYE 6739 — Goldsman 8/5/20 79 / 108

The Central Limit Theorem + Proof

Remarks: We have a lot to say about such an important theorem.

(1) If n is large, then X̄ ≈ Nor(µ, σ2/n).

(2) The Xi’s don’t have to be normal for the CLT to work! It even workson discrete distributions!

(3) You usually need n ≥ 30 observations for the approximation to work well.(Need fewer observations if the Xi’s come from a symmetric distribution.)

(4) You can almost always use the CLT if the observations are iid.

(5) In fact, there are versions of the CLT that are actually a lot more generalthan the theorem presented here!

ISYE 6739 — Goldsman 8/5/20 79 / 108

The Central Limit Theorem + Proof

Remarks: We have a lot to say about such an important theorem.

(1) If n is large, then X̄ ≈ Nor(µ, σ2/n).

(2) The Xi’s don’t have to be normal for the CLT to work! It even workson discrete distributions!

(3) You usually need n ≥ 30 observations for the approximation to work well.(Need fewer observations if the Xi’s come from a symmetric distribution.)

(4) You can almost always use the CLT if the observations are iid.

(5) In fact, there are versions of the CLT that are actually a lot more generalthan the theorem presented here!

ISYE 6739 — Goldsman 8/5/20 79 / 108

The Central Limit Theorem + Proof

Remarks: We have a lot to say about such an important theorem.

(1) If n is large, then X̄ ≈ Nor(µ, σ2/n).

(2) The Xi’s don’t have to be normal for the CLT to work! It even workson discrete distributions!

(3) You usually need n ≥ 30 observations for the approximation to work well.(Need fewer observations if the Xi’s come from a symmetric distribution.)

(4) You can almost always use the CLT if the observations are iid.

(5) In fact, there are versions of the CLT that are actually a lot more generalthan the theorem presented here!

ISYE 6739 — Goldsman 8/5/20 79 / 108

Central Limit Theorem Examples

1 Bernoulli and Binomial Distributions

2 Hypergeometric Distribution3 Geometric and Negative Binomial Distributions

4 Poisson Distribution5 Uniform, Exponential, and Friends6 Other Continuous Distributions7 Normal Distribution: Basics8 Standard Normal Distribution9 Sample Mean of Normals

10 The Central Limit Theorem + Proof

11 Central Limit Theorem Examples

12 Extensions — Multivariate Normal Distribution13 Extensions — Lognormal Distribution

14 Computer Stuff

ISYE 6739 — Goldsman 8/5/20 80 / 108

Central Limit Theorem Examples

Lesson 4.11 — Central Limit Theorem Examples

Example: To show that the Central Limit Theory really works, let’s add upjust n = 12 iid Unif(0,1)’s, U1, . . . , Un. Let Sn =

∑ni=1 Ui. Note that

E[Sn] = nE[Ui] = n/2, and Var(Sn) = nVar(Ui) = n/12. Therefore,

Zn ≡Sn − n/2√

n/12= Sn − 6 ≈ Nor(0, 1).

The histogram was compiled using 100,000 simulations of Z12. It works! 2

ISYE 6739 — Goldsman 8/5/20 81 / 108

Central Limit Theorem Examples

Lesson 4.11 — Central Limit Theorem Examples

Example: To show that the Central Limit Theory really works, let’s add upjust n = 12 iid Unif(0,1)’s, U1, . . . , Un. Let Sn =

∑ni=1 Ui.

Note thatE[Sn] = nE[Ui] = n/2, and Var(Sn) = nVar(Ui) = n/12. Therefore,

Zn ≡Sn − n/2√

n/12= Sn − 6 ≈ Nor(0, 1).

The histogram was compiled using 100,000 simulations of Z12. It works! 2

ISYE 6739 — Goldsman 8/5/20 81 / 108

Central Limit Theorem Examples

Lesson 4.11 — Central Limit Theorem Examples

Example: To show that the Central Limit Theory really works, let’s add upjust n = 12 iid Unif(0,1)’s, U1, . . . , Un. Let Sn =

∑ni=1 Ui. Note that

E[Sn] = nE[Ui] = n/2, and Var(Sn) = nVar(Ui) = n/12. Therefore,

Zn ≡Sn − n/2√

n/12= Sn − 6 ≈ Nor(0, 1).

The histogram was compiled using 100,000 simulations of Z12. It works! 2

ISYE 6739 — Goldsman 8/5/20 81 / 108

Central Limit Theorem Examples

Lesson 4.11 — Central Limit Theorem Examples

Example: To show that the Central Limit Theory really works, let’s add upjust n = 12 iid Unif(0,1)’s, U1, . . . , Un. Let Sn =

∑ni=1 Ui. Note that

E[Sn] = nE[Ui] = n/2, and Var(Sn) = nVar(Ui) = n/12. Therefore,

Zn ≡Sn − n/2√

n/12= Sn − 6 ≈ Nor(0, 1).

The histogram was compiled using 100,000 simulations of Z12. It works! 2

ISYE 6739 — Goldsman 8/5/20 81 / 108

Central Limit Theorem Examples

Lesson 4.11 — Central Limit Theorem Examples

Example: To show that the Central Limit Theory really works, let’s add upjust n = 12 iid Unif(0,1)’s, U1, . . . , Un. Let Sn =

∑ni=1 Ui. Note that

E[Sn] = nE[Ui] = n/2, and Var(Sn) = nVar(Ui) = n/12. Therefore,

Zn ≡Sn − n/2√

n/12= Sn − 6 ≈ Nor(0, 1).

The histogram was compiled using 100,000 simulations of Z12. It works! 2

ISYE 6739 — Goldsman 8/5/20 81 / 108

Central Limit Theorem Examples

Example: Suppose X1, . . . , X100iid∼ Exp(1/1000).

Find P (950 ≤ X̄ ≤ 1050).

Solution: Recall that if Xi ∼ Exp(λ), then E[Xi] = 1/λ andVar(Xi) = 1/λ2.

Further, if X̄ is the sample mean based on n observations, then

E[X̄] = E[Xi] = 1/λ and

Var(X̄) = Var(Xi)/n = 1/(nλ2).

For our problem, λ = 1/1000 and n = 100, so that E[X̄] = 1000 andVar(X̄) = 10000.

ISYE 6739 — Goldsman 8/5/20 82 / 108

Central Limit Theorem Examples

Example: Suppose X1, . . . , X100iid∼ Exp(1/1000).

Find P (950 ≤ X̄ ≤ 1050).

Solution: Recall that if Xi ∼ Exp(λ), then E[Xi] = 1/λ andVar(Xi) = 1/λ2.

Further, if X̄ is the sample mean based on n observations, then

E[X̄] = E[Xi] = 1/λ and

Var(X̄) = Var(Xi)/n = 1/(nλ2).

For our problem, λ = 1/1000 and n = 100, so that E[X̄] = 1000 andVar(X̄) = 10000.

ISYE 6739 — Goldsman 8/5/20 82 / 108

Central Limit Theorem Examples

Example: Suppose X1, . . . , X100iid∼ Exp(1/1000).

Find P (950 ≤ X̄ ≤ 1050).

Solution: Recall that if Xi ∼ Exp(λ), then E[Xi] = 1/λ andVar(Xi) = 1/λ2.

Further, if X̄ is the sample mean based on n observations, then

E[X̄] = E[Xi] = 1/λ and

Var(X̄) = Var(Xi)/n = 1/(nλ2).

For our problem, λ = 1/1000 and n = 100, so that E[X̄] = 1000 andVar(X̄) = 10000.

ISYE 6739 — Goldsman 8/5/20 82 / 108

Central Limit Theorem Examples

Example: Suppose X1, . . . , X100iid∼ Exp(1/1000).

Find P (950 ≤ X̄ ≤ 1050).

Solution: Recall that if Xi ∼ Exp(λ), then E[Xi] = 1/λ andVar(Xi) = 1/λ2.

Further, if X̄ is the sample mean based on n observations, then

E[X̄] = E[Xi] = 1/λ and

Var(X̄) = Var(Xi)/n = 1/(nλ2).

For our problem, λ = 1/1000 and n = 100, so that E[X̄] = 1000 andVar(X̄) = 10000.

ISYE 6739 — Goldsman 8/5/20 82 / 108

Central Limit Theorem Examples

Example: Suppose X1, . . . , X100iid∼ Exp(1/1000).

Find P (950 ≤ X̄ ≤ 1050).

Solution: Recall that if Xi ∼ Exp(λ), then E[Xi] = 1/λ andVar(Xi) = 1/λ2.

Further, if X̄ is the sample mean based on n observations, then

E[X̄] = E[Xi] = 1/λ and

Var(X̄) = Var(Xi)/n = 1/(nλ2).

For our problem, λ = 1/1000 and n = 100, so that E[X̄] = 1000 andVar(X̄) = 10000.

ISYE 6739 — Goldsman 8/5/20 82 / 108

Central Limit Theorem Examples

Example: Suppose X1, . . . , X100iid∼ Exp(1/1000).

Find P (950 ≤ X̄ ≤ 1050).

Solution: Recall that if Xi ∼ Exp(λ), then E[Xi] = 1/λ andVar(Xi) = 1/λ2.

Further, if X̄ is the sample mean based on n observations, then

E[X̄] = E[Xi] = 1/λ and

Var(X̄) = Var(Xi)/n = 1/(nλ2).

For our problem, λ = 1/1000 and n = 100, so that E[X̄] = 1000 andVar(X̄) = 10000.

ISYE 6739 — Goldsman 8/5/20 82 / 108

Central Limit Theorem Examples

Example: Suppose X1, . . . , X100iid∼ Exp(1/1000).

Find P (950 ≤ X̄ ≤ 1050).

Solution: Recall that if Xi ∼ Exp(λ), then E[Xi] = 1/λ andVar(Xi) = 1/λ2.

Further, if X̄ is the sample mean based on n observations, then

E[X̄] = E[Xi] = 1/λ and

Var(X̄) = Var(Xi)/n = 1/(nλ2).

For our problem, λ = 1/1000 and n = 100, so that E[X̄] = 1000 andVar(X̄) = 10000.

ISYE 6739 — Goldsman 8/5/20 82 / 108

Central Limit Theorem Examples

So by the CLT,

P (950 ≤ X̄ ≤ 1050)

= P

(950− E[X̄]√

Var(X̄)≤ X̄ − E[X̄]√

Var(X̄)≤ 1050− E[X̄]√

Var(X̄)

).= P

(950− 1000

100≤ Z ≤ 1050− 1000

100

)= P

(−1

2≤ Z ≤ 1

2

)= 2Φ(1/2)− 1 = 0.383. 2

Remark: This problem can be solved exactly if we have access to the ExcelErlang cdf function GAMMADIST. And what do you know, you end up withexactly the same answer of 0.383!

ISYE 6739 — Goldsman 8/5/20 83 / 108

Central Limit Theorem Examples

So by the CLT,

P (950 ≤ X̄ ≤ 1050)

= P

(950− E[X̄]√

Var(X̄)≤ X̄ − E[X̄]√

Var(X̄)≤ 1050− E[X̄]√

Var(X̄)

)

.= P

(950− 1000

100≤ Z ≤ 1050− 1000

100

)= P

(−1

2≤ Z ≤ 1

2

)= 2Φ(1/2)− 1 = 0.383. 2

Remark: This problem can be solved exactly if we have access to the ExcelErlang cdf function GAMMADIST. And what do you know, you end up withexactly the same answer of 0.383!

ISYE 6739 — Goldsman 8/5/20 83 / 108

Central Limit Theorem Examples

So by the CLT,

P (950 ≤ X̄ ≤ 1050)

= P

(950− E[X̄]√

Var(X̄)≤ X̄ − E[X̄]√

Var(X̄)≤ 1050− E[X̄]√

Var(X̄)

).= P

(950− 1000

100≤ Z ≤ 1050− 1000

100

)

= P

(−1

2≤ Z ≤ 1

2

)= 2Φ(1/2)− 1 = 0.383. 2

Remark: This problem can be solved exactly if we have access to the ExcelErlang cdf function GAMMADIST. And what do you know, you end up withexactly the same answer of 0.383!

ISYE 6739 — Goldsman 8/5/20 83 / 108

Central Limit Theorem Examples

So by the CLT,

P (950 ≤ X̄ ≤ 1050)

= P

(950− E[X̄]√

Var(X̄)≤ X̄ − E[X̄]√

Var(X̄)≤ 1050− E[X̄]√

Var(X̄)

).= P

(950− 1000

100≤ Z ≤ 1050− 1000

100

)= P

(−1

2≤ Z ≤ 1

2

)= 2Φ(1/2)− 1 = 0.383. 2

Remark: This problem can be solved exactly if we have access to the ExcelErlang cdf function GAMMADIST. And what do you know, you end up withexactly the same answer of 0.383!

ISYE 6739 — Goldsman 8/5/20 83 / 108

Central Limit Theorem Examples

So by the CLT,

P (950 ≤ X̄ ≤ 1050)

= P

(950− E[X̄]√

Var(X̄)≤ X̄ − E[X̄]√

Var(X̄)≤ 1050− E[X̄]√

Var(X̄)

).= P

(950− 1000

100≤ Z ≤ 1050− 1000

100

)= P

(−1

2≤ Z ≤ 1

2

)= 2Φ(1/2)− 1 = 0.383. 2

Remark: This problem can be solved exactly if we have access to the ExcelErlang cdf function GAMMADIST. And what do you know, you end up withexactly the same answer of 0.383!

ISYE 6739 — Goldsman 8/5/20 83 / 108

Central Limit Theorem Examples

Example: Suppose X1, . . . , X100 are iid from some distribution with mean1000 and standard deviation 1000.

Find P (950 ≤ X̄ ≤ 1050).

Solution: By exactly the same manipulations as in the previous example, theanswer .= 0.383.

Notice that we didn’t care whether or not the data came from an exponentialdistribution. We just needed the mean and variance. 2

ISYE 6739 — Goldsman 8/5/20 84 / 108

Central Limit Theorem Examples

Example: Suppose X1, . . . , X100 are iid from some distribution with mean1000 and standard deviation 1000. Find P (950 ≤ X̄ ≤ 1050).

Solution: By exactly the same manipulations as in the previous example, theanswer .= 0.383.

Notice that we didn’t care whether or not the data came from an exponentialdistribution. We just needed the mean and variance. 2

ISYE 6739 — Goldsman 8/5/20 84 / 108

Central Limit Theorem Examples

Example: Suppose X1, . . . , X100 are iid from some distribution with mean1000 and standard deviation 1000. Find P (950 ≤ X̄ ≤ 1050).

Solution: By exactly the same manipulations as in the previous example, theanswer .= 0.383.

Notice that we didn’t care whether or not the data came from an exponentialdistribution. We just needed the mean and variance. 2

ISYE 6739 — Goldsman 8/5/20 84 / 108

Central Limit Theorem Examples

Example: Suppose X1, . . . , X100 are iid from some distribution with mean1000 and standard deviation 1000. Find P (950 ≤ X̄ ≤ 1050).

Solution: By exactly the same manipulations as in the previous example, theanswer .= 0.383.

Notice that we didn’t care whether or not the data came from an exponentialdistribution. We just needed the mean and variance. 2

ISYE 6739 — Goldsman 8/5/20 84 / 108

Central Limit Theorem Examples

Normal Approximation to the Binomial

Suppose Y ∼ Bin(n, p), where n is very large. In such cases, we usuallyapproximate the Binomial via an appropriate Normal distribution.

The CLT applies since Y =∑n

i=1Xi, where the Xi’s are iid Bern(p).

ThenY − E[Y ]√

Var(Y )=

Y − np√npq

≈ Nor(0, 1).

The usual rule of thumb for the Normal approximation to the Binomial is thatit works pretty well as long as np ≥ 5 and nq ≥ 5.

ISYE 6739 — Goldsman 8/5/20 85 / 108

Central Limit Theorem Examples

Normal Approximation to the Binomial

Suppose Y ∼ Bin(n, p), where n is very large. In such cases, we usuallyapproximate the Binomial via an appropriate Normal distribution.

The CLT applies since Y =∑n

i=1Xi, where the Xi’s are iid Bern(p).

ThenY − E[Y ]√

Var(Y )=

Y − np√npq

≈ Nor(0, 1).

The usual rule of thumb for the Normal approximation to the Binomial is thatit works pretty well as long as np ≥ 5 and nq ≥ 5.

ISYE 6739 — Goldsman 8/5/20 85 / 108

Central Limit Theorem Examples

Normal Approximation to the Binomial

Suppose Y ∼ Bin(n, p), where n is very large. In such cases, we usuallyapproximate the Binomial via an appropriate Normal distribution.

The CLT applies since Y =∑n

i=1Xi, where the Xi’s are iid Bern(p).

ThenY − E[Y ]√

Var(Y )=

Y − np√npq

≈ Nor(0, 1).

The usual rule of thumb for the Normal approximation to the Binomial is thatit works pretty well as long as np ≥ 5 and nq ≥ 5.

ISYE 6739 — Goldsman 8/5/20 85 / 108

Central Limit Theorem Examples

Normal Approximation to the Binomial

Suppose Y ∼ Bin(n, p), where n is very large. In such cases, we usuallyapproximate the Binomial via an appropriate Normal distribution.

The CLT applies since Y =∑n

i=1Xi, where the Xi’s are iid Bern(p).

ThenY − E[Y ]√

Var(Y )=

Y − np√npq

≈ Nor(0, 1).

The usual rule of thumb for the Normal approximation to the Binomial is thatit works pretty well as long as np ≥ 5 and nq ≥ 5.

ISYE 6739 — Goldsman 8/5/20 85 / 108

Central Limit Theorem Examples

Normal Approximation to the Binomial

Suppose Y ∼ Bin(n, p), where n is very large. In such cases, we usuallyapproximate the Binomial via an appropriate Normal distribution.

The CLT applies since Y =∑n

i=1Xi, where the Xi’s are iid Bern(p).

ThenY − E[Y ]√

Var(Y )=

Y − np√npq

≈ Nor(0, 1).

The usual rule of thumb for the Normal approximation to the Binomial is thatit works pretty well as long as np ≥ 5 and nq ≥ 5.

ISYE 6739 — Goldsman 8/5/20 85 / 108

Central Limit Theorem Examples

Why do we need such an approximation?

Example: Suppose Y ∼ Bin(100, 0.8) and we want

P (Y ≥ 84) =100∑i=84

(100

i

)(0.8)i(0.2)100−i.

Good luck with the binomial coefficients (they’re too big) and the number ofterms to sum up (it’s going to get tedious). I’ll come back to visit you in anhour.

The next example shows how to use the approximation.

Note that it incorporates a “continuity correction” to account for the factthat the Binomial is discrete while the Normal is continuous. If you don’twant to use it, don’t worry too much.

ISYE 6739 — Goldsman 8/5/20 86 / 108

Central Limit Theorem Examples

Why do we need such an approximation?

Example: Suppose Y ∼ Bin(100, 0.8) and we want

P (Y ≥ 84) =

100∑i=84

(100

i

)(0.8)i(0.2)100−i.

Good luck with the binomial coefficients (they’re too big) and the number ofterms to sum up (it’s going to get tedious). I’ll come back to visit you in anhour.

The next example shows how to use the approximation.

Note that it incorporates a “continuity correction” to account for the factthat the Binomial is discrete while the Normal is continuous. If you don’twant to use it, don’t worry too much.

ISYE 6739 — Goldsman 8/5/20 86 / 108

Central Limit Theorem Examples

Why do we need such an approximation?

Example: Suppose Y ∼ Bin(100, 0.8) and we want

P (Y ≥ 84) =

100∑i=84

(100

i

)(0.8)i(0.2)100−i.

Good luck with the binomial coefficients (they’re too big) and the number ofterms to sum up (it’s going to get tedious). I’ll come back to visit you in anhour.

The next example shows how to use the approximation.

Note that it incorporates a “continuity correction” to account for the factthat the Binomial is discrete while the Normal is continuous. If you don’twant to use it, don’t worry too much.

ISYE 6739 — Goldsman 8/5/20 86 / 108

Central Limit Theorem Examples

Why do we need such an approximation?

Example: Suppose Y ∼ Bin(100, 0.8) and we want

P (Y ≥ 84) =

100∑i=84

(100

i

)(0.8)i(0.2)100−i.

Good luck with the binomial coefficients (they’re too big) and the number ofterms to sum up (it’s going to get tedious). I’ll come back to visit you in anhour.

The next example shows how to use the approximation.

Note that it incorporates a “continuity correction” to account for the factthat the Binomial is discrete while the Normal is continuous. If you don’twant to use it, don’t worry too much.

ISYE 6739 — Goldsman 8/5/20 86 / 108

Central Limit Theorem Examples

Why do we need such an approximation?

Example: Suppose Y ∼ Bin(100, 0.8) and we want

P (Y ≥ 84) =

100∑i=84

(100

i

)(0.8)i(0.2)100−i.

Good luck with the binomial coefficients (they’re too big) and the number ofterms to sum up (it’s going to get tedious). I’ll come back to visit you in anhour.

The next example shows how to use the approximation.

Note that it incorporates a “continuity correction” to account for the factthat the Binomial is discrete while the Normal is continuous. If you don’twant to use it, don’t worry too much.

ISYE 6739 — Goldsman 8/5/20 86 / 108

Central Limit Theorem Examples

Example: The Braves play 100 independent baseball games, each of whichthey have probability 0.8 of winning. What’s the probability that they win ≥84?

Y ∼ Bin(100, 0.8) and we want P (Y ≥ 84) (as in the last example). . .

P (Y ≥ 84) = P (Y ≥ 83.5) (“continuity correction”)

.= P

(Z ≥ 83.5− np

√npq

)(CLT)

= P

(Z ≥ 83.5− 80√

16

)= P (Z ≥ 0.875) = 0.1908.

The actual answer (using the true Bin(100,0.8) distribution) turns out to be0.1923 — pretty close! 2

ISYE 6739 — Goldsman 8/5/20 87 / 108

Central Limit Theorem Examples

Example: The Braves play 100 independent baseball games, each of whichthey have probability 0.8 of winning. What’s the probability that they win ≥84?

Y ∼ Bin(100, 0.8) and we want P (Y ≥ 84) (as in the last example). . .

P (Y ≥ 84) = P (Y ≥ 83.5) (“continuity correction”)

.= P

(Z ≥ 83.5− np

√npq

)(CLT)

= P

(Z ≥ 83.5− 80√

16

)= P (Z ≥ 0.875) = 0.1908.

The actual answer (using the true Bin(100,0.8) distribution) turns out to be0.1923 — pretty close! 2

ISYE 6739 — Goldsman 8/5/20 87 / 108

Central Limit Theorem Examples

Example: The Braves play 100 independent baseball games, each of whichthey have probability 0.8 of winning. What’s the probability that they win ≥84?

Y ∼ Bin(100, 0.8) and we want P (Y ≥ 84) (as in the last example). . .

P (Y ≥ 84) = P (Y ≥ 83.5) (“continuity correction”)

.= P

(Z ≥ 83.5− np

√npq

)(CLT)

= P

(Z ≥ 83.5− 80√

16

)= P (Z ≥ 0.875) = 0.1908.

The actual answer (using the true Bin(100,0.8) distribution) turns out to be0.1923 — pretty close! 2

ISYE 6739 — Goldsman 8/5/20 87 / 108

Central Limit Theorem Examples

Example: The Braves play 100 independent baseball games, each of whichthey have probability 0.8 of winning. What’s the probability that they win ≥84?

Y ∼ Bin(100, 0.8) and we want P (Y ≥ 84) (as in the last example). . .

P (Y ≥ 84) = P (Y ≥ 83.5) (“continuity correction”)

.= P

(Z ≥ 83.5− np

√npq

)(CLT)

= P

(Z ≥ 83.5− 80√

16

)= P (Z ≥ 0.875) = 0.1908.

The actual answer (using the true Bin(100,0.8) distribution) turns out to be0.1923 — pretty close! 2

ISYE 6739 — Goldsman 8/5/20 87 / 108

Central Limit Theorem Examples

Example: The Braves play 100 independent baseball games, each of whichthey have probability 0.8 of winning. What’s the probability that they win ≥84?

Y ∼ Bin(100, 0.8) and we want P (Y ≥ 84) (as in the last example). . .

P (Y ≥ 84) = P (Y ≥ 83.5) (“continuity correction”)

.= P

(Z ≥ 83.5− np

√npq

)(CLT)

= P

(Z ≥ 83.5− 80√

16

)= P (Z ≥ 0.875) = 0.1908.

The actual answer (using the true Bin(100,0.8) distribution) turns out to be0.1923 — pretty close! 2

ISYE 6739 — Goldsman 8/5/20 87 / 108

Central Limit Theorem Examples

Example: The Braves play 100 independent baseball games, each of whichthey have probability 0.8 of winning. What’s the probability that they win ≥84?

Y ∼ Bin(100, 0.8) and we want P (Y ≥ 84) (as in the last example). . .

P (Y ≥ 84) = P (Y ≥ 83.5) (“continuity correction”)

.= P

(Z ≥ 83.5− np

√npq

)(CLT)

= P

(Z ≥ 83.5− 80√

16

)= P (Z ≥ 0.875) = 0.1908.

The actual answer (using the true Bin(100,0.8) distribution) turns out to be0.1923 — pretty close! 2

ISYE 6739 — Goldsman 8/5/20 87 / 108

Extensions — Multivariate Normal Distribution

1 Bernoulli and Binomial Distributions

2 Hypergeometric Distribution3 Geometric and Negative Binomial Distributions

4 Poisson Distribution5 Uniform, Exponential, and Friends6 Other Continuous Distributions7 Normal Distribution: Basics8 Standard Normal Distribution9 Sample Mean of Normals

10 The Central Limit Theorem + Proof

11 Central Limit Theorem Examples

12 Extensions — Multivariate Normal Distribution13 Extensions — Lognormal Distribution

14 Computer Stuff

ISYE 6739 — Goldsman 8/5/20 88 / 108

Extensions — Multivariate Normal Distribution

Lesson 4.12 — Extensions — Multivariate Normal Distribution

Definition: (X,Y ) has the Bivariate Normal distribution if it has pdf

f(x, y) = C exp

−[z2X(x)− 2ρzX(x)zY (y) + z2Y (y)

]2(1− ρ2)

,

whereρ ≡ Corr(X,Y ), C ≡ 1

2πσXσY√

1− ρ2,

zX(x) ≡ x− µXσX

, and zY (y) ≡ y − µYσY

.

ISYE 6739 — Goldsman 8/5/20 89 / 108

Extensions — Multivariate Normal Distribution

Lesson 4.12 — Extensions — Multivariate Normal Distribution

Definition: (X,Y ) has the Bivariate Normal distribution if it has pdf

f(x, y) = C exp

−[z2X(x)− 2ρzX(x)zY (y) + z2Y (y)

]2(1− ρ2)

,

whereρ ≡ Corr(X,Y ), C ≡ 1

2πσXσY√

1− ρ2,

zX(x) ≡ x− µXσX

, and zY (y) ≡ y − µYσY

.

ISYE 6739 — Goldsman 8/5/20 89 / 108

Extensions — Multivariate Normal Distribution

Lesson 4.12 — Extensions — Multivariate Normal Distribution

Definition: (X,Y ) has the Bivariate Normal distribution if it has pdf

f(x, y) = C exp

−[z2X(x)− 2ρzX(x)zY (y) + z2Y (y)

]2(1− ρ2)

,

whereρ ≡ Corr(X,Y ), C ≡ 1

2πσXσY√

1− ρ2,

zX(x) ≡ x− µXσX

, and zY (y) ≡ y − µYσY

.

ISYE 6739 — Goldsman 8/5/20 89 / 108

Extensions — Multivariate Normal Distribution

Lesson 4.12 — Extensions — Multivariate Normal Distribution

Definition: (X,Y ) has the Bivariate Normal distribution if it has pdf

f(x, y) = C exp

−[z2X(x)− 2ρzX(x)zY (y) + z2Y (y)

]2(1− ρ2)

,

whereρ ≡ Corr(X,Y ), C ≡ 1

2πσXσY√

1− ρ2,

zX(x) ≡ x− µXσX

, and zY (y) ≡ y − µYσY

.

ISYE 6739 — Goldsman 8/5/20 89 / 108

Extensions — Multivariate Normal Distribution

Pretty nasty joint pdf, eh?

In fact, X ∼ Nor(µX , σ2X) and Y ∼ Nor(µY , σ2Y ).

Example: (X,Y ) could be a person’s (height,weight). The two quantitiesare marginally normal, but positively correlated.

If you want to calculate bivariate normal probabilities, you’ll need to evaluatequantities like

P (a < X < b, c < Y < d) =

∫ d

c

∫ b

af(x, y) dx dy,

which will probably require numerical integration techniques.

ISYE 6739 — Goldsman 8/5/20 90 / 108

Extensions — Multivariate Normal Distribution

Pretty nasty joint pdf, eh?

In fact, X ∼ Nor(µX , σ2X) and Y ∼ Nor(µY , σ2Y ).

Example: (X,Y ) could be a person’s (height,weight). The two quantitiesare marginally normal, but positively correlated.

If you want to calculate bivariate normal probabilities, you’ll need to evaluatequantities like

P (a < X < b, c < Y < d) =

∫ d

c

∫ b

af(x, y) dx dy,

which will probably require numerical integration techniques.

ISYE 6739 — Goldsman 8/5/20 90 / 108

Extensions — Multivariate Normal Distribution

Pretty nasty joint pdf, eh?

In fact, X ∼ Nor(µX , σ2X) and Y ∼ Nor(µY , σ2Y ).

Example: (X,Y ) could be a person’s (height,weight). The two quantitiesare marginally normal, but positively correlated.

If you want to calculate bivariate normal probabilities, you’ll need to evaluatequantities like

P (a < X < b, c < Y < d) =

∫ d

c

∫ b

af(x, y) dx dy,

which will probably require numerical integration techniques.

ISYE 6739 — Goldsman 8/5/20 90 / 108

Extensions — Multivariate Normal Distribution

Pretty nasty joint pdf, eh?

In fact, X ∼ Nor(µX , σ2X) and Y ∼ Nor(µY , σ2Y ).

Example: (X,Y ) could be a person’s (height,weight). The two quantitiesare marginally normal, but positively correlated.

If you want to calculate bivariate normal probabilities, you’ll need to evaluatequantities like

P (a < X < b, c < Y < d) =

∫ d

c

∫ b

af(x, y) dx dy,

which will probably require numerical integration techniques.

ISYE 6739 — Goldsman 8/5/20 90 / 108

Extensions — Multivariate Normal Distribution

Pretty nasty joint pdf, eh?

In fact, X ∼ Nor(µX , σ2X) and Y ∼ Nor(µY , σ2Y ).

Example: (X,Y ) could be a person’s (height,weight). The two quantitiesare marginally normal, but positively correlated.

If you want to calculate bivariate normal probabilities, you’ll need to evaluatequantities like

P (a < X < b, c < Y < d) =

∫ d

c

∫ b

af(x, y) dx dy,

which will probably require numerical integration techniques.

ISYE 6739 — Goldsman 8/5/20 90 / 108

Extensions — Multivariate Normal Distribution

Fun Fact (which will come up later when we discuss regression): Theconditional distribution of Y given that X = x is also normal. In particular,

Y |X = x ∼ Nor(µY + ρ(σY /σX)(x− µX), σ2Y (1− ρ2)

).

Information about X helps to update the distribution of Y .

Example: Consider students at a university. Let X be their combined SATscores (Math and Verbal), and Y their freshman GPA (out of 4).Suppose a study reveals that

µX = 1300, µY = 2.3,

σ2X = 6400, σ2Y = 0.25, ρ = 0.6.

Find P (Y ≥ 2|X = 900).

ISYE 6739 — Goldsman 8/5/20 91 / 108

Extensions — Multivariate Normal Distribution

Fun Fact (which will come up later when we discuss regression): Theconditional distribution of Y given that X = x is also normal. In particular,

Y |X = x ∼ Nor(µY + ρ(σY /σX)(x− µX), σ2Y (1− ρ2)

).

Information about X helps to update the distribution of Y .

Example: Consider students at a university. Let X be their combined SATscores (Math and Verbal), and Y their freshman GPA (out of 4).Suppose a study reveals that

µX = 1300, µY = 2.3,

σ2X = 6400, σ2Y = 0.25, ρ = 0.6.

Find P (Y ≥ 2|X = 900).

ISYE 6739 — Goldsman 8/5/20 91 / 108

Extensions — Multivariate Normal Distribution

Fun Fact (which will come up later when we discuss regression): Theconditional distribution of Y given that X = x is also normal. In particular,

Y |X = x ∼ Nor(µY + ρ(σY /σX)(x− µX), σ2Y (1− ρ2)

).

Information about X helps to update the distribution of Y .

Example: Consider students at a university. Let X be their combined SATscores (Math and Verbal), and Y their freshman GPA (out of 4).Suppose a study reveals that

µX = 1300, µY = 2.3,

σ2X = 6400, σ2Y = 0.25, ρ = 0.6.

Find P (Y ≥ 2|X = 900).

ISYE 6739 — Goldsman 8/5/20 91 / 108

Extensions — Multivariate Normal Distribution

Fun Fact (which will come up later when we discuss regression): Theconditional distribution of Y given that X = x is also normal. In particular,

Y |X = x ∼ Nor(µY + ρ(σY /σX)(x− µX), σ2Y (1− ρ2)

).

Information about X helps to update the distribution of Y .

Example: Consider students at a university. Let X be their combined SATscores (Math and Verbal), and Y their freshman GPA (out of 4).

Suppose a study reveals that

µX = 1300, µY = 2.3,

σ2X = 6400, σ2Y = 0.25, ρ = 0.6.

Find P (Y ≥ 2|X = 900).

ISYE 6739 — Goldsman 8/5/20 91 / 108

Extensions — Multivariate Normal Distribution

Fun Fact (which will come up later when we discuss regression): Theconditional distribution of Y given that X = x is also normal. In particular,

Y |X = x ∼ Nor(µY + ρ(σY /σX)(x− µX), σ2Y (1− ρ2)

).

Information about X helps to update the distribution of Y .

Example: Consider students at a university. Let X be their combined SATscores (Math and Verbal), and Y their freshman GPA (out of 4).Suppose a study reveals that

µX = 1300, µY = 2.3,

σ2X = 6400, σ2Y = 0.25, ρ = 0.6.

Find P (Y ≥ 2|X = 900).

ISYE 6739 — Goldsman 8/5/20 91 / 108

Extensions — Multivariate Normal Distribution

Fun Fact (which will come up later when we discuss regression): Theconditional distribution of Y given that X = x is also normal. In particular,

Y |X = x ∼ Nor(µY + ρ(σY /σX)(x− µX), σ2Y (1− ρ2)

).

Information about X helps to update the distribution of Y .

Example: Consider students at a university. Let X be their combined SATscores (Math and Verbal), and Y their freshman GPA (out of 4).Suppose a study reveals that

µX = 1300, µY = 2.3,

σ2X = 6400, σ2Y = 0.25, ρ = 0.6.

Find P (Y ≥ 2|X = 900).

ISYE 6739 — Goldsman 8/5/20 91 / 108

Extensions — Multivariate Normal Distribution

Fun Fact (which will come up later when we discuss regression): Theconditional distribution of Y given that X = x is also normal. In particular,

Y |X = x ∼ Nor(µY + ρ(σY /σX)(x− µX), σ2Y (1− ρ2)

).

Information about X helps to update the distribution of Y .

Example: Consider students at a university. Let X be their combined SATscores (Math and Verbal), and Y their freshman GPA (out of 4).Suppose a study reveals that

µX = 1300, µY = 2.3,

σ2X = 6400, σ2Y = 0.25, ρ = 0.6.

Find P (Y ≥ 2|X = 900).

ISYE 6739 — Goldsman 8/5/20 91 / 108

Extensions — Multivariate Normal Distribution

First,

E[Y |X = 900] = µY + ρ(σY /σX)(x− µX)

= 2.3 + (0.6)(√

0.25/6400 )(900− 1300) = 0.8,

indicating that the expected GPA of a kid with 900 SAT’s will be 0.8.

Second,Var(Y |X = 900) = σ2Y (1− ρ2) = 0.16.

Thus,Y |X = 900 ∼ Nor(0.8, 0.16).

Now we can calculate

P (Y ≥ 2|X = 900) = P(Z ≥ 2− 0.8√

0.16

)= 1− Φ(3) = 0.0013.

This guy doesn’t have much chance of having a good GPA. 2

ISYE 6739 — Goldsman 8/5/20 92 / 108

Extensions — Multivariate Normal Distribution

First,

E[Y |X = 900] = µY + ρ(σY /σX)(x− µX)

= 2.3 + (0.6)(√

0.25/6400 )(900− 1300) = 0.8,

indicating that the expected GPA of a kid with 900 SAT’s will be 0.8.

Second,Var(Y |X = 900) = σ2Y (1− ρ2) = 0.16.

Thus,Y |X = 900 ∼ Nor(0.8, 0.16).

Now we can calculate

P (Y ≥ 2|X = 900) = P(Z ≥ 2− 0.8√

0.16

)= 1− Φ(3) = 0.0013.

This guy doesn’t have much chance of having a good GPA. 2

ISYE 6739 — Goldsman 8/5/20 92 / 108

Extensions — Multivariate Normal Distribution

First,

E[Y |X = 900] = µY + ρ(σY /σX)(x− µX)

= 2.3 + (0.6)(√

0.25/6400 )(900− 1300) = 0.8,

indicating that the expected GPA of a kid with 900 SAT’s will be 0.8.

Second,Var(Y |X = 900) = σ2Y (1− ρ2) = 0.16.

Thus,Y |X = 900 ∼ Nor(0.8, 0.16).

Now we can calculate

P (Y ≥ 2|X = 900) = P(Z ≥ 2− 0.8√

0.16

)= 1− Φ(3) = 0.0013.

This guy doesn’t have much chance of having a good GPA. 2

ISYE 6739 — Goldsman 8/5/20 92 / 108

Extensions — Multivariate Normal Distribution

First,

E[Y |X = 900] = µY + ρ(σY /σX)(x− µX)

= 2.3 + (0.6)(√

0.25/6400 )(900− 1300) = 0.8,

indicating that the expected GPA of a kid with 900 SAT’s will be 0.8.

Second,Var(Y |X = 900) = σ2Y (1− ρ2) = 0.16.

Thus,Y |X = 900 ∼ Nor(0.8, 0.16).

Now we can calculate

P (Y ≥ 2|X = 900) = P(Z ≥ 2− 0.8√

0.16

)= 1− Φ(3) = 0.0013.

This guy doesn’t have much chance of having a good GPA. 2

ISYE 6739 — Goldsman 8/5/20 92 / 108

Extensions — Multivariate Normal Distribution

First,

E[Y |X = 900] = µY + ρ(σY /σX)(x− µX)

= 2.3 + (0.6)(√

0.25/6400 )(900− 1300) = 0.8,

indicating that the expected GPA of a kid with 900 SAT’s will be 0.8.

Second,Var(Y |X = 900) = σ2Y (1− ρ2) = 0.16.

Thus,Y |X = 900 ∼ Nor(0.8, 0.16).

Now we can calculate

P (Y ≥ 2|X = 900) = P(Z ≥ 2− 0.8√

0.16

)= 1− Φ(3) = 0.0013.

This guy doesn’t have much chance of having a good GPA. 2

ISYE 6739 — Goldsman 8/5/20 92 / 108

Extensions — Multivariate Normal Distribution

First,

E[Y |X = 900] = µY + ρ(σY /σX)(x− µX)

= 2.3 + (0.6)(√

0.25/6400 )(900− 1300) = 0.8,

indicating that the expected GPA of a kid with 900 SAT’s will be 0.8.

Second,Var(Y |X = 900) = σ2Y (1− ρ2) = 0.16.

Thus,Y |X = 900 ∼ Nor(0.8, 0.16).

Now we can calculate

P (Y ≥ 2|X = 900) = P(Z ≥ 2− 0.8√

0.16

)

= 1− Φ(3) = 0.0013.

This guy doesn’t have much chance of having a good GPA. 2

ISYE 6739 — Goldsman 8/5/20 92 / 108

Extensions — Multivariate Normal Distribution

First,

E[Y |X = 900] = µY + ρ(σY /σX)(x− µX)

= 2.3 + (0.6)(√

0.25/6400 )(900− 1300) = 0.8,

indicating that the expected GPA of a kid with 900 SAT’s will be 0.8.

Second,Var(Y |X = 900) = σ2Y (1− ρ2) = 0.16.

Thus,Y |X = 900 ∼ Nor(0.8, 0.16).

Now we can calculate

P (Y ≥ 2|X = 900) = P(Z ≥ 2− 0.8√

0.16

)= 1− Φ(3) = 0.0013.

This guy doesn’t have much chance of having a good GPA. 2

ISYE 6739 — Goldsman 8/5/20 92 / 108

Extensions — Multivariate Normal Distribution

First,

E[Y |X = 900] = µY + ρ(σY /σX)(x− µX)

= 2.3 + (0.6)(√

0.25/6400 )(900− 1300) = 0.8,

indicating that the expected GPA of a kid with 900 SAT’s will be 0.8.

Second,Var(Y |X = 900) = σ2Y (1− ρ2) = 0.16.

Thus,Y |X = 900 ∼ Nor(0.8, 0.16).

Now we can calculate

P (Y ≥ 2|X = 900) = P(Z ≥ 2− 0.8√

0.16

)= 1− Φ(3) = 0.0013.

This guy doesn’t have much chance of having a good GPA. 2

ISYE 6739 — Goldsman 8/5/20 92 / 108

Extensions — Multivariate Normal Distribution

The bivariate normal distribution is easily generalized to the multivariate case.

Honors Definition: The random vectorX = (X1, . . . , Xk)T has the

multivariate normal distribution with mean vector µ = (µ1, . . . , µk)T

and k × k covariance matrix Σ = (σij) if it has multivariate pdf

f(x) =1

(2π)k/2|Σ|1/2exp

{−(x− µ)TΣ−1(x− µ)

2

}, x ∈ Rk,

where |Σ| and Σ−1 are the determinant and inverse of Σ, respectively.

It turns out that

E[Xi] = µi, Var(Xi) = σii, Cov(Xi, Xj) = σij .

Notation: X ∼ Nork(µ, Σ).

ISYE 6739 — Goldsman 8/5/20 93 / 108

Extensions — Multivariate Normal Distribution

The bivariate normal distribution is easily generalized to the multivariate case.

Honors Definition: The random vectorX = (X1, . . . , Xk)T has the

multivariate normal distribution with mean vector µ = (µ1, . . . , µk)T

and k × k covariance matrix Σ = (σij) if it has multivariate pdf

f(x) =1

(2π)k/2|Σ|1/2exp

{−(x− µ)TΣ−1(x− µ)

2

}, x ∈ Rk,

where |Σ| and Σ−1 are the determinant and inverse of Σ, respectively.

It turns out that

E[Xi] = µi, Var(Xi) = σii, Cov(Xi, Xj) = σij .

Notation: X ∼ Nork(µ, Σ).

ISYE 6739 — Goldsman 8/5/20 93 / 108

Extensions — Multivariate Normal Distribution

The bivariate normal distribution is easily generalized to the multivariate case.

Honors Definition: The random vectorX = (X1, . . . , Xk)T has the

multivariate normal distribution with mean vector µ = (µ1, . . . , µk)T

and k × k covariance matrix Σ = (σij) if it has multivariate pdf

f(x) =1

(2π)k/2|Σ|1/2exp

{−(x− µ)TΣ−1(x− µ)

2

}, x ∈ Rk,

where |Σ| and Σ−1 are the determinant and inverse of Σ, respectively.

It turns out that

E[Xi] = µi, Var(Xi) = σii, Cov(Xi, Xj) = σij .

Notation: X ∼ Nork(µ, Σ).

ISYE 6739 — Goldsman 8/5/20 93 / 108

Extensions — Multivariate Normal Distribution

The bivariate normal distribution is easily generalized to the multivariate case.

Honors Definition: The random vectorX = (X1, . . . , Xk)T has the

multivariate normal distribution with mean vector µ = (µ1, . . . , µk)T

and k × k covariance matrix Σ = (σij) if it has multivariate pdf

f(x) =1

(2π)k/2|Σ|1/2exp

{−(x− µ)TΣ−1(x− µ)

2

}, x ∈ Rk,

where |Σ| and Σ−1 are the determinant and inverse of Σ, respectively.

It turns out that

E[Xi] = µi, Var(Xi) = σii, Cov(Xi, Xj) = σij .

Notation: X ∼ Nork(µ, Σ).

ISYE 6739 — Goldsman 8/5/20 93 / 108

Extensions — Multivariate Normal Distribution

The bivariate normal distribution is easily generalized to the multivariate case.

Honors Definition: The random vectorX = (X1, . . . , Xk)T has the

multivariate normal distribution with mean vector µ = (µ1, . . . , µk)T

and k × k covariance matrix Σ = (σij) if it has multivariate pdf

f(x) =1

(2π)k/2|Σ|1/2exp

{−(x− µ)TΣ−1(x− µ)

2

}, x ∈ Rk,

where |Σ| and Σ−1 are the determinant and inverse of Σ, respectively.

It turns out that

E[Xi] = µi, Var(Xi) = σii, Cov(Xi, Xj) = σij .

Notation: X ∼ Nork(µ, Σ).

ISYE 6739 — Goldsman 8/5/20 93 / 108

Extensions — Multivariate Normal Distribution

The bivariate normal distribution is easily generalized to the multivariate case.

Honors Definition: The random vectorX = (X1, . . . , Xk)T has the

multivariate normal distribution with mean vector µ = (µ1, . . . , µk)T

and k × k covariance matrix Σ = (σij) if it has multivariate pdf

f(x) =1

(2π)k/2|Σ|1/2exp

{−(x− µ)TΣ−1(x− µ)

2

}, x ∈ Rk,

where |Σ| and Σ−1 are the determinant and inverse of Σ, respectively.

It turns out that

E[Xi] = µi, Var(Xi) = σii, Cov(Xi, Xj) = σij .

Notation: X ∼ Nork(µ, Σ).

ISYE 6739 — Goldsman 8/5/20 93 / 108

Extensions — Lognormal Distribution

1 Bernoulli and Binomial Distributions

2 Hypergeometric Distribution3 Geometric and Negative Binomial Distributions

4 Poisson Distribution5 Uniform, Exponential, and Friends6 Other Continuous Distributions7 Normal Distribution: Basics8 Standard Normal Distribution9 Sample Mean of Normals

10 The Central Limit Theorem + Proof

11 Central Limit Theorem Examples

12 Extensions — Multivariate Normal Distribution13 Extensions — Lognormal Distribution

14 Computer Stuff

ISYE 6739 — Goldsman 8/5/20 94 / 108

Extensions — Lognormal Distribution

Lesson 4.13 — Extensions — Lognormal Distribution

Definition: If Y ∼ Nor(µY , σ2Y ), then X ≡ eY has the Lognormaldistribution with parameters (µY , σ

2Y ). This distribution has tremendous

uses, e.g., in the pricing of certain stock options.

Turns Out: The pdf and moments of the lognormal are

f(x) =1

xσY√

2πexp

{− 1

2σ2Y[`n(x)− µY ]2

}, x > 0,

E[X] = exp

{µY +

σ2Y2

},

Var(X) = exp{

2µY + σ2Y}(

exp{σ2Y}− 1).

ISYE 6739 — Goldsman 8/5/20 95 / 108

Extensions — Lognormal Distribution

Lesson 4.13 — Extensions — Lognormal Distribution

Definition: If Y ∼ Nor(µY , σ2Y ), then X ≡ eY has the Lognormaldistribution with parameters (µY , σ

2Y ). This distribution has tremendous

uses, e.g., in the pricing of certain stock options.

Turns Out: The pdf and moments of the lognormal are

f(x) =1

xσY√

2πexp

{− 1

2σ2Y[`n(x)− µY ]2

}, x > 0,

E[X] = exp

{µY +

σ2Y2

},

Var(X) = exp{

2µY + σ2Y}(

exp{σ2Y}− 1).

ISYE 6739 — Goldsman 8/5/20 95 / 108

Extensions — Lognormal Distribution

Lesson 4.13 — Extensions — Lognormal Distribution

Definition: If Y ∼ Nor(µY , σ2Y ), then X ≡ eY has the Lognormaldistribution with parameters (µY , σ

2Y ). This distribution has tremendous

uses, e.g., in the pricing of certain stock options.

Turns Out: The pdf and moments of the lognormal are

f(x) =1

xσY√

2πexp

{− 1

2σ2Y[`n(x)− µY ]2

}, x > 0,

E[X] = exp

{µY +

σ2Y2

},

Var(X) = exp{

2µY + σ2Y}(

exp{σ2Y}− 1).

ISYE 6739 — Goldsman 8/5/20 95 / 108

Extensions — Lognormal Distribution

Lesson 4.13 — Extensions — Lognormal Distribution

Definition: If Y ∼ Nor(µY , σ2Y ), then X ≡ eY has the Lognormaldistribution with parameters (µY , σ

2Y ). This distribution has tremendous

uses, e.g., in the pricing of certain stock options.

Turns Out: The pdf and moments of the lognormal are

f(x) =1

xσY√

2πexp

{− 1

2σ2Y[`n(x)− µY ]2

}, x > 0,

E[X] = exp

{µY +

σ2Y2

},

Var(X) = exp{

2µY + σ2Y}(

exp{σ2Y}− 1).

ISYE 6739 — Goldsman 8/5/20 95 / 108

Extensions — Lognormal Distribution

Lesson 4.13 — Extensions — Lognormal Distribution

Definition: If Y ∼ Nor(µY , σ2Y ), then X ≡ eY has the Lognormaldistribution with parameters (µY , σ

2Y ). This distribution has tremendous

uses, e.g., in the pricing of certain stock options.

Turns Out: The pdf and moments of the lognormal are

f(x) =1

xσY√

2πexp

{− 1

2σ2Y[`n(x)− µY ]2

}, x > 0,

E[X] = exp

{µY +

σ2Y2

},

Var(X) = exp{

2µY + σ2Y}(

exp{σ2Y}− 1).

ISYE 6739 — Goldsman 8/5/20 95 / 108

Extensions — Lognormal Distribution

Example: Suppose Y ∼ Nor(10, 4) and let X = eY . Then

P (X ≤ 1000) = P(Y ≤ `n(1000)

)= P

(Z ≤ `n(1000)− 10

2

)= Φ(−1.55) = 0.061. 2

Honors Example (How to Win a Nobel Prize:) It is well-known thatstock prices are closely related to the lognormal distribution. In fact, it’scommon to use the following model for a stock price at a fixed time t,

S(t) = S(0) exp

{(µ− σ2

2

)t+ σ

√tZ

}, t ≥ 0,

where µ is related to the “drift” of the stock price (i.e., the natural rate ofincrease), σ is its “volatility” (how much the stock bounces around), S(0) isthe initial price, and Z is a standard normal RV.

ISYE 6739 — Goldsman 8/5/20 96 / 108

Extensions — Lognormal Distribution

Example: Suppose Y ∼ Nor(10, 4) and let X = eY . Then

P (X ≤ 1000) = P(Y ≤ `n(1000)

)

= P(Z ≤ `n(1000)− 10

2

)= Φ(−1.55) = 0.061. 2

Honors Example (How to Win a Nobel Prize:) It is well-known thatstock prices are closely related to the lognormal distribution. In fact, it’scommon to use the following model for a stock price at a fixed time t,

S(t) = S(0) exp

{(µ− σ2

2

)t+ σ

√tZ

}, t ≥ 0,

where µ is related to the “drift” of the stock price (i.e., the natural rate ofincrease), σ is its “volatility” (how much the stock bounces around), S(0) isthe initial price, and Z is a standard normal RV.

ISYE 6739 — Goldsman 8/5/20 96 / 108

Extensions — Lognormal Distribution

Example: Suppose Y ∼ Nor(10, 4) and let X = eY . Then

P (X ≤ 1000) = P(Y ≤ `n(1000)

)= P

(Z ≤ `n(1000)− 10

2

)

= Φ(−1.55) = 0.061. 2

Honors Example (How to Win a Nobel Prize:) It is well-known thatstock prices are closely related to the lognormal distribution. In fact, it’scommon to use the following model for a stock price at a fixed time t,

S(t) = S(0) exp

{(µ− σ2

2

)t+ σ

√tZ

}, t ≥ 0,

where µ is related to the “drift” of the stock price (i.e., the natural rate ofincrease), σ is its “volatility” (how much the stock bounces around), S(0) isthe initial price, and Z is a standard normal RV.

ISYE 6739 — Goldsman 8/5/20 96 / 108

Extensions — Lognormal Distribution

Example: Suppose Y ∼ Nor(10, 4) and let X = eY . Then

P (X ≤ 1000) = P(Y ≤ `n(1000)

)= P

(Z ≤ `n(1000)− 10

2

)= Φ(−1.55) = 0.061. 2

Honors Example (How to Win a Nobel Prize:) It is well-known thatstock prices are closely related to the lognormal distribution. In fact, it’scommon to use the following model for a stock price at a fixed time t,

S(t) = S(0) exp

{(µ− σ2

2

)t+ σ

√tZ

}, t ≥ 0,

where µ is related to the “drift” of the stock price (i.e., the natural rate ofincrease), σ is its “volatility” (how much the stock bounces around), S(0) isthe initial price, and Z is a standard normal RV.

ISYE 6739 — Goldsman 8/5/20 96 / 108

Extensions — Lognormal Distribution

Example: Suppose Y ∼ Nor(10, 4) and let X = eY . Then

P (X ≤ 1000) = P(Y ≤ `n(1000)

)= P

(Z ≤ `n(1000)− 10

2

)= Φ(−1.55) = 0.061. 2

Honors Example (How to Win a Nobel Prize:) It is well-known thatstock prices are closely related to the lognormal distribution. In fact, it’scommon to use the following model for a stock price at a fixed time t,

S(t) = S(0) exp

{(µ− σ2

2

)t+ σ

√tZ

}, t ≥ 0,

where µ is related to the “drift” of the stock price (i.e., the natural rate ofincrease), σ is its “volatility” (how much the stock bounces around), S(0) isthe initial price, and Z is a standard normal RV.

ISYE 6739 — Goldsman 8/5/20 96 / 108

Extensions — Lognormal Distribution

Example: Suppose Y ∼ Nor(10, 4) and let X = eY . Then

P (X ≤ 1000) = P(Y ≤ `n(1000)

)= P

(Z ≤ `n(1000)− 10

2

)= Φ(−1.55) = 0.061. 2

Honors Example (How to Win a Nobel Prize:) It is well-known thatstock prices are closely related to the lognormal distribution. In fact, it’scommon to use the following model for a stock price at a fixed time t,

S(t) = S(0) exp

{(µ− σ2

2

)t+ σ

√tZ

}, t ≥ 0,

where µ is related to the “drift” of the stock price (i.e., the natural rate ofincrease), σ is its “volatility” (how much the stock bounces around), S(0) isthe initial price, and Z is a standard normal RV.

ISYE 6739 — Goldsman 8/5/20 96 / 108

Extensions — Lognormal Distribution

Example: Suppose Y ∼ Nor(10, 4) and let X = eY . Then

P (X ≤ 1000) = P(Y ≤ `n(1000)

)= P

(Z ≤ `n(1000)− 10

2

)= Φ(−1.55) = 0.061. 2

Honors Example (How to Win a Nobel Prize:) It is well-known thatstock prices are closely related to the lognormal distribution. In fact, it’scommon to use the following model for a stock price at a fixed time t,

S(t) = S(0) exp

{(µ− σ2

2

)t+ σ

√tZ

}, t ≥ 0,

where µ is related to the “drift” of the stock price (i.e., the natural rate ofincrease), σ is its “volatility” (how much the stock bounces around), S(0) isthe initial price, and Z is a standard normal RV.

ISYE 6739 — Goldsman 8/5/20 96 / 108

Extensions — Lognormal Distribution

An active area of finance is to estimate option prices.

For example, aso-called European call option C permits its owner, who pays an up-frontfee for the privilege, to purchase the stock at a pre-agreed strike price k, at apre-determined expiry date T .

For instance, suppose IBM is currently selling for $100 a share. If I think thatthe stock will go up in value, I may want to pay $3/share now for the right tobuy IBM at $105 three months from now.

If IBM is worth $120 three months from now, I’ll be able to buy it foronly $105, and will have made a profit of $120− $105− $3 = $12.

If IBM is selling for $107 three months hence, I can still buy it for $105,and will lose $1 (recouping $2 from my original option purchase).

If IBM is selling for $95, then I won’t exercise my option, and will walkaway with my tail between my legs having lost my original $3.

ISYE 6739 — Goldsman 8/5/20 97 / 108

Extensions — Lognormal Distribution

An active area of finance is to estimate option prices. For example, aso-called European call option C permits its owner, who pays an up-frontfee for the privilege, to purchase the stock at a pre-agreed strike price k, at apre-determined expiry date T .

For instance, suppose IBM is currently selling for $100 a share. If I think thatthe stock will go up in value, I may want to pay $3/share now for the right tobuy IBM at $105 three months from now.

If IBM is worth $120 three months from now, I’ll be able to buy it foronly $105, and will have made a profit of $120− $105− $3 = $12.

If IBM is selling for $107 three months hence, I can still buy it for $105,and will lose $1 (recouping $2 from my original option purchase).

If IBM is selling for $95, then I won’t exercise my option, and will walkaway with my tail between my legs having lost my original $3.

ISYE 6739 — Goldsman 8/5/20 97 / 108

Extensions — Lognormal Distribution

An active area of finance is to estimate option prices. For example, aso-called European call option C permits its owner, who pays an up-frontfee for the privilege, to purchase the stock at a pre-agreed strike price k, at apre-determined expiry date T .

For instance, suppose IBM is currently selling for $100 a share. If I think thatthe stock will go up in value, I may want to pay $3/share now for the right tobuy IBM at $105 three months from now.

If IBM is worth $120 three months from now, I’ll be able to buy it foronly $105, and will have made a profit of $120− $105− $3 = $12.

If IBM is selling for $107 three months hence, I can still buy it for $105,and will lose $1 (recouping $2 from my original option purchase).

If IBM is selling for $95, then I won’t exercise my option, and will walkaway with my tail between my legs having lost my original $3.

ISYE 6739 — Goldsman 8/5/20 97 / 108

Extensions — Lognormal Distribution

An active area of finance is to estimate option prices. For example, aso-called European call option C permits its owner, who pays an up-frontfee for the privilege, to purchase the stock at a pre-agreed strike price k, at apre-determined expiry date T .

For instance, suppose IBM is currently selling for $100 a share. If I think thatthe stock will go up in value, I may want to pay $3/share now for the right tobuy IBM at $105 three months from now.

If IBM is worth $120 three months from now, I’ll be able to buy it foronly $105, and will have made a profit of $120− $105− $3 = $12.

If IBM is selling for $107 three months hence, I can still buy it for $105,and will lose $1 (recouping $2 from my original option purchase).

If IBM is selling for $95, then I won’t exercise my option, and will walkaway with my tail between my legs having lost my original $3.

ISYE 6739 — Goldsman 8/5/20 97 / 108

Extensions — Lognormal Distribution

An active area of finance is to estimate option prices. For example, aso-called European call option C permits its owner, who pays an up-frontfee for the privilege, to purchase the stock at a pre-agreed strike price k, at apre-determined expiry date T .

For instance, suppose IBM is currently selling for $100 a share. If I think thatthe stock will go up in value, I may want to pay $3/share now for the right tobuy IBM at $105 three months from now.

If IBM is worth $120 three months from now, I’ll be able to buy it foronly $105, and will have made a profit of $120− $105− $3 = $12.

If IBM is selling for $107 three months hence, I can still buy it for $105,and will lose $1 (recouping $2 from my original option purchase).

If IBM is selling for $95, then I won’t exercise my option, and will walkaway with my tail between my legs having lost my original $3.

ISYE 6739 — Goldsman 8/5/20 97 / 108

Extensions — Lognormal Distribution

An active area of finance is to estimate option prices. For example, aso-called European call option C permits its owner, who pays an up-frontfee for the privilege, to purchase the stock at a pre-agreed strike price k, at apre-determined expiry date T .

For instance, suppose IBM is currently selling for $100 a share. If I think thatthe stock will go up in value, I may want to pay $3/share now for the right tobuy IBM at $105 three months from now.

If IBM is worth $120 three months from now, I’ll be able to buy it foronly $105, and will have made a profit of $120− $105− $3 = $12.

If IBM is selling for $107 three months hence, I can still buy it for $105,and will lose $1 (recouping $2 from my original option purchase).

If IBM is selling for $95, then I won’t exercise my option, and will walkaway with my tail between my legs having lost my original $3.

ISYE 6739 — Goldsman 8/5/20 97 / 108

Extensions — Lognormal Distribution

So what is the option worth (and what should I pay for it)? Its expected valueis given by

E[C] = e−rTE[(S(T )− k

)+],

where

x+ ≡ max{0, x}.r, the “risk-free” interest rate (e.g., what you can get from a U.S.Treasury bond). This is used instead of the drift µ.

The term e−rT denotes the time-value of money, i.e., a depreciation termcorresponding to the interest I could’ve made had I used my money tobuy a Treasury note.

Black and Scholes won a Nobel Prize for calculating E[C]. We’ll get thesame answer via a different method — but, alas, no Nobel Prize.

ISYE 6739 — Goldsman 8/5/20 98 / 108

Extensions — Lognormal Distribution

So what is the option worth (and what should I pay for it)? Its expected valueis given by

E[C] = e−rTE[(S(T )− k

)+],

where

x+ ≡ max{0, x}.r, the “risk-free” interest rate (e.g., what you can get from a U.S.Treasury bond). This is used instead of the drift µ.

The term e−rT denotes the time-value of money, i.e., a depreciation termcorresponding to the interest I could’ve made had I used my money tobuy a Treasury note.

Black and Scholes won a Nobel Prize for calculating E[C]. We’ll get thesame answer via a different method — but, alas, no Nobel Prize.

ISYE 6739 — Goldsman 8/5/20 98 / 108

Extensions — Lognormal Distribution

So what is the option worth (and what should I pay for it)? Its expected valueis given by

E[C] = e−rTE[(S(T )− k

)+],

where

x+ ≡ max{0, x}.

r, the “risk-free” interest rate (e.g., what you can get from a U.S.Treasury bond). This is used instead of the drift µ.

The term e−rT denotes the time-value of money, i.e., a depreciation termcorresponding to the interest I could’ve made had I used my money tobuy a Treasury note.

Black and Scholes won a Nobel Prize for calculating E[C]. We’ll get thesame answer via a different method — but, alas, no Nobel Prize.

ISYE 6739 — Goldsman 8/5/20 98 / 108

Extensions — Lognormal Distribution

So what is the option worth (and what should I pay for it)? Its expected valueis given by

E[C] = e−rTE[(S(T )− k

)+],

where

x+ ≡ max{0, x}.r, the “risk-free” interest rate (e.g., what you can get from a U.S.Treasury bond). This is used instead of the drift µ.

The term e−rT denotes the time-value of money, i.e., a depreciation termcorresponding to the interest I could’ve made had I used my money tobuy a Treasury note.

Black and Scholes won a Nobel Prize for calculating E[C]. We’ll get thesame answer via a different method — but, alas, no Nobel Prize.

ISYE 6739 — Goldsman 8/5/20 98 / 108

Extensions — Lognormal Distribution

So what is the option worth (and what should I pay for it)? Its expected valueis given by

E[C] = e−rTE[(S(T )− k

)+],

where

x+ ≡ max{0, x}.r, the “risk-free” interest rate (e.g., what you can get from a U.S.Treasury bond). This is used instead of the drift µ.

The term e−rT denotes the time-value of money, i.e., a depreciation termcorresponding to the interest I could’ve made had I used my money tobuy a Treasury note.

Black and Scholes won a Nobel Prize for calculating E[C]. We’ll get thesame answer via a different method — but, alas, no Nobel Prize.

ISYE 6739 — Goldsman 8/5/20 98 / 108

Extensions — Lognormal Distribution

So what is the option worth (and what should I pay for it)? Its expected valueis given by

E[C] = e−rTE[(S(T )− k

)+],

where

x+ ≡ max{0, x}.r, the “risk-free” interest rate (e.g., what you can get from a U.S.Treasury bond). This is used instead of the drift µ.

The term e−rT denotes the time-value of money, i.e., a depreciation termcorresponding to the interest I could’ve made had I used my money tobuy a Treasury note.

Black and Scholes won a Nobel Prize for calculating E[C]. We’ll get thesame answer via a different method — but, alas, no Nobel Prize.

ISYE 6739 — Goldsman 8/5/20 98 / 108

Extensions — Lognormal Distribution

E[C] = e−rTE

[S(0) exp

{(r − σ2

2

)T + σ

√TZ

}− k]+

= e−rT∫ ∞−∞

[S(0) exp

{(r − σ2

2

)T + σ

√Tz

}− k]+

φ(z) dz

(via the standard conditioning argument)

= S(0)Φ(b+ σ√T )− ke−rTΦ(b) (after lots of algebra),

where φ(·) and Φ(·) are the Nor(0,1) pdf and cdf, respectively, and

b ≡rT − σ2T

2 − `n(k/S(0))

σ√T

.

There are many generalizations of this problem that are used in practicalfinance problems, but this is the starting point. Meanwhile, get your tickets toNorway or Sweden or wherever they give out the Nobel!

ISYE 6739 — Goldsman 8/5/20 99 / 108

Extensions — Lognormal Distribution

E[C] = e−rTE

[S(0) exp

{(r − σ2

2

)T + σ

√TZ

}− k]+

= e−rT∫ ∞−∞

[S(0) exp

{(r − σ2

2

)T + σ

√Tz

}− k]+

φ(z) dz

(via the standard conditioning argument)

= S(0)Φ(b+ σ√T )− ke−rTΦ(b) (after lots of algebra),

where φ(·) and Φ(·) are the Nor(0,1) pdf and cdf, respectively, and

b ≡rT − σ2T

2 − `n(k/S(0))

σ√T

.

There are many generalizations of this problem that are used in practicalfinance problems, but this is the starting point. Meanwhile, get your tickets toNorway or Sweden or wherever they give out the Nobel!

ISYE 6739 — Goldsman 8/5/20 99 / 108

Extensions — Lognormal Distribution

E[C] = e−rTE

[S(0) exp

{(r − σ2

2

)T + σ

√TZ

}− k]+

= e−rT∫ ∞−∞

[S(0) exp

{(r − σ2

2

)T + σ

√Tz

}− k]+

φ(z) dz

(via the standard conditioning argument)

= S(0)Φ(b+ σ√T )− ke−rTΦ(b) (after lots of algebra),

where φ(·) and Φ(·) are the Nor(0,1) pdf and cdf, respectively, and

b ≡rT − σ2T

2 − `n(k/S(0))

σ√T

.

There are many generalizations of this problem that are used in practicalfinance problems, but this is the starting point. Meanwhile, get your tickets toNorway or Sweden or wherever they give out the Nobel!

ISYE 6739 — Goldsman 8/5/20 99 / 108

Extensions — Lognormal Distribution

E[C] = e−rTE

[S(0) exp

{(r − σ2

2

)T + σ

√TZ

}− k]+

= e−rT∫ ∞−∞

[S(0) exp

{(r − σ2

2

)T + σ

√Tz

}− k]+

φ(z) dz

(via the standard conditioning argument)

= S(0)Φ(b+ σ√T )− ke−rTΦ(b) (after lots of algebra),

where φ(·) and Φ(·) are the Nor(0,1) pdf and cdf, respectively, and

b ≡rT − σ2T

2 − `n(k/S(0))

σ√T

.

There are many generalizations of this problem that are used in practicalfinance problems, but this is the starting point. Meanwhile, get your tickets toNorway or Sweden or wherever they give out the Nobel!

ISYE 6739 — Goldsman 8/5/20 99 / 108

Extensions — Lognormal Distribution

E[C] = e−rTE

[S(0) exp

{(r − σ2

2

)T + σ

√TZ

}− k]+

= e−rT∫ ∞−∞

[S(0) exp

{(r − σ2

2

)T + σ

√Tz

}− k]+

φ(z) dz

(via the standard conditioning argument)

= S(0)Φ(b+ σ√T )− ke−rTΦ(b) (after lots of algebra),

where φ(·) and Φ(·) are the Nor(0,1) pdf and cdf, respectively, and

b ≡rT − σ2T

2 − `n(k/S(0))

σ√T

.

There are many generalizations of this problem that are used in practicalfinance problems, but this is the starting point. Meanwhile, get your tickets toNorway or Sweden or wherever they give out the Nobel!

ISYE 6739 — Goldsman 8/5/20 99 / 108

Extensions — Lognormal Distribution

E[C] = e−rTE

[S(0) exp

{(r − σ2

2

)T + σ

√TZ

}− k]+

= e−rT∫ ∞−∞

[S(0) exp

{(r − σ2

2

)T + σ

√Tz

}− k]+

φ(z) dz

(via the standard conditioning argument)

= S(0)Φ(b+ σ√T )− ke−rTΦ(b) (after lots of algebra),

where φ(·) and Φ(·) are the Nor(0,1) pdf and cdf, respectively, and

b ≡rT − σ2T

2 − `n(k/S(0))

σ√T

.

There are many generalizations of this problem that are used in practicalfinance problems, but this is the starting point. Meanwhile, get your tickets toNorway or Sweden or wherever they give out the Nobel!

ISYE 6739 — Goldsman 8/5/20 99 / 108

Computer Stuff

1 Bernoulli and Binomial Distributions

2 Hypergeometric Distribution3 Geometric and Negative Binomial Distributions

4 Poisson Distribution5 Uniform, Exponential, and Friends6 Other Continuous Distributions7 Normal Distribution: Basics8 Standard Normal Distribution9 Sample Mean of Normals

10 The Central Limit Theorem + Proof

11 Central Limit Theorem Examples

12 Extensions — Multivariate Normal Distribution13 Extensions — Lognormal Distribution

14 Computer Stuff

ISYE 6739 — Goldsman 8/5/20 100 / 108

Computer Stuff

Lesson 4.14 — Computer Stuff

Evaluating pmf’s / pdf’s and cdf’s

We can use various computer packages such as Excel, Minitab, R, SAS, etc.,to calculate pmf’s / pdf’s and cdf’s for a variety of common distributions. Forinstance, in Excel, we find the functions:

BINOMDIST = Binomial distributionEXPONDIST = ExponentialNEGBINOMDIST = Negative BinomialNORMDIST and NORMSDIST = Normal and Standard NormalPOISSON = Poisson

Functions such as NORMSINV and TINV can calculate the inverses of thestandard normal and t distributions, respectively.

ISYE 6739 — Goldsman 8/5/20 101 / 108

Computer Stuff

Lesson 4.14 — Computer Stuff

Evaluating pmf’s / pdf’s and cdf’s

We can use various computer packages such as Excel, Minitab, R, SAS, etc.,to calculate pmf’s / pdf’s and cdf’s for a variety of common distributions. Forinstance, in Excel, we find the functions:

BINOMDIST = Binomial distributionEXPONDIST = ExponentialNEGBINOMDIST = Negative BinomialNORMDIST and NORMSDIST = Normal and Standard NormalPOISSON = Poisson

Functions such as NORMSINV and TINV can calculate the inverses of thestandard normal and t distributions, respectively.

ISYE 6739 — Goldsman 8/5/20 101 / 108

Computer Stuff

Lesson 4.14 — Computer Stuff

Evaluating pmf’s / pdf’s and cdf’s

We can use various computer packages such as Excel, Minitab, R, SAS, etc.,to calculate pmf’s / pdf’s and cdf’s for a variety of common distributions. Forinstance, in Excel, we find the functions:

BINOMDIST = Binomial distributionEXPONDIST = ExponentialNEGBINOMDIST = Negative BinomialNORMDIST and NORMSDIST = Normal and Standard NormalPOISSON = Poisson

Functions such as NORMSINV and TINV can calculate the inverses of thestandard normal and t distributions, respectively.

ISYE 6739 — Goldsman 8/5/20 101 / 108

Computer Stuff

Lesson 4.14 — Computer Stuff

Evaluating pmf’s / pdf’s and cdf’s

We can use various computer packages such as Excel, Minitab, R, SAS, etc.,to calculate pmf’s / pdf’s and cdf’s for a variety of common distributions. Forinstance, in Excel, we find the functions:

BINOMDIST = Binomial distribution

EXPONDIST = ExponentialNEGBINOMDIST = Negative BinomialNORMDIST and NORMSDIST = Normal and Standard NormalPOISSON = Poisson

Functions such as NORMSINV and TINV can calculate the inverses of thestandard normal and t distributions, respectively.

ISYE 6739 — Goldsman 8/5/20 101 / 108

Computer Stuff

Lesson 4.14 — Computer Stuff

Evaluating pmf’s / pdf’s and cdf’s

We can use various computer packages such as Excel, Minitab, R, SAS, etc.,to calculate pmf’s / pdf’s and cdf’s for a variety of common distributions. Forinstance, in Excel, we find the functions:

BINOMDIST = Binomial distributionEXPONDIST = Exponential

NEGBINOMDIST = Negative BinomialNORMDIST and NORMSDIST = Normal and Standard NormalPOISSON = Poisson

Functions such as NORMSINV and TINV can calculate the inverses of thestandard normal and t distributions, respectively.

ISYE 6739 — Goldsman 8/5/20 101 / 108

Computer Stuff

Lesson 4.14 — Computer Stuff

Evaluating pmf’s / pdf’s and cdf’s

We can use various computer packages such as Excel, Minitab, R, SAS, etc.,to calculate pmf’s / pdf’s and cdf’s for a variety of common distributions. Forinstance, in Excel, we find the functions:

BINOMDIST = Binomial distributionEXPONDIST = ExponentialNEGBINOMDIST = Negative Binomial

NORMDIST and NORMSDIST = Normal and Standard NormalPOISSON = Poisson

Functions such as NORMSINV and TINV can calculate the inverses of thestandard normal and t distributions, respectively.

ISYE 6739 — Goldsman 8/5/20 101 / 108

Computer Stuff

Lesson 4.14 — Computer Stuff

Evaluating pmf’s / pdf’s and cdf’s

We can use various computer packages such as Excel, Minitab, R, SAS, etc.,to calculate pmf’s / pdf’s and cdf’s for a variety of common distributions. Forinstance, in Excel, we find the functions:

BINOMDIST = Binomial distributionEXPONDIST = ExponentialNEGBINOMDIST = Negative BinomialNORMDIST and NORMSDIST = Normal and Standard Normal

POISSON = Poisson

Functions such as NORMSINV and TINV can calculate the inverses of thestandard normal and t distributions, respectively.

ISYE 6739 — Goldsman 8/5/20 101 / 108

Computer Stuff

Lesson 4.14 — Computer Stuff

Evaluating pmf’s / pdf’s and cdf’s

We can use various computer packages such as Excel, Minitab, R, SAS, etc.,to calculate pmf’s / pdf’s and cdf’s for a variety of common distributions. Forinstance, in Excel, we find the functions:

BINOMDIST = Binomial distributionEXPONDIST = ExponentialNEGBINOMDIST = Negative BinomialNORMDIST and NORMSDIST = Normal and Standard NormalPOISSON = Poisson

Functions such as NORMSINV and TINV can calculate the inverses of thestandard normal and t distributions, respectively.

ISYE 6739 — Goldsman 8/5/20 101 / 108

Computer Stuff

Lesson 4.14 — Computer Stuff

Evaluating pmf’s / pdf’s and cdf’s

We can use various computer packages such as Excel, Minitab, R, SAS, etc.,to calculate pmf’s / pdf’s and cdf’s for a variety of common distributions. Forinstance, in Excel, we find the functions:

BINOMDIST = Binomial distributionEXPONDIST = ExponentialNEGBINOMDIST = Negative BinomialNORMDIST and NORMSDIST = Normal and Standard NormalPOISSON = Poisson

Functions such as NORMSINV and TINV can calculate the inverses of thestandard normal and t distributions, respectively.

ISYE 6739 — Goldsman 8/5/20 101 / 108

Computer Stuff

Simulating Random Variables

Motivation: Simulations are used to evaluate a variety of real-worldprocesses that contain inherent randomness, e.g., queueing systems, inventorysystems, manufacturing systems, etc. In order to run simulations, you need togenerate various RVs, e.g., arrival times, service times, failure times, etc.

Examples: There are numerous ways to simulate RVs.

The Excel function RAND simulates a Unif(0,1) RV.

If U1, U2iid∼ Unif(0,1), then U1 + U2 is Triangular(0,1,2). This is simply

RAND()+RAND() in Excel.

The Inverse Transform Theorem gives (−1/λ)`n(U) ∼ Exp(λ).

If U1, U2, . . . , Ukiid∼ Unif(0,1), then

∑ki=1(−1/λ)`n(Ui) is Erlangk(λ)

because it is the sum of iid Exp(λ)’s.

It can be shown that X = d`n(U)/`n(1− p)e ∼ Geom(p), where d·e isthe “ceiling” (integer round-up) function.

ISYE 6739 — Goldsman 8/5/20 102 / 108

Computer Stuff

Simulating Random Variables

Motivation: Simulations are used to evaluate a variety of real-worldprocesses that contain inherent randomness, e.g., queueing systems, inventorysystems, manufacturing systems, etc. In order to run simulations, you need togenerate various RVs, e.g., arrival times, service times, failure times, etc.

Examples: There are numerous ways to simulate RVs.

The Excel function RAND simulates a Unif(0,1) RV.

If U1, U2iid∼ Unif(0,1), then U1 + U2 is Triangular(0,1,2). This is simply

RAND()+RAND() in Excel.

The Inverse Transform Theorem gives (−1/λ)`n(U) ∼ Exp(λ).

If U1, U2, . . . , Ukiid∼ Unif(0,1), then

∑ki=1(−1/λ)`n(Ui) is Erlangk(λ)

because it is the sum of iid Exp(λ)’s.

It can be shown that X = d`n(U)/`n(1− p)e ∼ Geom(p), where d·e isthe “ceiling” (integer round-up) function.

ISYE 6739 — Goldsman 8/5/20 102 / 108

Computer Stuff

Simulating Random Variables

Motivation: Simulations are used to evaluate a variety of real-worldprocesses that contain inherent randomness, e.g., queueing systems, inventorysystems, manufacturing systems, etc. In order to run simulations, you need togenerate various RVs, e.g., arrival times, service times, failure times, etc.

Examples: There are numerous ways to simulate RVs.

The Excel function RAND simulates a Unif(0,1) RV.

If U1, U2iid∼ Unif(0,1), then U1 + U2 is Triangular(0,1,2). This is simply

RAND()+RAND() in Excel.

The Inverse Transform Theorem gives (−1/λ)`n(U) ∼ Exp(λ).

If U1, U2, . . . , Ukiid∼ Unif(0,1), then

∑ki=1(−1/λ)`n(Ui) is Erlangk(λ)

because it is the sum of iid Exp(λ)’s.

It can be shown that X = d`n(U)/`n(1− p)e ∼ Geom(p), where d·e isthe “ceiling” (integer round-up) function.

ISYE 6739 — Goldsman 8/5/20 102 / 108

Computer Stuff

Simulating Random Variables

Motivation: Simulations are used to evaluate a variety of real-worldprocesses that contain inherent randomness, e.g., queueing systems, inventorysystems, manufacturing systems, etc. In order to run simulations, you need togenerate various RVs, e.g., arrival times, service times, failure times, etc.

Examples: There are numerous ways to simulate RVs.

The Excel function RAND simulates a Unif(0,1) RV.

If U1, U2iid∼ Unif(0,1), then U1 + U2 is Triangular(0,1,2). This is simply

RAND()+RAND() in Excel.

The Inverse Transform Theorem gives (−1/λ)`n(U) ∼ Exp(λ).

If U1, U2, . . . , Ukiid∼ Unif(0,1), then

∑ki=1(−1/λ)`n(Ui) is Erlangk(λ)

because it is the sum of iid Exp(λ)’s.

It can be shown that X = d`n(U)/`n(1− p)e ∼ Geom(p), where d·e isthe “ceiling” (integer round-up) function.

ISYE 6739 — Goldsman 8/5/20 102 / 108

Computer Stuff

Simulating Random Variables

Motivation: Simulations are used to evaluate a variety of real-worldprocesses that contain inherent randomness, e.g., queueing systems, inventorysystems, manufacturing systems, etc. In order to run simulations, you need togenerate various RVs, e.g., arrival times, service times, failure times, etc.

Examples: There are numerous ways to simulate RVs.

The Excel function RAND simulates a Unif(0,1) RV.

If U1, U2iid∼ Unif(0,1), then U1 + U2 is Triangular(0,1,2). This is simply

RAND()+RAND() in Excel.

The Inverse Transform Theorem gives (−1/λ)`n(U) ∼ Exp(λ).

If U1, U2, . . . , Ukiid∼ Unif(0,1), then

∑ki=1(−1/λ)`n(Ui) is Erlangk(λ)

because it is the sum of iid Exp(λ)’s.

It can be shown that X = d`n(U)/`n(1− p)e ∼ Geom(p), where d·e isthe “ceiling” (integer round-up) function.

ISYE 6739 — Goldsman 8/5/20 102 / 108

Computer Stuff

Simulating Random Variables

Motivation: Simulations are used to evaluate a variety of real-worldprocesses that contain inherent randomness, e.g., queueing systems, inventorysystems, manufacturing systems, etc. In order to run simulations, you need togenerate various RVs, e.g., arrival times, service times, failure times, etc.

Examples: There are numerous ways to simulate RVs.

The Excel function RAND simulates a Unif(0,1) RV.

If U1, U2iid∼ Unif(0,1), then U1 + U2 is Triangular(0,1,2). This is simply

RAND()+RAND() in Excel.

The Inverse Transform Theorem gives (−1/λ)`n(U) ∼ Exp(λ).

If U1, U2, . . . , Ukiid∼ Unif(0,1), then

∑ki=1(−1/λ)`n(Ui) is Erlangk(λ)

because it is the sum of iid Exp(λ)’s.

It can be shown that X = d`n(U)/`n(1− p)e ∼ Geom(p), where d·e isthe “ceiling” (integer round-up) function.

ISYE 6739 — Goldsman 8/5/20 102 / 108

Computer Stuff

Simulating Random Variables

Motivation: Simulations are used to evaluate a variety of real-worldprocesses that contain inherent randomness, e.g., queueing systems, inventorysystems, manufacturing systems, etc. In order to run simulations, you need togenerate various RVs, e.g., arrival times, service times, failure times, etc.

Examples: There are numerous ways to simulate RVs.

The Excel function RAND simulates a Unif(0,1) RV.

If U1, U2iid∼ Unif(0,1), then U1 + U2 is Triangular(0,1,2). This is simply

RAND()+RAND() in Excel.

The Inverse Transform Theorem gives (−1/λ)`n(U) ∼ Exp(λ).

If U1, U2, . . . , Ukiid∼ Unif(0,1), then

∑ki=1(−1/λ)`n(Ui) is Erlangk(λ)

because it is the sum of iid Exp(λ)’s.

It can be shown that X = d`n(U)/`n(1− p)e ∼ Geom(p), where d·e isthe “ceiling” (integer round-up) function.

ISYE 6739 — Goldsman 8/5/20 102 / 108

Computer Stuff

Simulating Random Variables

Motivation: Simulations are used to evaluate a variety of real-worldprocesses that contain inherent randomness, e.g., queueing systems, inventorysystems, manufacturing systems, etc. In order to run simulations, you need togenerate various RVs, e.g., arrival times, service times, failure times, etc.

Examples: There are numerous ways to simulate RVs.

The Excel function RAND simulates a Unif(0,1) RV.

If U1, U2iid∼ Unif(0,1), then U1 + U2 is Triangular(0,1,2). This is simply

RAND()+RAND() in Excel.

The Inverse Transform Theorem gives (−1/λ)`n(U) ∼ Exp(λ).

If U1, U2, . . . , Ukiid∼ Unif(0,1), then

∑ki=1(−1/λ)`n(Ui) is Erlangk(λ)

because it is the sum of iid Exp(λ)’s.

It can be shown that X = d`n(U)/`n(1− p)e ∼ Geom(p), where d·e isthe “ceiling” (integer round-up) function.

ISYE 6739 — Goldsman 8/5/20 102 / 108

Computer Stuff

The simulation of RVs is actually the topic of another course, but we will endthis module with a remarkable method for generating normal RVs.

Theorem (Box and Muller): If U1, U2iid∼ Unif(0, 1), then

Z1 =√−2`n(U1) cos(2πU2) and Z2 =

√−2`n(U1) sin(2πU2)

are iid Nor(0,1).

Example: Suppose that U1 = 0.3 and U2 = 0.8 are realizations of two iidUnif(0,1)’s. Box–Muller gives the following two iid standard normals.

Z1 =√−2`n(U1) cos(2πU2) = 0.480

Z2 =√−2`n(U1) sin(2πU2) = − 1.476. 2

ISYE 6739 — Goldsman 8/5/20 103 / 108

Computer Stuff

The simulation of RVs is actually the topic of another course, but we will endthis module with a remarkable method for generating normal RVs.

Theorem (Box and Muller): If U1, U2iid∼ Unif(0, 1), then

Z1 =√−2`n(U1) cos(2πU2) and Z2 =

√−2`n(U1) sin(2πU2)

are iid Nor(0,1).

Example: Suppose that U1 = 0.3 and U2 = 0.8 are realizations of two iidUnif(0,1)’s. Box–Muller gives the following two iid standard normals.

Z1 =√−2`n(U1) cos(2πU2) = 0.480

Z2 =√−2`n(U1) sin(2πU2) = − 1.476. 2

ISYE 6739 — Goldsman 8/5/20 103 / 108

Computer Stuff

The simulation of RVs is actually the topic of another course, but we will endthis module with a remarkable method for generating normal RVs.

Theorem (Box and Muller): If U1, U2iid∼ Unif(0, 1), then

Z1 =√−2`n(U1) cos(2πU2) and

Z2 =√−2`n(U1) sin(2πU2)

are iid Nor(0,1).

Example: Suppose that U1 = 0.3 and U2 = 0.8 are realizations of two iidUnif(0,1)’s. Box–Muller gives the following two iid standard normals.

Z1 =√−2`n(U1) cos(2πU2) = 0.480

Z2 =√−2`n(U1) sin(2πU2) = − 1.476. 2

ISYE 6739 — Goldsman 8/5/20 103 / 108

Computer Stuff

The simulation of RVs is actually the topic of another course, but we will endthis module with a remarkable method for generating normal RVs.

Theorem (Box and Muller): If U1, U2iid∼ Unif(0, 1), then

Z1 =√−2`n(U1) cos(2πU2) and Z2 =

√−2`n(U1) sin(2πU2)

are iid Nor(0,1).

Example: Suppose that U1 = 0.3 and U2 = 0.8 are realizations of two iidUnif(0,1)’s. Box–Muller gives the following two iid standard normals.

Z1 =√−2`n(U1) cos(2πU2) = 0.480

Z2 =√−2`n(U1) sin(2πU2) = − 1.476. 2

ISYE 6739 — Goldsman 8/5/20 103 / 108

Computer Stuff

The simulation of RVs is actually the topic of another course, but we will endthis module with a remarkable method for generating normal RVs.

Theorem (Box and Muller): If U1, U2iid∼ Unif(0, 1), then

Z1 =√−2`n(U1) cos(2πU2) and Z2 =

√−2`n(U1) sin(2πU2)

are iid Nor(0,1).

Example: Suppose that U1 = 0.3 and U2 = 0.8 are realizations of two iidUnif(0,1)’s. Box–Muller gives the following two iid standard normals.

Z1 =√−2`n(U1) cos(2πU2) = 0.480

Z2 =√−2`n(U1) sin(2πU2) = − 1.476. 2

ISYE 6739 — Goldsman 8/5/20 103 / 108

Computer Stuff

The simulation of RVs is actually the topic of another course, but we will endthis module with a remarkable method for generating normal RVs.

Theorem (Box and Muller): If U1, U2iid∼ Unif(0, 1), then

Z1 =√−2`n(U1) cos(2πU2) and Z2 =

√−2`n(U1) sin(2πU2)

are iid Nor(0,1).

Example: Suppose that U1 = 0.3 and U2 = 0.8 are realizations of two iidUnif(0,1)’s. Box–Muller gives the following two iid standard normals.

Z1 =√−2`n(U1) cos(2πU2) = 0.480

Z2 =√−2`n(U1) sin(2πU2) = − 1.476. 2

ISYE 6739 — Goldsman 8/5/20 103 / 108

Computer Stuff

The simulation of RVs is actually the topic of another course, but we will endthis module with a remarkable method for generating normal RVs.

Theorem (Box and Muller): If U1, U2iid∼ Unif(0, 1), then

Z1 =√−2`n(U1) cos(2πU2) and Z2 =

√−2`n(U1) sin(2πU2)

are iid Nor(0,1).

Example: Suppose that U1 = 0.3 and U2 = 0.8 are realizations of two iidUnif(0,1)’s. Box–Muller gives the following two iid standard normals.

Z1 =√−2`n(U1) cos(2πU2) = 0.480

Z2 =√−2`n(U1) sin(2πU2) = − 1.476. 2

ISYE 6739 — Goldsman 8/5/20 103 / 108

Computer Stuff

The simulation of RVs is actually the topic of another course, but we will endthis module with a remarkable method for generating normal RVs.

Theorem (Box and Muller): If U1, U2iid∼ Unif(0, 1), then

Z1 =√−2`n(U1) cos(2πU2) and Z2 =

√−2`n(U1) sin(2πU2)

are iid Nor(0,1).

Example: Suppose that U1 = 0.3 and U2 = 0.8 are realizations of two iidUnif(0,1)’s. Box–Muller gives the following two iid standard normals.

Z1 =√−2`n(U1) cos(2πU2) = 0.480

Z2 =√−2`n(U1) sin(2πU2) = − 1.476. 2

ISYE 6739 — Goldsman 8/5/20 103 / 108

Computer Stuff

Remarks:

There are many other ways to generate Nor(0,1)’s, but this is the perhapsthe easiest.

It is essential that the cosine and sine must be calculated in radians, notdegrees.

To get X ∼ Nor(µ, σ2) from Z ∼ Nor(0, 1), just take X = µ+ σZ.

Amazingly, it’s “Muller”, not “Müller”.See https://www.youtube.com/watch?v=nntGTK2Fhb0

ISYE 6739 — Goldsman 8/5/20 104 / 108

Computer Stuff

Remarks:

There are many other ways to generate Nor(0,1)’s, but this is the perhapsthe easiest.

It is essential that the cosine and sine must be calculated in radians, notdegrees.

To get X ∼ Nor(µ, σ2) from Z ∼ Nor(0, 1), just take X = µ+ σZ.

Amazingly, it’s “Muller”, not “Müller”.See https://www.youtube.com/watch?v=nntGTK2Fhb0

ISYE 6739 — Goldsman 8/5/20 104 / 108

Computer Stuff

Remarks:

There are many other ways to generate Nor(0,1)’s, but this is the perhapsthe easiest.

It is essential that the cosine and sine must be calculated in radians, notdegrees.

To get X ∼ Nor(µ, σ2) from Z ∼ Nor(0, 1), just take X = µ+ σZ.

Amazingly, it’s “Muller”, not “Müller”.See https://www.youtube.com/watch?v=nntGTK2Fhb0

ISYE 6739 — Goldsman 8/5/20 104 / 108

Computer Stuff

Remarks:

There are many other ways to generate Nor(0,1)’s, but this is the perhapsthe easiest.

It is essential that the cosine and sine must be calculated in radians, notdegrees.

To get X ∼ Nor(µ, σ2) from Z ∼ Nor(0, 1), just take X = µ+ σZ.

Amazingly, it’s “Muller”, not “Müller”.See https://www.youtube.com/watch?v=nntGTK2Fhb0

ISYE 6739 — Goldsman 8/5/20 104 / 108

Computer Stuff

Honors Proof: We follow the method from Module 3 to calculate the jointpdf of a function of two RVs. Namely, if we can express U1 = k1(Z1, Z2) andU2 = k2(Z1, Z2) for some functions k1(·, ·) and k2(·, ·), then the joint pdf of(Z1, Z2) is given by

g(z1, z2) = fU1

(k1(z1, z2)

)fU2

(k2(z1, z2)

) ∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣=

∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣ (U1 and U2 are iid Unif(0,1)

).

In order to obtain the functions k1(Z1, Z2) and k2(Z1, Z2), note that

Z21 + Z2

2 = − 2 `n(U1)[

cos2(2πU2) + sin2(2πU2)]

= − 2 `n(U1),

so thatU1 = e−(Z

21+Z

22 )/2.

ISYE 6739 — Goldsman 8/5/20 105 / 108

Computer Stuff

Honors Proof: We follow the method from Module 3 to calculate the jointpdf of a function of two RVs. Namely, if we can express U1 = k1(Z1, Z2) andU2 = k2(Z1, Z2) for some functions k1(·, ·) and k2(·, ·), then the joint pdf of(Z1, Z2) is given by

g(z1, z2) = fU1

(k1(z1, z2)

)fU2

(k2(z1, z2)

) ∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣

=

∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣ (U1 and U2 are iid Unif(0,1)

).

In order to obtain the functions k1(Z1, Z2) and k2(Z1, Z2), note that

Z21 + Z2

2 = − 2 `n(U1)[

cos2(2πU2) + sin2(2πU2)]

= − 2 `n(U1),

so thatU1 = e−(Z

21+Z

22 )/2.

ISYE 6739 — Goldsman 8/5/20 105 / 108

Computer Stuff

Honors Proof: We follow the method from Module 3 to calculate the jointpdf of a function of two RVs. Namely, if we can express U1 = k1(Z1, Z2) andU2 = k2(Z1, Z2) for some functions k1(·, ·) and k2(·, ·), then the joint pdf of(Z1, Z2) is given by

g(z1, z2) = fU1

(k1(z1, z2)

)fU2

(k2(z1, z2)

) ∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣=

∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣ (U1 and U2 are iid Unif(0,1)

).

In order to obtain the functions k1(Z1, Z2) and k2(Z1, Z2), note that

Z21 + Z2

2 = − 2 `n(U1)[

cos2(2πU2) + sin2(2πU2)]

= − 2 `n(U1),

so thatU1 = e−(Z

21+Z

22 )/2.

ISYE 6739 — Goldsman 8/5/20 105 / 108

Computer Stuff

Honors Proof: We follow the method from Module 3 to calculate the jointpdf of a function of two RVs. Namely, if we can express U1 = k1(Z1, Z2) andU2 = k2(Z1, Z2) for some functions k1(·, ·) and k2(·, ·), then the joint pdf of(Z1, Z2) is given by

g(z1, z2) = fU1

(k1(z1, z2)

)fU2

(k2(z1, z2)

) ∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣=

∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣ (U1 and U2 are iid Unif(0,1)

).

In order to obtain the functions k1(Z1, Z2) and k2(Z1, Z2), note that

Z21 + Z2

2 = − 2 `n(U1)[

cos2(2πU2) + sin2(2πU2)]

= − 2 `n(U1),

so thatU1 = e−(Z

21+Z

22 )/2.

ISYE 6739 — Goldsman 8/5/20 105 / 108

Computer Stuff

Honors Proof: We follow the method from Module 3 to calculate the jointpdf of a function of two RVs. Namely, if we can express U1 = k1(Z1, Z2) andU2 = k2(Z1, Z2) for some functions k1(·, ·) and k2(·, ·), then the joint pdf of(Z1, Z2) is given by

g(z1, z2) = fU1

(k1(z1, z2)

)fU2

(k2(z1, z2)

) ∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣=

∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣ (U1 and U2 are iid Unif(0,1)

).

In order to obtain the functions k1(Z1, Z2) and k2(Z1, Z2), note that

Z21 + Z2

2 = − 2 `n(U1)[

cos2(2πU2) + sin2(2πU2)]

= − 2 `n(U1),

so thatU1 = e−(Z

21+Z

22 )/2.

ISYE 6739 — Goldsman 8/5/20 105 / 108

Computer Stuff

Honors Proof: We follow the method from Module 3 to calculate the jointpdf of a function of two RVs. Namely, if we can express U1 = k1(Z1, Z2) andU2 = k2(Z1, Z2) for some functions k1(·, ·) and k2(·, ·), then the joint pdf of(Z1, Z2) is given by

g(z1, z2) = fU1

(k1(z1, z2)

)fU2

(k2(z1, z2)

) ∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣=

∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣ (U1 and U2 are iid Unif(0,1)

).

In order to obtain the functions k1(Z1, Z2) and k2(Z1, Z2), note that

Z21 + Z2

2 = − 2 `n(U1)[

cos2(2πU2) + sin2(2πU2)]

= − 2 `n(U1),

so thatU1 = e−(Z

21+Z

22 )/2.

ISYE 6739 — Goldsman 8/5/20 105 / 108

Computer Stuff

This immediately implies that

Z21 = −2 `n(U1) cos2(2πU2)

= −2 `n(e−(Z

21+Z

22 )/2)

cos2(2πU2)

= (Z21 + Z2

2 ) cos2(2πU2),

so that

U2 =1

2πarccos

√Z21

Z21 + Z2

2

)=

1

2πarccos

(√Z21

Z21 + Z2

2

),

where we (non-rigorously) get rid of the “±” to balance off the fact that therange of y = arccos(x) is only regarded to be 0 ≤ y ≤ π (not 0 ≤ y ≤ 2π).

ISYE 6739 — Goldsman 8/5/20 106 / 108

Computer Stuff

This immediately implies that

Z21 = −2 `n(U1) cos2(2πU2)

= −2 `n(e−(Z

21+Z

22 )/2)

cos2(2πU2)

= (Z21 + Z2

2 ) cos2(2πU2),

so that

U2 =1

2πarccos

√Z21

Z21 + Z2

2

)=

1

2πarccos

(√Z21

Z21 + Z2

2

),

where we (non-rigorously) get rid of the “±” to balance off the fact that therange of y = arccos(x) is only regarded to be 0 ≤ y ≤ π (not 0 ≤ y ≤ 2π).

ISYE 6739 — Goldsman 8/5/20 106 / 108

Computer Stuff

This immediately implies that

Z21 = −2 `n(U1) cos2(2πU2)

= −2 `n(e−(Z

21+Z

22 )/2)

cos2(2πU2)

= (Z21 + Z2

2 ) cos2(2πU2),

so that

U2 =1

2πarccos

√Z21

Z21 + Z2

2

)=

1

2πarccos

(√Z21

Z21 + Z2

2

),

where we (non-rigorously) get rid of the “±” to balance off the fact that therange of y = arccos(x) is only regarded to be 0 ≤ y ≤ π (not 0 ≤ y ≤ 2π).

ISYE 6739 — Goldsman 8/5/20 106 / 108

Computer Stuff

This immediately implies that

Z21 = −2 `n(U1) cos2(2πU2)

= −2 `n(e−(Z

21+Z

22 )/2)

cos2(2πU2)

= (Z21 + Z2

2 ) cos2(2πU2),

so that

U2 =1

2πarccos

√Z21

Z21 + Z2

2

)=

1

2πarccos

(√Z21

Z21 + Z2

2

),

where we (non-rigorously) get rid of the “±” to balance off the fact that therange of y = arccos(x) is only regarded to be 0 ≤ y ≤ π (not 0 ≤ y ≤ 2π).

ISYE 6739 — Goldsman 8/5/20 106 / 108

Computer Stuff

This immediately implies that

Z21 = −2 `n(U1) cos2(2πU2)

= −2 `n(e−(Z

21+Z

22 )/2)

cos2(2πU2)

= (Z21 + Z2

2 ) cos2(2πU2),

so that

U2 =1

2πarccos

√Z21

Z21 + Z2

2

)=

1

2πarccos

(√Z21

Z21 + Z2

2

),

where we (non-rigorously) get rid of the “±” to balance off the fact that therange of y = arccos(x) is only regarded to be 0 ≤ y ≤ π (not 0 ≤ y ≤ 2π).

ISYE 6739 — Goldsman 8/5/20 106 / 108

Computer Stuff

Now some derivative fun:

∂u1∂zi

=∂

∂zie−(z

21+z

22)/2 = − zi e−(z

21+z

22)/2, i = 1, 2,

∂u2∂z1

=∂

∂z1

1

2πarccos

(√z21

z21 + z22

)

=1

−1√1− z21

z21+z22

∂z1

√z21

z21 + z22(chain rule)

=1

−1√z22

z21+z22

1

2

(z21

z21 + z22

)−1/2 ∂

∂z1

z21z21 + z22

(chain rule again)

=−(z21 + z22)

4πz1z2

2z1z22

(z21 + z22)2=

−z22π(z21 + z22)

, and

∂u2∂z2

=z1

2π(z21 + z22)(after similar algebra).

ISYE 6739 — Goldsman 8/5/20 107 / 108

Computer Stuff

Now some derivative fun:∂u1∂zi

=∂

∂zie−(z

21+z

22)/2

= − zi e−(z21+z

22)/2, i = 1, 2,

∂u2∂z1

=∂

∂z1

1

2πarccos

(√z21

z21 + z22

)

=1

−1√1− z21

z21+z22

∂z1

√z21

z21 + z22(chain rule)

=1

−1√z22

z21+z22

1

2

(z21

z21 + z22

)−1/2 ∂

∂z1

z21z21 + z22

(chain rule again)

=−(z21 + z22)

4πz1z2

2z1z22

(z21 + z22)2=

−z22π(z21 + z22)

, and

∂u2∂z2

=z1

2π(z21 + z22)(after similar algebra).

ISYE 6739 — Goldsman 8/5/20 107 / 108

Computer Stuff

Now some derivative fun:∂u1∂zi

=∂

∂zie−(z

21+z

22)/2 = − zi e−(z

21+z

22)/2, i = 1, 2,

∂u2∂z1

=∂

∂z1

1

2πarccos

(√z21

z21 + z22

)

=1

−1√1− z21

z21+z22

∂z1

√z21

z21 + z22(chain rule)

=1

−1√z22

z21+z22

1

2

(z21

z21 + z22

)−1/2 ∂

∂z1

z21z21 + z22

(chain rule again)

=−(z21 + z22)

4πz1z2

2z1z22

(z21 + z22)2=

−z22π(z21 + z22)

, and

∂u2∂z2

=z1

2π(z21 + z22)(after similar algebra).

ISYE 6739 — Goldsman 8/5/20 107 / 108

Computer Stuff

Now some derivative fun:∂u1∂zi

=∂

∂zie−(z

21+z

22)/2 = − zi e−(z

21+z

22)/2, i = 1, 2,

∂u2∂z1

=∂

∂z1

1

2πarccos

(√z21

z21 + z22

)

=1

−1√1− z21

z21+z22

∂z1

√z21

z21 + z22(chain rule)

=1

−1√z22

z21+z22

1

2

(z21

z21 + z22

)−1/2 ∂

∂z1

z21z21 + z22

(chain rule again)

=−(z21 + z22)

4πz1z2

2z1z22

(z21 + z22)2=

−z22π(z21 + z22)

, and

∂u2∂z2

=z1

2π(z21 + z22)(after similar algebra).

ISYE 6739 — Goldsman 8/5/20 107 / 108

Computer Stuff

Now some derivative fun:∂u1∂zi

=∂

∂zie−(z

21+z

22)/2 = − zi e−(z

21+z

22)/2, i = 1, 2,

∂u2∂z1

=∂

∂z1

1

2πarccos

(√z21

z21 + z22

)

=1

−1√1− z21

z21+z22

∂z1

√z21

z21 + z22(chain rule)

=1

−1√z22

z21+z22

1

2

(z21

z21 + z22

)−1/2 ∂

∂z1

z21z21 + z22

(chain rule again)

=−(z21 + z22)

4πz1z2

2z1z22

(z21 + z22)2=

−z22π(z21 + z22)

, and

∂u2∂z2

=z1

2π(z21 + z22)(after similar algebra).

ISYE 6739 — Goldsman 8/5/20 107 / 108

Computer Stuff

Now some derivative fun:∂u1∂zi

=∂

∂zie−(z

21+z

22)/2 = − zi e−(z

21+z

22)/2, i = 1, 2,

∂u2∂z1

=∂

∂z1

1

2πarccos

(√z21

z21 + z22

)

=1

−1√1− z21

z21+z22

∂z1

√z21

z21 + z22(chain rule)

=1

−1√z22

z21+z22

1

2

(z21

z21 + z22

)−1/2 ∂

∂z1

z21z21 + z22

(chain rule again)

=−(z21 + z22)

4πz1z2

2z1z22

(z21 + z22)2=

−z22π(z21 + z22)

, and

∂u2∂z2

=z1

2π(z21 + z22)(after similar algebra).

ISYE 6739 — Goldsman 8/5/20 107 / 108

Computer Stuff

Now some derivative fun:∂u1∂zi

=∂

∂zie−(z

21+z

22)/2 = − zi e−(z

21+z

22)/2, i = 1, 2,

∂u2∂z1

=∂

∂z1

1

2πarccos

(√z21

z21 + z22

)

=1

−1√1− z21

z21+z22

∂z1

√z21

z21 + z22(chain rule)

=1

−1√z22

z21+z22

1

2

(z21

z21 + z22

)−1/2 ∂

∂z1

z21z21 + z22

(chain rule again)

=−(z21 + z22)

4πz1z2

2z1z22

(z21 + z22)2

=−z2

2π(z21 + z22), and

∂u2∂z2

=z1

2π(z21 + z22)(after similar algebra).

ISYE 6739 — Goldsman 8/5/20 107 / 108

Computer Stuff

Now some derivative fun:∂u1∂zi

=∂

∂zie−(z

21+z

22)/2 = − zi e−(z

21+z

22)/2, i = 1, 2,

∂u2∂z1

=∂

∂z1

1

2πarccos

(√z21

z21 + z22

)

=1

−1√1− z21

z21+z22

∂z1

√z21

z21 + z22(chain rule)

=1

−1√z22

z21+z22

1

2

(z21

z21 + z22

)−1/2 ∂

∂z1

z21z21 + z22

(chain rule again)

=−(z21 + z22)

4πz1z2

2z1z22

(z21 + z22)2=

−z22π(z21 + z22)

, and

∂u2∂z2

=z1

2π(z21 + z22)(after similar algebra).

ISYE 6739 — Goldsman 8/5/20 107 / 108

Computer Stuff

Now some derivative fun:∂u1∂zi

=∂

∂zie−(z

21+z

22)/2 = − zi e−(z

21+z

22)/2, i = 1, 2,

∂u2∂z1

=∂

∂z1

1

2πarccos

(√z21

z21 + z22

)

=1

−1√1− z21

z21+z22

∂z1

√z21

z21 + z22(chain rule)

=1

−1√z22

z21+z22

1

2

(z21

z21 + z22

)−1/2 ∂

∂z1

z21z21 + z22

(chain rule again)

=−(z21 + z22)

4πz1z2

2z1z22

(z21 + z22)2=

−z22π(z21 + z22)

, and

∂u2∂z2

=z1

2π(z21 + z22)(after similar algebra).

ISYE 6739 — Goldsman 8/5/20 107 / 108

Computer Stuff

Then we finally have

g(z1, z2) =

∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣=

∣∣∣∣−z1 e−(z21+z22)/2 z12π(z21 + z22)

− z22π(z21 + z22)

z2 e−(z21+z22)/2

∣∣∣∣=

1

2πe−(z

21+z

22)/2 =

(1√2π

e−z21/2

)(1√2π

e−z22/2

),

which is the product of two iid Nor(0,1) pdf’s, so we are done!

ISYE 6739 — Goldsman 8/5/20 108 / 108

Computer Stuff

Then we finally have

g(z1, z2) =

∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣

=

∣∣∣∣−z1 e−(z21+z22)/2 z12π(z21 + z22)

− z22π(z21 + z22)

z2 e−(z21+z22)/2

∣∣∣∣=

1

2πe−(z

21+z

22)/2 =

(1√2π

e−z21/2

)(1√2π

e−z22/2

),

which is the product of two iid Nor(0,1) pdf’s, so we are done!

ISYE 6739 — Goldsman 8/5/20 108 / 108

Computer Stuff

Then we finally have

g(z1, z2) =

∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣=

∣∣∣∣−z1 e−(z21+z22)/2 z12π(z21 + z22)

− z22π(z21 + z22)

z2 e−(z21+z22)/2

∣∣∣∣

=1

2πe−(z

21+z

22)/2 =

(1√2π

e−z21/2

)(1√2π

e−z22/2

),

which is the product of two iid Nor(0,1) pdf’s, so we are done!

ISYE 6739 — Goldsman 8/5/20 108 / 108

Computer Stuff

Then we finally have

g(z1, z2) =

∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣=

∣∣∣∣−z1 e−(z21+z22)/2 z12π(z21 + z22)

− z22π(z21 + z22)

z2 e−(z21+z22)/2

∣∣∣∣=

1

2πe−(z

21+z

22)/2

=

(1√2π

e−z21/2

)(1√2π

e−z22/2

),

which is the product of two iid Nor(0,1) pdf’s, so we are done!

ISYE 6739 — Goldsman 8/5/20 108 / 108

Computer Stuff

Then we finally have

g(z1, z2) =

∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣=

∣∣∣∣−z1 e−(z21+z22)/2 z12π(z21 + z22)

− z22π(z21 + z22)

z2 e−(z21+z22)/2

∣∣∣∣=

1

2πe−(z

21+z

22)/2 =

(1√2π

e−z21/2

)(1√2π

e−z22/2

),

which is the product of two iid Nor(0,1) pdf’s, so we are done!

ISYE 6739 — Goldsman 8/5/20 108 / 108

Computer Stuff

Then we finally have

g(z1, z2) =

∣∣∣∣∂u1∂z1

∂u2∂z2− ∂u2∂z1

∂u1∂z2

∣∣∣∣=

∣∣∣∣−z1 e−(z21+z22)/2 z12π(z21 + z22)

− z22π(z21 + z22)

z2 e−(z21+z22)/2

∣∣∣∣=

1

2πe−(z

21+z

22)/2 =

(1√2π

e−z21/2

)(1√2π

e−z22/2

),

which is the product of two iid Nor(0,1) pdf’s, so we are done!

ISYE 6739 — Goldsman 8/5/20 108 / 108