stat 302 - university of british columbiaruben/stat302website/lecture12.pdf · stat 302 ruben zamar...

191
Stat 302 Ruben Zamar [email protected] Asymptotic Results Ruben Zamar [email protected] Module 12 Asymptotic Results 1 / 60

Upload: phungque

Post on 31-Mar-2018

215 views

Category:

Documents


2 download

TRANSCRIPT

Stat 302

Ruben [email protected]

Asymptotic Results

Ruben Zamar [email protected] () Module 12 Asymptotic Results 1 / 60

Motivation

Suppose we are interested on the overall performance of UBCstudents in the Stat 302 Final Exam.

Some questions we may have are:

What is the mean performance across all UBC students?Do women do better than men on average?Do CPSC students do better than other FOSC students?How likely is for a student to fail this test? Does this prob changeacross disciplines?

Ruben Zamar [email protected] () Module 12 Asymptotic Results 2 / 60

Motivation

Suppose we are interested on the overall performance of UBCstudents in the Stat 302 Final Exam.

Some questions we may have are:

What is the mean performance across all UBC students?Do women do better than men on average?Do CPSC students do better than other FOSC students?How likely is for a student to fail this test? Does this prob changeacross disciplines?

Ruben Zamar [email protected] () Module 12 Asymptotic Results 2 / 60

Motivation

Suppose we are interested on the overall performance of UBCstudents in the Stat 302 Final Exam.

Some questions we may have are:

What is the mean performance across all UBC students?

Do women do better than men on average?Do CPSC students do better than other FOSC students?How likely is for a student to fail this test? Does this prob changeacross disciplines?

Ruben Zamar [email protected] () Module 12 Asymptotic Results 2 / 60

Motivation

Suppose we are interested on the overall performance of UBCstudents in the Stat 302 Final Exam.

Some questions we may have are:

What is the mean performance across all UBC students?Do women do better than men on average?

Do CPSC students do better than other FOSC students?How likely is for a student to fail this test? Does this prob changeacross disciplines?

Ruben Zamar [email protected] () Module 12 Asymptotic Results 2 / 60

Motivation

Suppose we are interested on the overall performance of UBCstudents in the Stat 302 Final Exam.

Some questions we may have are:

What is the mean performance across all UBC students?Do women do better than men on average?Do CPSC students do better than other FOSC students?

How likely is for a student to fail this test? Does this prob changeacross disciplines?

Ruben Zamar [email protected] () Module 12 Asymptotic Results 2 / 60

Motivation

Suppose we are interested on the overall performance of UBCstudents in the Stat 302 Final Exam.

Some questions we may have are:

What is the mean performance across all UBC students?Do women do better than men on average?Do CPSC students do better than other FOSC students?How likely is for a student to fail this test? Does this prob changeacross disciplines?

Ruben Zamar [email protected] () Module 12 Asymptotic Results 2 / 60

Motivation (Continued)

Suppose we measure the independent performances of n = 65students (e.g. in April 12...).

The student’s performances Xi could be model as iid rv’s with(unknown) mean µ and variance σ2.

A quantity of possible interest is

P (|X − µ| < 5) =?

If this probability is large then we are confident that X estimates µwithin a 5 points error margin

Ruben Zamar [email protected] () Module 12 Asymptotic Results 3 / 60

Motivation (Continued)

Suppose we measure the independent performances of n = 65students (e.g. in April 12...).

The student’s performances Xi could be model as iid rv’s with(unknown) mean µ and variance σ2.

A quantity of possible interest is

P (|X − µ| < 5) =?

If this probability is large then we are confident that X estimates µwithin a 5 points error margin

Ruben Zamar [email protected] () Module 12 Asymptotic Results 3 / 60

Motivation (Continued)

Suppose we measure the independent performances of n = 65students (e.g. in April 12...).

The student’s performances Xi could be model as iid rv’s with(unknown) mean µ and variance σ2.

A quantity of possible interest is

P (|X − µ| < 5) =?

If this probability is large then we are confident that X estimates µwithin a 5 points error margin

Ruben Zamar [email protected] () Module 12 Asymptotic Results 3 / 60

Motivation (Continued)

Suppose we measure the independent performances of n = 65students (e.g. in April 12...).

The student’s performances Xi could be model as iid rv’s with(unknown) mean µ and variance σ2.

A quantity of possible interest is

P (|X − µ| < 5) =?

If this probability is large then we are confident that X estimates µwithin a 5 points error margin

Ruben Zamar [email protected] () Module 12 Asymptotic Results 3 / 60

Confidence Intervals

Other examples of quantities of possible interest are

P (L (X1, ...,Xn) < µ < U (X1, ...,Xn)) = 0.95?where L (X1, ...,Xn) and U (X1, ...,Xn) are some functions (calledestimates) that only depend on the data

P (L (X1, ...,Xn) < µW − µM < U (X1, ...,Xn)) = 0.95?

Unfortunately we cannot compute these probabilities exactly becausewe don’t know the actual distribution of the Xi .

Even if we knew the cdf of Xi , it may be inconvenient or unfeasible tocalculate the exact probabilities when L (X1, ...,Xn) andU (X1, ...,Xn) are complicated (non-linear) functions.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 4 / 60

Confidence Intervals

Other examples of quantities of possible interest are

P (L (X1, ...,Xn) < µ < U (X1, ...,Xn)) = 0.95?where L (X1, ...,Xn) and U (X1, ...,Xn) are some functions (calledestimates) that only depend on the data

P (L (X1, ...,Xn) < µW − µM < U (X1, ...,Xn)) = 0.95?

Unfortunately we cannot compute these probabilities exactly becausewe don’t know the actual distribution of the Xi .

Even if we knew the cdf of Xi , it may be inconvenient or unfeasible tocalculate the exact probabilities when L (X1, ...,Xn) andU (X1, ...,Xn) are complicated (non-linear) functions.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 4 / 60

Confidence Intervals

Other examples of quantities of possible interest are

P (L (X1, ...,Xn) < µ < U (X1, ...,Xn)) = 0.95?where L (X1, ...,Xn) and U (X1, ...,Xn) are some functions (calledestimates) that only depend on the data

P (L (X1, ...,Xn) < µW − µM < U (X1, ...,Xn)) = 0.95?

Unfortunately we cannot compute these probabilities exactly becausewe don’t know the actual distribution of the Xi .

Even if we knew the cdf of Xi , it may be inconvenient or unfeasible tocalculate the exact probabilities when L (X1, ...,Xn) andU (X1, ...,Xn) are complicated (non-linear) functions.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 4 / 60

Confidence Intervals

Other examples of quantities of possible interest are

P (L (X1, ...,Xn) < µ < U (X1, ...,Xn)) = 0.95?where L (X1, ...,Xn) and U (X1, ...,Xn) are some functions (calledestimates) that only depend on the data

P (L (X1, ...,Xn) < µW − µM < U (X1, ...,Xn)) = 0.95?

Unfortunately we cannot compute these probabilities exactly becausewe don’t know the actual distribution of the Xi .

Even if we knew the cdf of Xi , it may be inconvenient or unfeasible tocalculate the exact probabilities when L (X1, ...,Xn) andU (X1, ...,Xn) are complicated (non-linear) functions.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 4 / 60

Confidence Intervals

Other examples of quantities of possible interest are

P (L (X1, ...,Xn) < µ < U (X1, ...,Xn)) = 0.95?where L (X1, ...,Xn) and U (X1, ...,Xn) are some functions (calledestimates) that only depend on the data

P (L (X1, ...,Xn) < µW − µM < U (X1, ...,Xn)) = 0.95?

Unfortunately we cannot compute these probabilities exactly becausewe don’t know the actual distribution of the Xi .

Even if we knew the cdf of Xi , it may be inconvenient or unfeasible tocalculate the exact probabilities when L (X1, ...,Xn) andU (X1, ...,Xn) are complicated (non-linear) functions.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 4 / 60

Asymptotic Approximation

We will learn in this course how to approximate these quantities whenn is “large”

These approximations are called “asymptotic calculations”

They are obtained by taking limit for n→ ∞The limiting results often give good approximation for moderatevalues of n (e.g. n = 20)

We will see that the standard normal cdf Φ (z) is a main tool forasymptotic calculations

Ruben Zamar [email protected] () Module 12 Asymptotic Results 5 / 60

Asymptotic Approximation

We will learn in this course how to approximate these quantities whenn is “large”

These approximations are called “asymptotic calculations”

They are obtained by taking limit for n→ ∞The limiting results often give good approximation for moderatevalues of n (e.g. n = 20)

We will see that the standard normal cdf Φ (z) is a main tool forasymptotic calculations

Ruben Zamar [email protected] () Module 12 Asymptotic Results 5 / 60

Asymptotic Approximation

We will learn in this course how to approximate these quantities whenn is “large”

These approximations are called “asymptotic calculations”

They are obtained by taking limit for n→ ∞

The limiting results often give good approximation for moderatevalues of n (e.g. n = 20)

We will see that the standard normal cdf Φ (z) is a main tool forasymptotic calculations

Ruben Zamar [email protected] () Module 12 Asymptotic Results 5 / 60

Asymptotic Approximation

We will learn in this course how to approximate these quantities whenn is “large”

These approximations are called “asymptotic calculations”

They are obtained by taking limit for n→ ∞The limiting results often give good approximation for moderatevalues of n (e.g. n = 20)

We will see that the standard normal cdf Φ (z) is a main tool forasymptotic calculations

Ruben Zamar [email protected] () Module 12 Asymptotic Results 5 / 60

Asymptotic Approximation

We will learn in this course how to approximate these quantities whenn is “large”

These approximations are called “asymptotic calculations”

They are obtained by taking limit for n→ ∞The limiting results often give good approximation for moderatevalues of n (e.g. n = 20)

We will see that the standard normal cdf Φ (z) is a main tool forasymptotic calculations

Ruben Zamar [email protected] () Module 12 Asymptotic Results 5 / 60

The General Setting

Let Xn be a sequence of random variables

Xn ∼ Fn, n = 1, 2, ...

There are several ways in which the sequence Xn may approach a“target” random variable X ∼ F as n→ ∞.Sometimes the target random variable is in fact a constant c

i.e. F (x) = I[c ,∞) (x) and so X = c with probability 1

Ruben Zamar [email protected] () Module 12 Asymptotic Results 6 / 60

The General Setting

Let Xn be a sequence of random variables

Xn ∼ Fn, n = 1, 2, ...

There are several ways in which the sequence Xn may approach a“target” random variable X ∼ F as n→ ∞.

Sometimes the target random variable is in fact a constant c

i.e. F (x) = I[c ,∞) (x) and so X = c with probability 1

Ruben Zamar [email protected] () Module 12 Asymptotic Results 6 / 60

The General Setting

Let Xn be a sequence of random variables

Xn ∼ Fn, n = 1, 2, ...

There are several ways in which the sequence Xn may approach a“target” random variable X ∼ F as n→ ∞.Sometimes the target random variable is in fact a constant c

i.e. F (x) = I[c ,∞) (x) and so X = c with probability 1

Ruben Zamar [email protected] () Module 12 Asymptotic Results 6 / 60

The General Setting

Let Xn be a sequence of random variables

Xn ∼ Fn, n = 1, 2, ...

There are several ways in which the sequence Xn may approach a“target” random variable X ∼ F as n→ ∞.Sometimes the target random variable is in fact a constant c

i.e. F (x) = I[c ,∞) (x) and so X = c with probability 1

Ruben Zamar [email protected] () Module 12 Asymptotic Results 6 / 60

Different Types of Convergence

An (incomplete) list of different types of convergence:

(1) Convergence with Probability One (Almost Sure Convergence)(2) Convergence in Probability(3) Convergence in Distribution(4) Convergence in Quadratic Mean

Ruben Zamar [email protected] () Module 12 Asymptotic Results 7 / 60

Almost Sure Convergence

Suppose that the rv’s Xn and X are all defined on the same samplespace Ω.

Notation:Xn → X a.s.

means “Xn converges almost surely to X as n→ ∞”Definition: Xn → X a.s. if

P(w : lim

n→∞Xn (w) = X (w)

)= 1

This is a very strong type of convergence.

a.s. convergence implies ( no proved here)

convergence in probability andconvergence in distribution.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 8 / 60

Almost Sure Convergence

Suppose that the rv’s Xn and X are all defined on the same samplespace Ω.Notation:

Xn → X a.s.

means “Xn converges almost surely to X as n→ ∞”

Definition: Xn → X a.s. if

P(w : lim

n→∞Xn (w) = X (w)

)= 1

This is a very strong type of convergence.

a.s. convergence implies ( no proved here)

convergence in probability andconvergence in distribution.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 8 / 60

Almost Sure Convergence

Suppose that the rv’s Xn and X are all defined on the same samplespace Ω.Notation:

Xn → X a.s.

means “Xn converges almost surely to X as n→ ∞”Definition: Xn → X a.s. if

P(w : lim

n→∞Xn (w) = X (w)

)= 1

This is a very strong type of convergence.

a.s. convergence implies ( no proved here)

convergence in probability andconvergence in distribution.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 8 / 60

Almost Sure Convergence

Suppose that the rv’s Xn and X are all defined on the same samplespace Ω.Notation:

Xn → X a.s.

means “Xn converges almost surely to X as n→ ∞”Definition: Xn → X a.s. if

P(w : lim

n→∞Xn (w) = X (w)

)= 1

This is a very strong type of convergence.

a.s. convergence implies ( no proved here)

convergence in probability andconvergence in distribution.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 8 / 60

Almost Sure Convergence

Suppose that the rv’s Xn and X are all defined on the same samplespace Ω.Notation:

Xn → X a.s.

means “Xn converges almost surely to X as n→ ∞”Definition: Xn → X a.s. if

P(w : lim

n→∞Xn (w) = X (w)

)= 1

This is a very strong type of convergence.

a.s. convergence implies ( no proved here)

convergence in probability andconvergence in distribution.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 8 / 60

Almost Sure Convergence

Suppose that the rv’s Xn and X are all defined on the same samplespace Ω.Notation:

Xn → X a.s.

means “Xn converges almost surely to X as n→ ∞”Definition: Xn → X a.s. if

P(w : lim

n→∞Xn (w) = X (w)

)= 1

This is a very strong type of convergence.

a.s. convergence implies ( no proved here)

convergence in probability and

convergence in distribution.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 8 / 60

Almost Sure Convergence

Suppose that the rv’s Xn and X are all defined on the same samplespace Ω.Notation:

Xn → X a.s.

means “Xn converges almost surely to X as n→ ∞”Definition: Xn → X a.s. if

P(w : lim

n→∞Xn (w) = X (w)

)= 1

This is a very strong type of convergence.

a.s. convergence implies ( no proved here)

convergence in probability andconvergence in distribution.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 8 / 60

The Strong Law of Large Numbers (SLLN)

Suppose that the random variables Xn are independent and all havethe same mean µ

The SLLN states that

1n

n

∑i=1Xi → µ a.s.

as n→ ∞.In this case the limiting rv is a constant: the common mean µ

Ruben Zamar [email protected] () Module 12 Asymptotic Results 9 / 60

The Strong Law of Large Numbers (SLLN)

Suppose that the random variables Xn are independent and all havethe same mean µ

The SLLN states that

1n

n

∑i=1Xi → µ a.s.

as n→ ∞.

In this case the limiting rv is a constant: the common mean µ

Ruben Zamar [email protected] () Module 12 Asymptotic Results 9 / 60

The Strong Law of Large Numbers (SLLN)

Suppose that the random variables Xn are independent and all havethe same mean µ

The SLLN states that

1n

n

∑i=1Xi → µ a.s.

as n→ ∞.In this case the limiting rv is a constant: the common mean µ

Ruben Zamar [email protected] () Module 12 Asymptotic Results 9 / 60

Examples

Example 1: Suppose that Xn ∼ Binom(n, p)

Notice that Xn = Y1 + Y2 + · · ·+ Yn , the Yi are iid Bernoulli(1, p)E (Yi ) = p, for all iIn this case E (Xn) = np, for all nBy the SLLN

pn =Xnn=Y1 + Y2 + · · ·+ Yn

n→ p a.s.

pn is called the "sample proportion”and p is called the “populationproportion”

Ruben Zamar [email protected] () Module 12 Asymptotic Results 10 / 60

Examples

Example 1: Suppose that Xn ∼ Binom(n, p)

Notice that Xn = Y1 + Y2 + · · ·+ Yn , the Yi are iid Bernoulli(1, p)

E (Yi ) = p, for all iIn this case E (Xn) = np, for all nBy the SLLN

pn =Xnn=Y1 + Y2 + · · ·+ Yn

n→ p a.s.

pn is called the "sample proportion”and p is called the “populationproportion”

Ruben Zamar [email protected] () Module 12 Asymptotic Results 10 / 60

Examples

Example 1: Suppose that Xn ∼ Binom(n, p)

Notice that Xn = Y1 + Y2 + · · ·+ Yn , the Yi are iid Bernoulli(1, p)E (Yi ) = p, for all i

In this case E (Xn) = np, for all nBy the SLLN

pn =Xnn=Y1 + Y2 + · · ·+ Yn

n→ p a.s.

pn is called the "sample proportion”and p is called the “populationproportion”

Ruben Zamar [email protected] () Module 12 Asymptotic Results 10 / 60

Examples

Example 1: Suppose that Xn ∼ Binom(n, p)

Notice that Xn = Y1 + Y2 + · · ·+ Yn , the Yi are iid Bernoulli(1, p)E (Yi ) = p, for all iIn this case E (Xn) = np, for all n

By the SLLN

pn =Xnn=Y1 + Y2 + · · ·+ Yn

n→ p a.s.

pn is called the "sample proportion”and p is called the “populationproportion”

Ruben Zamar [email protected] () Module 12 Asymptotic Results 10 / 60

Examples

Example 1: Suppose that Xn ∼ Binom(n, p)

Notice that Xn = Y1 + Y2 + · · ·+ Yn , the Yi are iid Bernoulli(1, p)E (Yi ) = p, for all iIn this case E (Xn) = np, for all nBy the SLLN

pn =Xnn=Y1 + Y2 + · · ·+ Yn

n→ p a.s.

pn is called the "sample proportion”and p is called the “populationproportion”

Ruben Zamar [email protected] () Module 12 Asymptotic Results 10 / 60

Examples

Example 1: Suppose that Xn ∼ Binom(n, p)

Notice that Xn = Y1 + Y2 + · · ·+ Yn , the Yi are iid Bernoulli(1, p)E (Yi ) = p, for all iIn this case E (Xn) = np, for all nBy the SLLN

pn =Xnn=Y1 + Y2 + · · ·+ Yn

n→ p a.s.

pn is called the "sample proportion”and p is called the “populationproportion”

Ruben Zamar [email protected] () Module 12 Asymptotic Results 10 / 60

a.s. Convergence and Continuity

Suppose that Xn → X a.s. and g (x) is a continuous function

It’s enough for g(x) to be continuous on the range of X

We can (easily) show that

g (Xn)→ g (X ) a.s.

Proof: just notice that if

Xn (w)→ X (w)

theng [Xn (w)]→ g [X (w)] .

Ruben Zamar [email protected] () Module 12 Asymptotic Results 11 / 60

a.s. Convergence and Continuity

Suppose that Xn → X a.s. and g (x) is a continuous function

It’s enough for g(x) to be continuous on the range of X

We can (easily) show that

g (Xn)→ g (X ) a.s.

Proof: just notice that if

Xn (w)→ X (w)

theng [Xn (w)]→ g [X (w)] .

Ruben Zamar [email protected] () Module 12 Asymptotic Results 11 / 60

a.s. Convergence and Continuity

Suppose that Xn → X a.s. and g (x) is a continuous function

It’s enough for g(x) to be continuous on the range of X

We can (easily) show that

g (Xn)→ g (X ) a.s.

Proof: just notice that if

Xn (w)→ X (w)

theng [Xn (w)]→ g [X (w)] .

Ruben Zamar [email protected] () Module 12 Asymptotic Results 11 / 60

a.s. Convergence and Continuity

Suppose that Xn → X a.s. and g (x) is a continuous function

It’s enough for g(x) to be continuous on the range of X

We can (easily) show that

g (Xn)→ g (X ) a.s.

Proof: just notice that if

Xn (w)→ X (w)

theng [Xn (w)]→ g [X (w)] .

Ruben Zamar [email protected] () Module 12 Asymptotic Results 11 / 60

Examples (Continued)

Example 2: Suppose that Xn are iid Exp(λ)

In this case E (Xn) = 1/λ, for all nBy the SLLN

X =1n

n

∑i=1

Xi → 1/λ a.s.

Now we use that a.s. convergence is preserved by continuous functions:

1X=

n∑ni=1 Xi

→ λ a.s.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 12 / 60

Examples (Continued)

Example 2: Suppose that Xn are iid Exp(λ)

In this case E (Xn) = 1/λ, for all n

By the SLLN

X =1n

n

∑i=1

Xi → 1/λ a.s.

Now we use that a.s. convergence is preserved by continuous functions:

1X=

n∑ni=1 Xi

→ λ a.s.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 12 / 60

Examples (Continued)

Example 2: Suppose that Xn are iid Exp(λ)

In this case E (Xn) = 1/λ, for all nBy the SLLN

X =1n

n

∑i=1

Xi → 1/λ a.s.

Now we use that a.s. convergence is preserved by continuous functions:

1X=

n∑ni=1 Xi

→ λ a.s.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 12 / 60

Examples (Continued)

Example 2: Suppose that Xn are iid Exp(λ)

In this case E (Xn) = 1/λ, for all nBy the SLLN

X =1n

n

∑i=1

Xi → 1/λ a.s.

Now we use that a.s. convergence is preserved by continuous functions:

1X=

n∑ni=1 Xi

→ λ a.s.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 12 / 60

The Sample Variance

Example 3: Suppose that Xn are iid with mean µ and variance σ2

By the SLLN

X =1n

n

∑i=1

Xi → µ a.s.

Moreover, since E(X 2i)= σ2 + µ2 we have that

X 2 =1n

n

∑i=1

X 2i → σ2 + µ2 a.s.

Since a.s. convergence is preserved by continuous functions we have

X 2 − X 2 → σ2 + µ2 − µ2 = σ2 a.s.

That is1n

n

∑i=1(Xi − X )2 → σ2 a.s.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 13 / 60

The Sample Variance

Example 3: Suppose that Xn are iid with mean µ and variance σ2

By the SLLN

X =1n

n

∑i=1

Xi → µ a.s.

Moreover, since E(X 2i)= σ2 + µ2 we have that

X 2 =1n

n

∑i=1

X 2i → σ2 + µ2 a.s.

Since a.s. convergence is preserved by continuous functions we have

X 2 − X 2 → σ2 + µ2 − µ2 = σ2 a.s.

That is1n

n

∑i=1(Xi − X )2 → σ2 a.s.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 13 / 60

The Sample Variance

Example 3: Suppose that Xn are iid with mean µ and variance σ2

By the SLLN

X =1n

n

∑i=1

Xi → µ a.s.

Moreover, since E(X 2i)= σ2 + µ2 we have that

X 2 =1n

n

∑i=1

X 2i → σ2 + µ2 a.s.

Since a.s. convergence is preserved by continuous functions we have

X 2 − X 2 → σ2 + µ2 − µ2 = σ2 a.s.

That is1n

n

∑i=1(Xi − X )2 → σ2 a.s.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 13 / 60

The Sample Variance

Example 3: Suppose that Xn are iid with mean µ and variance σ2

By the SLLN

X =1n

n

∑i=1

Xi → µ a.s.

Moreover, since E(X 2i)= σ2 + µ2 we have that

X 2 =1n

n

∑i=1

X 2i → σ2 + µ2 a.s.

Since a.s. convergence is preserved by continuous functions we have

X 2 − X 2 → σ2 + µ2 − µ2 = σ2 a.s.

That is1n

n

∑i=1(Xi − X )2 → σ2 a.s.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 13 / 60

The Sample Variance

Example 3: Suppose that Xn are iid with mean µ and variance σ2

By the SLLN

X =1n

n

∑i=1

Xi → µ a.s.

Moreover, since E(X 2i)= σ2 + µ2 we have that

X 2 =1n

n

∑i=1

X 2i → σ2 + µ2 a.s.

Since a.s. convergence is preserved by continuous functions we have

X 2 − X 2 → σ2 + µ2 − µ2 = σ2 a.s.

That is1n

n

∑i=1(Xi − X )2 → σ2 a.s.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 13 / 60

Convergence in Probability

Suppose that the rv’s Xn and X are all defined on the same samplespace Ω.

Notation:Xn →p X

means “Xn converges in probability to X as n→ ∞”Definition: Xn →p X if for all ε > 0,

limn→∞

P (|Xn − X | > ε) = 0

This type of convergence is weaker than a.s. convergence

Ruben Zamar [email protected] () Module 12 Asymptotic Results 14 / 60

Convergence in Probability

Suppose that the rv’s Xn and X are all defined on the same samplespace Ω.Notation:

Xn →p X

means “Xn converges in probability to X as n→ ∞”

Definition: Xn →p X if for all ε > 0,

limn→∞

P (|Xn − X | > ε) = 0

This type of convergence is weaker than a.s. convergence

Ruben Zamar [email protected] () Module 12 Asymptotic Results 14 / 60

Convergence in Probability

Suppose that the rv’s Xn and X are all defined on the same samplespace Ω.Notation:

Xn →p X

means “Xn converges in probability to X as n→ ∞”Definition: Xn →p X if for all ε > 0,

limn→∞

P (|Xn − X | > ε) = 0

This type of convergence is weaker than a.s. convergence

Ruben Zamar [email protected] () Module 12 Asymptotic Results 14 / 60

Convergence in Probability

Suppose that the rv’s Xn and X are all defined on the same samplespace Ω.Notation:

Xn →p X

means “Xn converges in probability to X as n→ ∞”Definition: Xn →p X if for all ε > 0,

limn→∞

P (|Xn − X | > ε) = 0

This type of convergence is weaker than a.s. convergence

Ruben Zamar [email protected] () Module 12 Asymptotic Results 14 / 60

The Weak Law of Large Numbers (WLLN)

Suppose that the random variables Xn are independent and all havethe same mean µ and finite variance σ2

The WLLN states that

1n

n

∑i=1Xi →p µ

as n→ ∞.We will give a simple proof of this result using Chevychev’s Inequality:

P (|X − µ| > ε) ≤ σ2

ε2

Ruben Zamar [email protected] () Module 12 Asymptotic Results 15 / 60

The Weak Law of Large Numbers (WLLN)

Suppose that the random variables Xn are independent and all havethe same mean µ and finite variance σ2

The WLLN states that

1n

n

∑i=1Xi →p µ

as n→ ∞.

We will give a simple proof of this result using Chevychev’s Inequality:

P (|X − µ| > ε) ≤ σ2

ε2

Ruben Zamar [email protected] () Module 12 Asymptotic Results 15 / 60

The Weak Law of Large Numbers (WLLN)

Suppose that the random variables Xn are independent and all havethe same mean µ and finite variance σ2

The WLLN states that

1n

n

∑i=1Xi →p µ

as n→ ∞.We will give a simple proof of this result using Chevychev’s Inequality:

P (|X − µ| > ε) ≤ σ2

ε2

Ruben Zamar [email protected] () Module 12 Asymptotic Results 15 / 60

Proof for the WLLN

We have

The SLLN states that

E

(1n

n

∑i=1Xi

)= µ

and

Var

(1n

n

∑i=1Xi

)=

σ2

n

By Chevychev’s Inequality

limn→∞

P (|X − µ| > ε) ≤ limn→∞

σ2

nε2= 0

Ruben Zamar [email protected] () Module 12 Asymptotic Results 16 / 60

Proof for the WLLN

We have

The SLLN states that

E

(1n

n

∑i=1Xi

)= µ

and

Var

(1n

n

∑i=1Xi

)=

σ2

n

By Chevychev’s Inequality

limn→∞

P (|X − µ| > ε) ≤ limn→∞

σ2

nε2= 0

Ruben Zamar [email protected] () Module 12 Asymptotic Results 16 / 60

Proof for the WLLN

We have

The SLLN states that

E

(1n

n

∑i=1Xi

)= µ

and

Var

(1n

n

∑i=1Xi

)=

σ2

n

By Chevychev’s Inequality

limn→∞

P (|X − µ| > ε) ≤ limn→∞

σ2

nε2= 0

Ruben Zamar [email protected] () Module 12 Asymptotic Results 16 / 60

Some Remarks

It can be shown (with some more work) that convergence inprobability is preserved by continuous functions.

More precisely, suppose that

Xn →p X and Yn →p Y ,g (s, t) is a continuous functionThen g (Xn ,Yn)→p g (X ,Y )

Example: suppose that Xn →p Z , where Z ∼ N (0, 1) thenX 2n →p Z 2, and Z 2 ∼ Gamma (1/2, 1/2) , which is call Chi-squarewith one degree of freedom.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 17 / 60

Some Remarks

It can be shown (with some more work) that convergence inprobability is preserved by continuous functions.

More precisely, suppose that

Xn →p X and Yn →p Y ,g (s, t) is a continuous functionThen g (Xn ,Yn)→p g (X ,Y )

Example: suppose that Xn →p Z , where Z ∼ N (0, 1) thenX 2n →p Z 2, and Z 2 ∼ Gamma (1/2, 1/2) , which is call Chi-squarewith one degree of freedom.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 17 / 60

Some Remarks

It can be shown (with some more work) that convergence inprobability is preserved by continuous functions.

More precisely, suppose that

Xn →p X and Yn →p Y ,

g (s, t) is a continuous functionThen g (Xn ,Yn)→p g (X ,Y )

Example: suppose that Xn →p Z , where Z ∼ N (0, 1) thenX 2n →p Z 2, and Z 2 ∼ Gamma (1/2, 1/2) , which is call Chi-squarewith one degree of freedom.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 17 / 60

Some Remarks

It can be shown (with some more work) that convergence inprobability is preserved by continuous functions.

More precisely, suppose that

Xn →p X and Yn →p Y ,g (s, t) is a continuous function

Then g (Xn ,Yn)→p g (X ,Y )

Example: suppose that Xn →p Z , where Z ∼ N (0, 1) thenX 2n →p Z 2, and Z 2 ∼ Gamma (1/2, 1/2) , which is call Chi-squarewith one degree of freedom.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 17 / 60

Some Remarks

It can be shown (with some more work) that convergence inprobability is preserved by continuous functions.

More precisely, suppose that

Xn →p X and Yn →p Y ,g (s, t) is a continuous functionThen g (Xn ,Yn)→p g (X ,Y )

Example: suppose that Xn →p Z , where Z ∼ N (0, 1) thenX 2n →p Z 2, and Z 2 ∼ Gamma (1/2, 1/2) , which is call Chi-squarewith one degree of freedom.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 17 / 60

Some Remarks

It can be shown (with some more work) that convergence inprobability is preserved by continuous functions.

More precisely, suppose that

Xn →p X and Yn →p Y ,g (s, t) is a continuous functionThen g (Xn ,Yn)→p g (X ,Y )

Example: suppose that Xn →p Z , where Z ∼ N (0, 1) thenX 2n →p Z 2, and Z 2 ∼ Gamma (1/2, 1/2) , which is call Chi-squarewith one degree of freedom.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 17 / 60

Max and Min of Independent Unif(a,b)

Suppose that X1,X2, ...,Xn are iid Unif(a, b)

The common cdf is

F (x) =x − ab− a , for a < x < b

LetUn = Min X1,X2, ...,Xn

andVn = Max X1,X2, ...,Xn

Ruben Zamar [email protected] () Module 12 Asymptotic Results 18 / 60

Max and Min of Independent Unif(a,b)

Suppose that X1,X2, ...,Xn are iid Unif(a, b)

The common cdf is

F (x) =x − ab− a , for a < x < b

LetUn = Min X1,X2, ...,Xn

andVn = Max X1,X2, ...,Xn

Ruben Zamar [email protected] () Module 12 Asymptotic Results 18 / 60

Max and Min of Independent Unif(a,b)

Suppose that X1,X2, ...,Xn are iid Unif(a, b)

The common cdf is

F (x) =x − ab− a , for a < x < b

LetUn = Min X1,X2, ...,Xn

andVn = Max X1,X2, ...,Xn

Ruben Zamar [email protected] () Module 12 Asymptotic Results 18 / 60

The Min Converges in Probability to a

We will show that Un →p a

Recall that for all a < u < b

FUn (u) = 1−[1− u − a

b− a

]n= 1−

(b− ub− a

)n

Ruben Zamar [email protected] () Module 12 Asymptotic Results 19 / 60

The Min Converges in Probability to a

We will show that Un →p a

Recall that for all a < u < b

FUn (u) = 1−[1− u − a

b− a

]n= 1−

(b− ub− a

)n

Ruben Zamar [email protected] () Module 12 Asymptotic Results 19 / 60

The Min Converges in Probability to a

In fact, let 0 < ε < b− a be given. Then,

P (|Un − a| > ε) = P (Un > a+ ε) +

=0︷ ︸︸ ︷P (Un < a− ε)

=

(b− a− ε

b− a

)n=

(1− ε

b− a

)n→ 0,

as n→ ∞

Students should show that Vn →p b.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 20 / 60

The Min Converges in Probability to a

In fact, let 0 < ε < b− a be given. Then,

P (|Un − a| > ε) = P (Un > a+ ε) +

=0︷ ︸︸ ︷P (Un < a− ε)

=

(b− a− ε

b− a

)n=

(1− ε

b− a

)n→ 0,

as n→ ∞Students should show that Vn →p b.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 20 / 60

Convergence in Distribution. Preliminary Concepts

Unlike a.s. convergence and convergence in probability, convergencein distribution doesn’t require the Xn to be all defined in the samesample space.

Technical concept: Given a cdf F (x) we define the setCF = x : F is continuous at xCF is the set of “continuity points of F”.It can be shown that C cF (the set of discontinuity points of F ) iseither finite or at most countable.

Proving this is a challenge question worth 5/100 marks added toyour midterm exam.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 21 / 60

Convergence in Distribution. Preliminary Concepts

Unlike a.s. convergence and convergence in probability, convergencein distribution doesn’t require the Xn to be all defined in the samesample space.

Technical concept: Given a cdf F (x) we define the setCF = x : F is continuous at x

CF is the set of “continuity points of F”.It can be shown that C cF (the set of discontinuity points of F ) iseither finite or at most countable.

Proving this is a challenge question worth 5/100 marks added toyour midterm exam.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 21 / 60

Convergence in Distribution. Preliminary Concepts

Unlike a.s. convergence and convergence in probability, convergencein distribution doesn’t require the Xn to be all defined in the samesample space.

Technical concept: Given a cdf F (x) we define the setCF = x : F is continuous at xCF is the set of “continuity points of F”.

It can be shown that C cF (the set of discontinuity points of F ) iseither finite or at most countable.

Proving this is a challenge question worth 5/100 marks added toyour midterm exam.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 21 / 60

Convergence in Distribution. Preliminary Concepts

Unlike a.s. convergence and convergence in probability, convergencein distribution doesn’t require the Xn to be all defined in the samesample space.

Technical concept: Given a cdf F (x) we define the setCF = x : F is continuous at xCF is the set of “continuity points of F”.It can be shown that C cF (the set of discontinuity points of F ) iseither finite or at most countable.

Proving this is a challenge question worth 5/100 marks added toyour midterm exam.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 21 / 60

Convergence in Distribution. Preliminary Concepts

Unlike a.s. convergence and convergence in probability, convergencein distribution doesn’t require the Xn to be all defined in the samesample space.

Technical concept: Given a cdf F (x) we define the setCF = x : F is continuous at xCF is the set of “continuity points of F”.It can be shown that C cF (the set of discontinuity points of F ) iseither finite or at most countable.

Proving this is a challenge question worth 5/100 marks added toyour midterm exam.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 21 / 60

Convergence in Distribution

Let F (x) be the cdf of the “target” rv X , and let CF be its set ofcontinuity points

Definition: We say that Xn converges in distribution to X andwrite Xn →d X if Fn (x)→ F (x) for all x ∈ CF .Note: the convergence of Fn (x) to F (x) may fail for points outsideCF

Ruben Zamar [email protected] () Module 12 Asymptotic Results 22 / 60

Convergence in Distribution

Let F (x) be the cdf of the “target” rv X , and let CF be its set ofcontinuity points

Definition: We say that Xn converges in distribution to X andwrite Xn →d X if Fn (x)→ F (x) for all x ∈ CF .

Note: the convergence of Fn (x) to F (x) may fail for points outsideCF

Ruben Zamar [email protected] () Module 12 Asymptotic Results 22 / 60

Convergence in Distribution

Let F (x) be the cdf of the “target” rv X , and let CF be its set ofcontinuity points

Definition: We say that Xn converges in distribution to X andwrite Xn →d X if Fn (x)→ F (x) for all x ∈ CF .Note: the convergence of Fn (x) to F (x) may fail for points outsideCF

Ruben Zamar [email protected] () Module 12 Asymptotic Results 22 / 60

Restriction to Continuity Points

Let Xn = 1/n with probability one and let X = 0 with probabilityone.

Then

Fn (x) =0 if x < 1/n1 if x ≥ 1/n

and

F (x) =0 if x < 01 if x ≥ 0

In this case CF = x : x 6= 0Notice that Fn (0) = 0 for all n and F (0) = 1

That is, Fn (0)9 F (0) .

However, Fn (x)→ F (x) for all x 6= 0. Therefore 1/n→ 0 indistribution!

Ruben Zamar [email protected] () Module 12 Asymptotic Results 23 / 60

Restriction to Continuity Points

Let Xn = 1/n with probability one and let X = 0 with probabilityone.

Then

Fn (x) =0 if x < 1/n1 if x ≥ 1/n

and

F (x) =0 if x < 01 if x ≥ 0

In this case CF = x : x 6= 0Notice that Fn (0) = 0 for all n and F (0) = 1

That is, Fn (0)9 F (0) .

However, Fn (x)→ F (x) for all x 6= 0. Therefore 1/n→ 0 indistribution!

Ruben Zamar [email protected] () Module 12 Asymptotic Results 23 / 60

Restriction to Continuity Points

Let Xn = 1/n with probability one and let X = 0 with probabilityone.

Then

Fn (x) =0 if x < 1/n1 if x ≥ 1/n

and

F (x) =0 if x < 01 if x ≥ 0

In this case CF = x : x 6= 0

Notice that Fn (0) = 0 for all n and F (0) = 1

That is, Fn (0)9 F (0) .

However, Fn (x)→ F (x) for all x 6= 0. Therefore 1/n→ 0 indistribution!

Ruben Zamar [email protected] () Module 12 Asymptotic Results 23 / 60

Restriction to Continuity Points

Let Xn = 1/n with probability one and let X = 0 with probabilityone.

Then

Fn (x) =0 if x < 1/n1 if x ≥ 1/n

and

F (x) =0 if x < 01 if x ≥ 0

In this case CF = x : x 6= 0Notice that Fn (0) = 0 for all n and F (0) = 1

That is, Fn (0)9 F (0) .

However, Fn (x)→ F (x) for all x 6= 0. Therefore 1/n→ 0 indistribution!

Ruben Zamar [email protected] () Module 12 Asymptotic Results 23 / 60

Restriction to Continuity Points

Let Xn = 1/n with probability one and let X = 0 with probabilityone.

Then

Fn (x) =0 if x < 1/n1 if x ≥ 1/n

and

F (x) =0 if x < 01 if x ≥ 0

In this case CF = x : x 6= 0Notice that Fn (0) = 0 for all n and F (0) = 1

That is, Fn (0)9 F (0) .

However, Fn (x)→ F (x) for all x 6= 0. Therefore 1/n→ 0 indistribution!

Ruben Zamar [email protected] () Module 12 Asymptotic Results 23 / 60

Restriction to Continuity Points

Let Xn = 1/n with probability one and let X = 0 with probabilityone.

Then

Fn (x) =0 if x < 1/n1 if x ≥ 1/n

and

F (x) =0 if x < 01 if x ≥ 0

In this case CF = x : x 6= 0Notice that Fn (0) = 0 for all n and F (0) = 1

That is, Fn (0)9 F (0) .

However, Fn (x)→ F (x) for all x 6= 0. Therefore 1/n→ 0 indistribution!

Ruben Zamar [email protected] () Module 12 Asymptotic Results 23 / 60

Asymptotic Distribution of the Max of Unif(a,b)

Example: Suppose that X1,X2, ...,Xn are iid Unif (a, b)

We have shown above that

Vn = max X1,X2, ...,Xn →p b

We will show now that

n (b− Vn)→d Exp(

1b− a

)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 24 / 60

Asymptotic Distribution of the Max of Unif(a,b)

Example: Suppose that X1,X2, ...,Xn are iid Unif (a, b)

We have shown above that

Vn = max X1,X2, ...,Xn →p b

We will show now that

n (b− Vn)→d Exp(

1b− a

)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 24 / 60

Asymptotic Distribution of the Max of Unif(a,b)

Example: Suppose that X1,X2, ...,Xn are iid Unif (a, b)

We have shown above that

Vn = max X1,X2, ...,Xn →p b

We will show now that

n (b− Vn)→d Exp(

1b− a

)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 24 / 60

Max of Unif(a,b) (continued)

We will first obtain the cdf for n (b− Vn) .

For any d > 0 such that d/n < b, we have

P [n (b− Vn) ≤ d ] = P[b− d

n≤ Vn

]= 1− P

[Vn < b−

dn

]= 1−

[(b− d

n

)− a

b− a

]n= 1−

[1− d

n (b− a)

]n

Ruben Zamar [email protected] () Module 12 Asymptotic Results 25 / 60

Max of Unif(a,b) (continued)

We will first obtain the cdf for n (b− Vn) .For any d > 0 such that d/n < b, we have

P [n (b− Vn) ≤ d ] = P[b− d

n≤ Vn

]= 1− P

[Vn < b−

dn

]= 1−

[(b− d

n

)− a

b− a

]n= 1−

[1− d

n (b− a)

]n

Ruben Zamar [email protected] () Module 12 Asymptotic Results 25 / 60

Max of Unif(a,b) (continued)

Recall that (1+

xn

)n→ ex , for all x as n→ ∞

Therefore

1−[1− d

n (b− a)

]n→ 1− exp

− db− a

, for all d > 0.

This is the cdf of an Exp( 1b−a)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 26 / 60

Max of Unif(a,b) (continued)

Recall that (1+

xn

)n→ ex , for all x as n→ ∞

Therefore

1−[1− d

n (b− a)

]n→ 1− exp

− db− a

, for all d > 0.

This is the cdf of an Exp( 1b−a)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 26 / 60

Max of Unif(a,b) (continued)

Recall that (1+

xn

)n→ ex , for all x as n→ ∞

Therefore

1−[1− d

n (b− a)

]n→ 1− exp

− db− a

, for all d > 0.

This is the cdf of an Exp( 1b−a)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 26 / 60

Asymptotic Distribution of the Min of Unif(a,b)

The derivation of the asymptotic distribution of n (Un − a) is achallenge problem worth 1/100 increase in the midterm grade.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 27 / 60

Convergence in Distribution and Continuity

It can be shown (this time with considerable level of diffi culty) thatconvergence in distribution is preserved by continuous functions

More precisely, if Xn →d X and g (x) is continuous, theng (Xn)→d g (X )

Example: Suppose that Xn →d X ∼ N (0, 1) Then

2+ 3Xn →d 2+ 3X ∼ N (2, 9)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 28 / 60

Convergence in Distribution and Continuity

It can be shown (this time with considerable level of diffi culty) thatconvergence in distribution is preserved by continuous functions

More precisely, if Xn →d X and g (x) is continuous, theng (Xn)→d g (X )

Example: Suppose that Xn →d X ∼ N (0, 1) Then

2+ 3Xn →d 2+ 3X ∼ N (2, 9)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 28 / 60

Convergence in Distribution and Continuity

It can be shown (this time with considerable level of diffi culty) thatconvergence in distribution is preserved by continuous functions

More precisely, if Xn →d X and g (x) is continuous, theng (Xn)→d g (X )

Example: Suppose that Xn →d X ∼ N (0, 1) Then

2+ 3Xn →d 2+ 3X ∼ N (2, 9)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 28 / 60

Another Example

Example: Suppose that Xn →d X ∼ Unif (0, 1) . What is thelimiting distribution of − log (Xn)?

Solution: By continuity

− log (Xn)→d − log (X )

Moreover, for all y > 0,

P (− log (X ) ≤ y) = P (log (X ) ≥ −y)= P

(X ≥ e−y

)= 1− e−y

Therefore, − log (Xn)→d Exp (1)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 29 / 60

Another Example

Example: Suppose that Xn →d X ∼ Unif (0, 1) . What is thelimiting distribution of − log (Xn)?Solution: By continuity

− log (Xn)→d − log (X )

Moreover, for all y > 0,

P (− log (X ) ≤ y) = P (log (X ) ≥ −y)= P

(X ≥ e−y

)= 1− e−y

Therefore, − log (Xn)→d Exp (1)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 29 / 60

Another Example

Example: Suppose that Xn →d X ∼ Unif (0, 1) . What is thelimiting distribution of − log (Xn)?Solution: By continuity

− log (Xn)→d − log (X )

Moreover, for all y > 0,

P (− log (X ) ≤ y) = P (log (X ) ≥ −y)= P

(X ≥ e−y

)= 1− e−y

Therefore, − log (Xn)→d Exp (1)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 29 / 60

Another Example

Example: Suppose that Xn →d X ∼ Unif (0, 1) . What is thelimiting distribution of − log (Xn)?Solution: By continuity

− log (Xn)→d − log (X )

Moreover, for all y > 0,

P (− log (X ) ≤ y) = P (log (X ) ≥ −y)= P

(X ≥ e−y

)= 1− e−y

Therefore, − log (Xn)→d Exp (1)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 29 / 60

An important Technical Result (on MGF’s)

Suppose that Xn ∼ Fn and Mn (t) = E(etXn

), for all n

Suppose that X ∼ F and M (t) = E(etX)

Suppose that Mn (t)→ M (t) for all −ε < t < ε, for some ε > 0

Then Fn (x)→ F (x) for all x ∈ CFIn other words Mn (t)→ M (t) for all −ε < t < ε, for some ε > 0implies that Xn →d X .

Ruben Zamar [email protected] () Module 12 Asymptotic Results 30 / 60

An important Technical Result (on MGF’s)

Suppose that Xn ∼ Fn and Mn (t) = E(etXn

), for all n

Suppose that X ∼ F and M (t) = E(etX)

Suppose that Mn (t)→ M (t) for all −ε < t < ε, for some ε > 0

Then Fn (x)→ F (x) for all x ∈ CFIn other words Mn (t)→ M (t) for all −ε < t < ε, for some ε > 0implies that Xn →d X .

Ruben Zamar [email protected] () Module 12 Asymptotic Results 30 / 60

An important Technical Result (on MGF’s)

Suppose that Xn ∼ Fn and Mn (t) = E(etXn

), for all n

Suppose that X ∼ F and M (t) = E(etX)

Suppose that Mn (t)→ M (t) for all −ε < t < ε, for some ε > 0

Then Fn (x)→ F (x) for all x ∈ CFIn other words Mn (t)→ M (t) for all −ε < t < ε, for some ε > 0implies that Xn →d X .

Ruben Zamar [email protected] () Module 12 Asymptotic Results 30 / 60

An important Technical Result (on MGF’s)

Suppose that Xn ∼ Fn and Mn (t) = E(etXn

), for all n

Suppose that X ∼ F and M (t) = E(etX)

Suppose that Mn (t)→ M (t) for all −ε < t < ε, for some ε > 0

Then Fn (x)→ F (x) for all x ∈ CF

In other words Mn (t)→ M (t) for all −ε < t < ε, for some ε > 0implies that Xn →d X .

Ruben Zamar [email protected] () Module 12 Asymptotic Results 30 / 60

An important Technical Result (on MGF’s)

Suppose that Xn ∼ Fn and Mn (t) = E(etXn

), for all n

Suppose that X ∼ F and M (t) = E(etX)

Suppose that Mn (t)→ M (t) for all −ε < t < ε, for some ε > 0

Then Fn (x)→ F (x) for all x ∈ CFIn other words Mn (t)→ M (t) for all −ε < t < ε, for some ε > 0implies that Xn →d X .

Ruben Zamar [email protected] () Module 12 Asymptotic Results 30 / 60

The Central Limit Theorem (CLT)

Suppose that X1,X2, ...,Xn are iid with common mean µ andcommon variance σ2

Then

Zn =Xn − E (Xn)SD (Xn)

=√n(Xn − µ)

σ→d Z ∼ N (0, 1)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 31 / 60

The Central Limit Theorem (CLT)

Suppose that X1,X2, ...,Xn are iid with common mean µ andcommon variance σ2

Then

Zn =Xn − E (Xn)SD (Xn)

=√n(Xn − µ)

σ→d Z ∼ N (0, 1)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 31 / 60

CLT (continued)

To summarize, for large n (e.g. n ≥ 20)

√n(Xn − µ)

σ≈ N (0, 1)

⇒ Xn − µ ≈ σ√nN (0, 1) = N

(0,

σ2

n

)

⇒ Xn ≈ µ+N(0,

σ2

n

)= N

(µ,

σ2

n

)

⇒ Xn ≈ N [E (Xn) ,Var (Xn)]

Ruben Zamar [email protected] () Module 12 Asymptotic Results 32 / 60

CLT (continued)

To summarize, for large n (e.g. n ≥ 20)

√n(Xn − µ)

σ≈ N (0, 1)

⇒ Xn − µ ≈ σ√nN (0, 1) = N

(0,

σ2

n

)

⇒ Xn ≈ µ+N(0,

σ2

n

)= N

(µ,

σ2

n

)

⇒ Xn ≈ N [E (Xn) ,Var (Xn)]

Ruben Zamar [email protected] () Module 12 Asymptotic Results 32 / 60

CLT (continued)

To summarize, for large n (e.g. n ≥ 20)

√n(Xn − µ)

σ≈ N (0, 1)

⇒ Xn − µ ≈ σ√nN (0, 1) = N

(0,

σ2

n

)

⇒ Xn ≈ µ+N(0,

σ2

n

)= N

(µ,

σ2

n

)

⇒ Xn ≈ N [E (Xn) ,Var (Xn)]

Ruben Zamar [email protected] () Module 12 Asymptotic Results 32 / 60

CLT (continued)

To summarize, for large n (e.g. n ≥ 20)

√n(Xn − µ)

σ≈ N (0, 1)

⇒ Xn − µ ≈ σ√nN (0, 1) = N

(0,

σ2

n

)

⇒ Xn ≈ µ+N(0,

σ2

n

)= N

(µ,

σ2

n

)

⇒ Xn ≈ N [E (Xn) ,Var (Xn)]

Ruben Zamar [email protected] () Module 12 Asymptotic Results 32 / 60

CLT (continued)

Proof. We can assume without loss of generality that µ = 0 andσ2 = 1.

Challenge Question: Show that we can indeed assume that µ = 0and σ2 = 1 without loss of generality. This is for an increase of0.5/30 in your MT 2 grade.

Then Zn =√nXn and using the iid assumption:

MZn (t) = M√nXn (t)

= M(∑Xi )/√n (t)

= M∑Xi(t/√n)

=[M(t/√n)]n

Ruben Zamar [email protected] () Module 12 Asymptotic Results 33 / 60

CLT (continued)

Proof. We can assume without loss of generality that µ = 0 andσ2 = 1.

Challenge Question: Show that we can indeed assume that µ = 0and σ2 = 1 without loss of generality. This is for an increase of0.5/30 in your MT 2 grade.

Then Zn =√nXn and using the iid assumption:

MZn (t) = M√nXn (t)

= M(∑Xi )/√n (t)

= M∑Xi(t/√n)

=[M(t/√n)]n

Ruben Zamar [email protected] () Module 12 Asymptotic Results 33 / 60

CLT (continued)

Proof. We can assume without loss of generality that µ = 0 andσ2 = 1.

Challenge Question: Show that we can indeed assume that µ = 0and σ2 = 1 without loss of generality. This is for an increase of0.5/30 in your MT 2 grade.

Then Zn =√nXn and using the iid assumption:

MZn (t) = M√nXn (t)

= M(∑Xi )/√n (t)

= M∑Xi(t/√n)

=[M(t/√n)]n

Ruben Zamar [email protected] () Module 12 Asymptotic Results 33 / 60

CLT (continued)

Moreover,

M(t/√n)= E

exp

(tX1√n

)= E

1+

tX1√n+t2X 212n

+ o

(t√n

)

= 1+t2

2n+ o

(t√n

)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 34 / 60

CLT (continued)

Hence

M(t/√n)n=

[1+

t2

2n+ o

(t√n

)]n→ exp

(t2

2

)which is the MGF of a standard normal random variable. This provesthe result.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 35 / 60

Some Remarks

We have shown that Xn →p µ By continuity Xn − µ→p 0 andtherefore

Xn − µ

σ→p 0

The CLT shows that the right “magnification”of the differencebetween Xn and µ to appreciate its behavior when n is large is

√n.

Less magnification (such as n1/3 or log (n)) would not be enoughbecause the magnified difference would still collapse to zero

More magnification (such as n3 or en ) would be to much because themagnified absolute difference would blow out to infinity.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 36 / 60

Some Remarks

We have shown that Xn →p µ By continuity Xn − µ→p 0 andtherefore

Xn − µ

σ→p 0

The CLT shows that the right “magnification”of the differencebetween Xn and µ to appreciate its behavior when n is large is

√n.

Less magnification (such as n1/3 or log (n)) would not be enoughbecause the magnified difference would still collapse to zero

More magnification (such as n3 or en ) would be to much because themagnified absolute difference would blow out to infinity.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 36 / 60

Some Remarks

We have shown that Xn →p µ By continuity Xn − µ→p 0 andtherefore

Xn − µ

σ→p 0

The CLT shows that the right “magnification”of the differencebetween Xn and µ to appreciate its behavior when n is large is

√n.

Less magnification (such as n1/3 or log (n)) would not be enoughbecause the magnified difference would still collapse to zero

More magnification (such as n3 or en ) would be to much because themagnified absolute difference would blow out to infinity.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 36 / 60

Some Remarks

We have shown that Xn →p µ By continuity Xn − µ→p 0 andtherefore

Xn − µ

σ→p 0

The CLT shows that the right “magnification”of the differencebetween Xn and µ to appreciate its behavior when n is large is

√n.

Less magnification (such as n1/3 or log (n)) would not be enoughbecause the magnified difference would still collapse to zero

More magnification (such as n3 or en ) would be to much because themagnified absolute difference would blow out to infinity.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 36 / 60

Bounded in Probability

A normal random variable is unbounded (it can take any valuebetween −∞ and ∞)

However Z ∼ N (0, 1) is bounded in probability because theP (|Z | > K ) = 2 (1−Φ (K ))→ 0 as K → ∞.A sequence Yn is bounded in probability if for all δ > 0, there existsKδ > 0 and Nδ > 0 such that

P (|Yn | ≤ Kδ) ≥ 1− δ,

for all n ≥ Nδ.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 37 / 60

Bounded in Probability

A normal random variable is unbounded (it can take any valuebetween −∞ and ∞)However Z ∼ N (0, 1) is bounded in probability because theP (|Z | > K ) = 2 (1−Φ (K ))→ 0 as K → ∞.

A sequence Yn is bounded in probability if for all δ > 0, there existsKδ > 0 and Nδ > 0 such that

P (|Yn | ≤ Kδ) ≥ 1− δ,

for all n ≥ Nδ.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 37 / 60

Bounded in Probability

A normal random variable is unbounded (it can take any valuebetween −∞ and ∞)However Z ∼ N (0, 1) is bounded in probability because theP (|Z | > K ) = 2 (1−Φ (K ))→ 0 as K → ∞.A sequence Yn is bounded in probability if for all δ > 0, there existsKδ > 0 and Nδ > 0 such that

P (|Yn | ≤ Kδ) ≥ 1− δ,

for all n ≥ Nδ.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 37 / 60

Bounded in Probability and the CLT

The CLT also means that the difference∣∣√n (Xn − µ)

∣∣ is boundedin probability

Notation: Xn − µ = Op(1/√n)meaning that∣∣∣∣ Xn − µ

1/√n

∣∣∣∣ = ∣∣√n (Xn − µ)∣∣ is bounded in probability

Ruben Zamar [email protected] () Module 12 Asymptotic Results 38 / 60

Bounded in Probability and the CLT

The CLT also means that the difference∣∣√n (Xn − µ)

∣∣ is boundedin probability

Notation: Xn − µ = Op(1/√n)meaning that∣∣∣∣ Xn − µ

1/√n

∣∣∣∣ = ∣∣√n (Xn − µ)∣∣ is bounded in probability

Ruben Zamar [email protected] () Module 12 Asymptotic Results 38 / 60

CLT ApplicationNormal Approximation for the Binomial

Suppose that Xn ∼ Binom (n, p)

Then

Xn = Y1 + Y2 + · · ·+ Yn, Yi ∼ iid Binom (1, p)

E (Yi ) = p, Var (Yi ) = p (1− p)

We can writeXnn=Y1 + Y2 + · · ·+ Yn

n= Yn

Ruben Zamar [email protected] () Module 12 Asymptotic Results 39 / 60

CLT ApplicationNormal Approximation for the Binomial

Suppose that Xn ∼ Binom (n, p)Then

Xn = Y1 + Y2 + · · ·+ Yn, Yi ∼ iid Binom (1, p)

E (Yi ) = p, Var (Yi ) = p (1− p)

We can writeXnn=Y1 + Y2 + · · ·+ Yn

n= Yn

Ruben Zamar [email protected] () Module 12 Asymptotic Results 39 / 60

CLT ApplicationNormal Approximation for the Binomial

Suppose that Xn ∼ Binom (n, p)Then

Xn = Y1 + Y2 + · · ·+ Yn, Yi ∼ iid Binom (1, p)

E (Yi ) = p, Var (Yi ) = p (1− p)

We can writeXnn=Y1 + Y2 + · · ·+ Yn

n= Yn

Ruben Zamar [email protected] () Module 12 Asymptotic Results 39 / 60

CLT Application (continued)Normal Approximation for the Binomial

By the CLT,

√n

(Xnn − p√p (1− p)

)=√n

(Yn − p√p (1− p)

)→ Z ∼ N (0, 1)

This means that

√n

(Xnn − p√p (1− p)

)w N (0, 1)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 40 / 60

CLT Application (continued)Normal Approximation for the Binomial

By the CLT,

√n

(Xnn − p√p (1− p)

)=√n

(Yn − p√p (1− p)

)→ Z ∼ N (0, 1)

This means that

√n

(Xnn − p√p (1− p)

)w N (0, 1)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 40 / 60

CLT Application (continued)Normal Approximation for the Binomial

Therefore

=⇒Xnn − p√p (1− p)

w 1√nN (0, 1) = N

(0,1n

)

=⇒ Xnn− p w

√p (1− p)N

(0,1n

)= N

(0,p (1− p)

n

)

=⇒ Xnnw p +N

(0,p (1− p)

n

)= N

(p,p (1− p)

n

)

=⇒ Xn w nN(p,p (1− p)

n

)= N

(np,

n2p (1− p)n

)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 41 / 60

CLT Application (continued)Normal Approximation for the Binomial

In summary

Binomial (n, p) w N (np, np (1− p)) = N (E (Xn) ,Var (Xn))

That is, when n is large and p is not too small (so that the originaldistribution is not too asymmetric)

Rule of thumb: min np, n (1− p) ≥ 5

Ruben Zamar [email protected] () Module 12 Asymptotic Results 42 / 60

CLT Application (continued)Normal Approximation for the Binomial

In summary

Binomial (n, p) w N (np, np (1− p)) = N (E (Xn) ,Var (Xn))

That is, when n is large and p is not too small (so that the originaldistribution is not too asymmetric)

Rule of thumb: min np, n (1− p) ≥ 5

Ruben Zamar [email protected] () Module 12 Asymptotic Results 42 / 60

CLT Application (continued)Normal Approximation for the Binomial

In summary

Binomial (n, p) w N (np, np (1− p)) = N (E (Xn) ,Var (Xn))

That is, when n is large and p is not too small (so that the originaldistribution is not too asymmetric)

Rule of thumb: min np, n (1− p) ≥ 5

Ruben Zamar [email protected] () Module 12 Asymptotic Results 42 / 60

CLT Application (continued)Normal Approximation for the Binomial

As a numerical example, let X ∼ Binom (20, 0.4)

Use the normal distribution to approximate

(a) P (X ≥ 6) (b) P (6 < X < 9)

Solution

E (X ) = 20× 0.4 = 8,Var (X ) = 20× 0.4× 0.6 = 4.8SD (X ) =

√4.8 = 2.1909

Ruben Zamar [email protected] () Module 12 Asymptotic Results 43 / 60

CLT Application (continued)Normal Approximation for the Binomial

As a numerical example, let X ∼ Binom (20, 0.4)Use the normal distribution to approximate

(a) P (X ≥ 6) (b) P (6 < X < 9)

Solution

E (X ) = 20× 0.4 = 8,Var (X ) = 20× 0.4× 0.6 = 4.8SD (X ) =

√4.8 = 2.1909

Ruben Zamar [email protected] () Module 12 Asymptotic Results 43 / 60

CLT Application (continued)Normal Approximation for the Binomial

As a numerical example, let X ∼ Binom (20, 0.4)Use the normal distribution to approximate

(a) P (X ≥ 6) (b) P (6 < X < 9)

Solution

E (X ) = 20× 0.4 = 8,Var (X ) = 20× 0.4× 0.6 = 4.8SD (X ) =

√4.8 = 2.1909

Ruben Zamar [email protected] () Module 12 Asymptotic Results 43 / 60

CLT Application (continued)Normal Approximation for the Binomial

As a numerical example, let X ∼ Binom (20, 0.4)Use the normal distribution to approximate

(a) P (X ≥ 6) (b) P (6 < X < 9)

Solution

E (X ) = 20× 0.4 = 8,Var (X ) = 20× 0.4× 0.6 = 4.8SD (X ) =

√4.8 = 2.1909

Ruben Zamar [email protected] () Module 12 Asymptotic Results 43 / 60

CLT Application (continued)

(a)

P (X ≥ 6) = 1− P (X < 6) = 1− P (X ≤ 5)

= 1−Φ(5− 82.1909

)= 1−Φ (−1.3693)

= 0.9145472

Ruben Zamar [email protected] () Module 12 Asymptotic Results 44 / 60

CLT Application (continued)

(b)

P (6 < X < 9) = P (6 < X ≤ 8)

= Φ(8− 82.1909

)−Φ

(6− 82.1909

)= 0.5−Φ (−0.91287) = 0.3193445

Ruben Zamar [email protected] () Module 12 Asymptotic Results 45 / 60

CLT Application (continued)Normal Approximation for the Binomial

The exact probability calculated using the Binomial distribution are0.874401 and 0.3455881

With “correction for discreteness”we have

1−Φ(5.5−82.1909

)= 1−Φ (−1.1411) = 0.8730858 and

Φ(8.5−82.1909

)−Φ

(6.5−82.1909

)= Φ (0.228 22)−Φ (−0.68465) = 0.34348

Fortunately for all of us, these approximations are nowadaysobsolete as the exact quantities can be easily computed usingstatistical software.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 46 / 60

CLT Application (continued)Normal Approximation for the Binomial

The exact probability calculated using the Binomial distribution are0.874401 and 0.3455881

With “correction for discreteness”we have

1−Φ(5.5−82.1909

)= 1−Φ (−1.1411) = 0.8730858 and

Φ(8.5−82.1909

)−Φ

(6.5−82.1909

)= Φ (0.228 22)−Φ (−0.68465) = 0.34348

Fortunately for all of us, these approximations are nowadaysobsolete as the exact quantities can be easily computed usingstatistical software.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 46 / 60

CLT Application (continued)Normal Approximation for the Binomial

The exact probability calculated using the Binomial distribution are0.874401 and 0.3455881

With “correction for discreteness”we have

1−Φ(5.5−82.1909

)= 1−Φ (−1.1411) = 0.8730858 and

Φ(8.5−82.1909

)−Φ

(6.5−82.1909

)= Φ (0.228 22)−Φ (−0.68465) = 0.34348

Fortunately for all of us, these approximations are nowadaysobsolete as the exact quantities can be easily computed usingstatistical software.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 46 / 60

CLT Application (continued)Normal Approximation for the Binomial

The exact probability calculated using the Binomial distribution are0.874401 and 0.3455881

With “correction for discreteness”we have

1−Φ(5.5−82.1909

)= 1−Φ (−1.1411) = 0.8730858 and

Φ(8.5−82.1909

)−Φ

(6.5−82.1909

)= Φ (0.228 22)−Φ (−0.68465) = 0.34348

Fortunately for all of us, these approximations are nowadaysobsolete as the exact quantities can be easily computed usingstatistical software.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 46 / 60

CLT Application (continued)Normal Approximation for the Binomial

The exact probability calculated using the Binomial distribution are0.874401 and 0.3455881

With “correction for discreteness”we have

1−Φ(5.5−82.1909

)= 1−Φ (−1.1411) = 0.8730858 and

Φ(8.5−82.1909

)−Φ

(6.5−82.1909

)= Φ (0.228 22)−Φ (−0.68465) = 0.34348

Fortunately for all of us, these approximations are nowadaysobsolete as the exact quantities can be easily computed usingstatistical software.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 46 / 60

CLT ApplicationMonte Carlo Integration

Suppose we wish to compute the integral

A =∫ b

ag (x) dx

We notice that

I = (b− a) 1b− a

∫ b

ag (x) dx

= (b− a)E g (U) , U ∼ Unif (a, b)

= (b− a)A, A = E g (U)Therefore

I = (b− a)A, where A = E g (U) .

Ruben Zamar [email protected] () Module 12 Asymptotic Results 47 / 60

CLT ApplicationMonte Carlo Integration

Suppose we wish to compute the integral

A =∫ b

ag (x) dx

We notice that

I = (b− a) 1b− a

∫ b

ag (x) dx

= (b− a)E g (U) , U ∼ Unif (a, b)

= (b− a)A, A = E g (U)

ThereforeI = (b− a)A, where A = E g (U) .

Ruben Zamar [email protected] () Module 12 Asymptotic Results 47 / 60

CLT ApplicationMonte Carlo Integration

Suppose we wish to compute the integral

A =∫ b

ag (x) dx

We notice that

I = (b− a) 1b− a

∫ b

ag (x) dx

= (b− a)E g (U) , U ∼ Unif (a, b)

= (b− a)A, A = E g (U)Therefore

I = (b− a)A, where A = E g (U) .

Ruben Zamar [email protected] () Module 12 Asymptotic Results 47 / 60

CLT Application (Continued)Monte Carlo Integration

Monte Carlo Integration Method:

1 generate a large enough number N of iid Ui with common distributionUnif (a, b)

2 Estimate the “population mean” A by the “sample mean” A, where

A =1N

N

∑i=1

g (Ui )

Ruben Zamar [email protected] () Module 12 Asymptotic Results 48 / 60

CLT Application (Continued)Monte Carlo Integration

Monte Carlo Integration Method:1 generate a large enough number N of iid Ui with common distributionUnif (a, b)

2 Estimate the “population mean” A by the “sample mean” A, where

A =1N

N

∑i=1

g (Ui )

Ruben Zamar [email protected] () Module 12 Asymptotic Results 48 / 60

CLT Application (Continued)Monte Carlo Integration

Monte Carlo Integration Method:1 generate a large enough number N of iid Ui with common distributionUnif (a, b)

2 Estimate the “population mean” A by the “sample mean” A, where

A =1N

N

∑i=1

g (Ui )

Ruben Zamar [email protected] () Module 12 Asymptotic Results 48 / 60

CLT Application (Continued)

Properties of the Monte Carlo Estimate

1 A is an unbiased estimate of A : E(A)= A

Because the sample mean is an unbiased estimate of the populationmean

2 A→ A a.s. (and also in probability)Because of the SLLN (and the WLLN)

3√N(A− A

)→d N

(0, τ2

), where τ2 = Var (g (U)) (by the CLT)

4 τ2 = (1/N)∑ g2 (Ui )− [(1/N)∑ g (Ui )]→ τ2 a.s. (and also inprobability)Because of the SLLN (and the WLLN)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 49 / 60

CLT Application (Continued)

Properties of the Monte Carlo Estimate1 A is an unbiased estimate of A : E

(A)= A

Because the sample mean is an unbiased estimate of the populationmean

2 A→ A a.s. (and also in probability)Because of the SLLN (and the WLLN)

3√N(A− A

)→d N

(0, τ2

), where τ2 = Var (g (U)) (by the CLT)

4 τ2 = (1/N)∑ g2 (Ui )− [(1/N)∑ g (Ui )]→ τ2 a.s. (and also inprobability)Because of the SLLN (and the WLLN)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 49 / 60

CLT Application (Continued)

Properties of the Monte Carlo Estimate1 A is an unbiased estimate of A : E

(A)= A

Because the sample mean is an unbiased estimate of the populationmean

2 A→ A a.s. (and also in probability)Because of the SLLN (and the WLLN)

3√N(A− A

)→d N

(0, τ2

), where τ2 = Var (g (U)) (by the CLT)

4 τ2 = (1/N)∑ g2 (Ui )− [(1/N)∑ g (Ui )]→ τ2 a.s. (and also inprobability)Because of the SLLN (and the WLLN)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 49 / 60

CLT Application (Continued)

Properties of the Monte Carlo Estimate1 A is an unbiased estimate of A : E

(A)= A

Because the sample mean is an unbiased estimate of the populationmean

2 A→ A a.s. (and also in probability)Because of the SLLN (and the WLLN)

3√N(A− A

)→d N

(0, τ2

), where τ2 = Var (g (U)) (by the CLT)

4 τ2 = (1/N)∑ g2 (Ui )− [(1/N)∑ g (Ui )]→ τ2 a.s. (and also inprobability)Because of the SLLN (and the WLLN)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 49 / 60

CLT Application (Continued)

Properties of the Monte Carlo Estimate1 A is an unbiased estimate of A : E

(A)= A

Because the sample mean is an unbiased estimate of the populationmean

2 A→ A a.s. (and also in probability)Because of the SLLN (and the WLLN)

3√N(A− A

)→d N

(0, τ2

), where τ2 = Var (g (U)) (by the CLT)

4 τ2 = (1/N)∑ g2 (Ui )− [(1/N)∑ g (Ui )]→ τ2 a.s. (and also inprobability)Because of the SLLN (and the WLLN)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 49 / 60

Slutzky’s Results

Suppose that Xn →d X and Yn →p c. Then

(a) Xn + Yn →d X + c(b) XnYn →d Xc(c) XnYn →d X/c , provided c 6= 0.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 50 / 60

Slutzky’s Results

Suppose that Xn →d X and Yn →p c. Then

(a) Xn + Yn →d X + c

(b) XnYn →d Xc(c) XnYn →d X/c , provided c 6= 0.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 50 / 60

Slutzky’s Results

Suppose that Xn →d X and Yn →p c. Then

(a) Xn + Yn →d X + c(b) XnYn →d Xc

(c) XnYn →d X/c , provided c 6= 0.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 50 / 60

Slutzky’s Results

Suppose that Xn →d X and Yn →p c. Then

(a) Xn + Yn →d X + c(b) XnYn →d Xc(c) XnYn →d X/c , provided c 6= 0.

Ruben Zamar [email protected] () Module 12 Asymptotic Results 50 / 60

Application of Slutzky’s Results to MC Integration

By property 3 we have

√N

(A− A

→d N (0, 1) (*)

By property 4 we haveτ2 →p τ2

By continuity we also have

τ →p τ

andτ

τ→p 1 (**)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 51 / 60

Application of Slutzky’s Results to MC Integration

By property 3 we have

√N

(A− A

→d N (0, 1) (*)

By property 4 we haveτ2 →p τ2

By continuity we also have

τ →p τ

andτ

τ→p 1 (**)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 51 / 60

Application of Slutzky’s Results to MC Integration

By property 3 we have

√N

(A− A

→d N (0, 1) (*)

By property 4 we haveτ2 →p τ2

By continuity we also have

τ →p τ

andτ

τ→p 1 (**)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 51 / 60

Application of Slutzky’s Results to MC Integration(continued)

By (*), (**), and Slutzky’s Result (c) we have

√N

(A− A

τ

τ→d N (0, 1)

That is√N

(A− A

→d N (0, 1)

In summary,

A w N(A,

τ2

N

)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 52 / 60

Application of Slutzky’s Results to MC Integration(continued)

By (*), (**), and Slutzky’s Result (c) we have

√N

(A− A

τ

τ→d N (0, 1)

That is√N

(A− A

→d N (0, 1)

In summary,

A w N(A,

τ2

N

)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 52 / 60

Application of Slutzky’s Results to MC Integration(continued)

By (*), (**), and Slutzky’s Result (c) we have

√N

(A− A

τ

τ→d N (0, 1)

That is√N

(A− A

→d N (0, 1)

In summary,

A w N(A,

τ2

N

)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 52 / 60

Controlling the MC-Integration Error

Fix the desired estimation precision (or relative precision): ε

Fix the desired probability of achieving the desired estimationprecision: 1− α

Run an “appropriate sized”pilot sample to calculate τ0

Ruben Zamar [email protected] () Module 12 Asymptotic Results 53 / 60

Controlling the MC-Integration Error

Fix the desired estimation precision (or relative precision): ε

Fix the desired probability of achieving the desired estimationprecision: 1− α

Run an “appropriate sized”pilot sample to calculate τ0

Ruben Zamar [email protected] () Module 12 Asymptotic Results 53 / 60

Controlling the MC-Integration Error

Fix the desired estimation precision (or relative precision): ε

Fix the desired probability of achieving the desired estimationprecision: 1− α

Run an “appropriate sized”pilot sample to calculate τ0

Ruben Zamar [email protected] () Module 12 Asymptotic Results 53 / 60

Controlling the MC-Integration Error

Notice that

P(∣∣A− A∣∣ < ε

)= P

(√N

∣∣∣∣ A− Aτ

∣∣∣∣ < √N ε

τ

)≈ 2Φ

(√N

ε

τ

)− 1

≈ 2Φ(√

τ0

)− 1

Ruben Zamar [email protected] () Module 12 Asymptotic Results 54 / 60

Controlling the MC-Integration Error

Solve for N in the following equation:

2Φ(√

τ0

)− 1 = 1− α

N∗ =

[τ0ε

Φ−1(1− α

2

)]2

Ruben Zamar [email protected] () Module 12 Asymptotic Results 55 / 60

Numerical Example

Example: Calculate the integral

I =1√2π

∫ 1.5

−0.5exp

(−12x2)dx

within a 0.001 error margin with probability 0.999.

In this case ε = 0.001 and α = 0.0001, b− a = 2,

g (x) = exp(−12x2)

and

A =12

∫ 1.5

−0.5exp

(−12x2)dx

Ruben Zamar [email protected] () Module 12 Asymptotic Results 56 / 60

Numerical Example

Example: Calculate the integral

I =1√2π

∫ 1.5

−0.5exp

(−12x2)dx

within a 0.001 error margin with probability 0.999.

In this case ε = 0.001 and α = 0.0001, b− a = 2,

g (x) = exp(−12x2)

and

A =12

∫ 1.5

−0.5exp

(−12x2)dx

Ruben Zamar [email protected] () Module 12 Asymptotic Results 56 / 60

Numerical Example (Continued)

Hence

I =2√2πA =

√2πA

We generate (using R) 100000 iid Unif (−0.5, 1.5) random variablesUi to compute

τ20 =11000

N

∑i=1

[exp

(−12U2i

)]2−[11000

N

∑i=1exp

(−12U2i

)]2

=11000

N

∑i=1exp

(−U2i

)−[11000

N

∑i=1exp

(−12U2i

)]2= 0.0457856

τ0 = 0.2139757

Ruben Zamar [email protected] () Module 12 Asymptotic Results 57 / 60

Numerical Example (Continued)

Hence

I =2√2πA =

√2πA

We generate (using R) 100000 iid Unif (−0.5, 1.5) random variablesUi to compute

τ20 =11000

N

∑i=1

[exp

(−12U2i

)]2−[11000

N

∑i=1exp

(−12U2i

)]2

=11000

N

∑i=1exp

(−U2i

)−[11000

N

∑i=1exp

(−12U2i

)]2= 0.0457856

τ0 = 0.2139757

Ruben Zamar [email protected] () Module 12 Asymptotic Results 57 / 60

Numerical Example (Continued)

Therefore

N∗ =

[τ0ε

Φ−1(1− α

2

)]2

=

[0.21397570.001

Φ−1(1− 0.0001

2

)]2N∗ = 693, 044 (rounding up)

Ruben Zamar [email protected] () Module 12 Asymptotic Results 58 / 60

Numerical Example (Continued)

Now we generate 693, 044 iid Unif (−0.5, 1.5) random variables Ui tocompute

A =1N∗

N ∗

∑i=1exp

(−12U2i

)= 0.782887

Therefore

I =

√2πA =

√2π0.7830537 = 0.624654

Compare with

pnorm(1.5)− pnorm(−0.5) = 0.6246553

∣∣I − [pnorm(1.5)− pnorm(−0.5)]∣∣ = 0.000002

Ruben Zamar [email protected] () Module 12 Asymptotic Results 59 / 60

Numerical Example (Continued)

Now we generate 693, 044 iid Unif (−0.5, 1.5) random variables Ui tocompute

A =1N∗

N ∗

∑i=1exp

(−12U2i

)= 0.782887

Therefore

I =

√2πA =

√2π0.7830537 = 0.624654

Compare with

pnorm(1.5)− pnorm(−0.5) = 0.6246553

∣∣I − [pnorm(1.5)− pnorm(−0.5)]∣∣ = 0.000002

Ruben Zamar [email protected] () Module 12 Asymptotic Results 59 / 60

Numerical Example (Continued)

Now we generate 693, 044 iid Unif (−0.5, 1.5) random variables Ui tocompute

A =1N∗

N ∗

∑i=1exp

(−12U2i

)= 0.782887

Therefore

I =

√2πA =

√2π0.7830537 = 0.624654

Compare with

pnorm(1.5)− pnorm(−0.5) = 0.6246553

∣∣I − [pnorm(1.5)− pnorm(−0.5)]∣∣ = 0.000002Ruben Zamar [email protected] () Module 12 Asymptotic Results 59 / 60

Numerical Example (Continued)

In practice we don’t know I

Still we may wish to check if the desired precision has been likelyachieved

Recompute N∗∗ using the full Monte Carlo sample

N∗∗ = 692, 553

Compare N∗∗ with N∗

N∗∗ = 692, 553 < N∗ = 693, 044

Ruben Zamar [email protected] () Module 12 Asymptotic Results 60 / 60

Numerical Example (Continued)

In practice we don’t know I

Still we may wish to check if the desired precision has been likelyachieved

Recompute N∗∗ using the full Monte Carlo sample

N∗∗ = 692, 553

Compare N∗∗ with N∗

N∗∗ = 692, 553 < N∗ = 693, 044

Ruben Zamar [email protected] () Module 12 Asymptotic Results 60 / 60

Numerical Example (Continued)

In practice we don’t know I

Still we may wish to check if the desired precision has been likelyachieved

Recompute N∗∗ using the full Monte Carlo sample

N∗∗ = 692, 553

Compare N∗∗ with N∗

N∗∗ = 692, 553 < N∗ = 693, 044

Ruben Zamar [email protected] () Module 12 Asymptotic Results 60 / 60

Numerical Example (Continued)

In practice we don’t know I

Still we may wish to check if the desired precision has been likelyachieved

Recompute N∗∗ using the full Monte Carlo sample

N∗∗ = 692, 553

Compare N∗∗ with N∗

N∗∗ = 692, 553 < N∗ = 693, 044

Ruben Zamar [email protected] () Module 12 Asymptotic Results 60 / 60