guide:452fd94468: Difference between revisions

From Stochiki
No edit summary
mNo edit summary
Line 6: Line 6:
\newcommand{\NA}{{\rm NA}}
\newcommand{\NA}{{\rm NA}}
\newcommand{\mathds}{\mathbb}</math></div>
\newcommand{\mathds}{\mathbb}</math></div>
\label{sec 9.4}
 
We have seen in Section~\ref{sec 9.3} that the distribution function for the sum of
We have seen in Section~\ref{sec 9.3} that the distribution function for the sum of
a large number <math>n</math> of independent discrete random variables with mean <math>\mu</math> and
a large number <math>n</math> of independent discrete random variables with mean <math>\mu</math> and
Line 14: Line 14:
also holds true for continuous random variables having a common density function.
also holds true for continuous random variables having a common density function.


 
Let us begin by looking at some examples to see whether such a result is even plausible.
Let us begin by looking at some examples to see whether such a result is even
plausible.


===Standardized Sums===
===Standardized Sums===
Line 42: Line 40:
function for <math>S_n</math> (see Figure \ref{fig 9.7}).
function for <math>S_n</math> (see Figure \ref{fig 9.7}).


<div id="PSfig9-7" class="d-flex justify-content-center">
<div id="fig 9.7" class="d-flex justify-content-center">
[[File:guide_e6d15_PSfig9-7.ps | 400px | thumb | ]]
[[File:guide_e6d15_PSfig9-7.png | 400px | thumb | Density function for <math>S^*_n</math> (uniform case, <math>n = 2, 3, 4, 10</math>). ]]
</div>  
</div>  
'''Example'''  
'''Example'''  
Let us do the same thing, but now choose numbers from the interval
Let us do the same thing, but now choose numbers from the interval
Line 76: Line 75:
9.9}.
9.9}.


<div id="PSfig9-9" class="d-flex justify-content-center">
<div id="fig 9.9" class="d-flex justify-content-center">
[[File:guide_e6d15_PSfig9-9.ps | 400px | thumb |  ]]
[[File:guide_e6d15_PSfig9-9.png | 400px | thumb |Density function for <math>S^*_n</math> (exponential case,
<math>n = 2, 3, 10, 30</math>, <math>\lambda = 1</math>). ]]
</div>  
</div>  
These examples make it seem plausible that the density function for the
These examples make it seem plausible that the density function for the
normalized random variable <math>S_n^*</math> for large <math>n</math> will look very much like the
normalized random variable <math>S_n^*</math> for large <math>n</math> will look very much like the
Line 99: Line 100:
We shall give a proof of this theorem in Section \ref{sec
We shall give a proof of this theorem in Section \ref{sec
10.3}.  We will now look at some examples.
10.3}.  We will now look at some examples.
<span id="exam 9.10"/>
<span id="exam 9.10"/>
'''Example'''  
'''Example'''  
Line 109: Line 111:
standard deviation <math>\sigma = .0002</math> (so, if the errors are approximately normally  
standard deviation <math>\sigma = .0002</math> (so, if the errors are approximately normally  
distributed, then his measurements are within 1 foot of the correct distance about
distributed, then his measurements are within 1 foot of the correct distance about
65\% of the time).  What can he say about the average?
65% of the time).  What can he say about the average?




Line 198: Line 200:
</math>
</math>
The interval <math>(\bar \mu - .0002, \bar \mu
The interval <math>(\bar \mu - .0002, \bar \mu
+ .0002)</math> is called ''the 99.7\% confidence interval'' for <math>\mu</math> (see
+ .0002)</math> is called ''the 99.7% confidence interval'' for <math>\mu</math> (see
[[guide:146f3c94d0#exam 9.4.1 |Example]]).
[[guide:146f3c94d0#exam 9.4.1 |Example]]).


Line 247: Line 249:
<math>n = 8</math> the <math>t</math>-density and normal density are practically the same  
<math>n = 8</math> the <math>t</math>-density and normal density are practically the same  
(see Figure \ref{fig 9.12}).
(see Figure \ref{fig 9.12}).
<div id="PSfig9-12" class="d-flex justify-content-center">
<div id="fig 9.12" class="d-flex justify-content-center">
[[File:guide_e6d15_PSfig9-12.ps | 400px | thumb | ]]
[[File:guide_e6d15_PSfig9-12.png | 400px | thumb | Graph of <math>t-</math>density for <math>n= 1, 3, 8</math> and the normal density with <math>\mu = 0, \sigma = 1</math>. ]]
</div>  
</div>  
\exercises
 
\indent ''Notes on computer problems'':
'''Notes on computer problems''':
\begin{description}
 
\item[(a)] <math>\ </math>Simulation: Recall (see [[guide:D26a5cb8f7#cor 5.2 |Corollary]]) that
<ul style="list-style-type:lower-alpha">
<li>
<math>\ </math>Simulation: Recall (see [[guide:D26a5cb8f7#cor 5.2 |Corollary]]) that


<math display="block">
<math display="block">
Line 271: Line 275:
X = \sigma\sqrt {-2\log(rnd)} \cos 2\pi(rnd) + \mu\ .
X = \sigma\sqrt {-2\log(rnd)} \cos 2\pi(rnd) + \mu\ .
</math>
</math>
\item[(b)] <math>\ </math>Bar graphs: you should aim for about 20 to 30 bars (of equal width) in   
</li>
 
<li><math>\ </math>Bar graphs: you should aim for about 20 to 30 bars (of equal width) in   
your graph.  You can achieve this by a good choice of the range <math>[x{\rm min}, x{\rm min}]</math> and the
your graph.  You can achieve this by a good choice of the range <math>[x{\rm min}, x{\rm min}]</math> and the
number of bars (for instance, <math>[\mu - 3\sigma, \mu + 3\sigma]</math> with 30 bars will work in many
number of bars (for instance, <math>[\mu - 3\sigma, \mu + 3\sigma]</math> with 30 bars will work in many
cases).  Experiment!
cases).  Experiment!
\end{description}
</li>
\vskip .1in
</ul>
 
==General references==
 
\choice{}{\setcounter{chapter}{9}==General references==
{{cite web |url=https://math.dartmouth.edu/~prob/prob/prob.pdf |title=Grinstead and Snell’s Introduction to Probability |last=Doyle |first=Peter G.|date=2006 |access-date=June 6, 2024}}
{{cite web |url=https://math.dartmouth.edu/~prob/prob/prob.pdf |title=Grinstead and Snell’s Introduction to Probability |last=Doyle |first=Peter G.|date=2006 |access-date=June 6, 2024}}
==Notes==
==Notes==
{{Reflist|group=Notes}}
{{Reflist|group=Notes}}

Revision as of 03:00, 11 June 2024

[math] \newcommand{\NA}{{\rm NA}} \newcommand{\mat}[1]{{\bf#1}} \newcommand{\exref}[1]{\ref{##1}} \newcommand{\secstoprocess}{\all} \newcommand{\NA}{{\rm NA}} \newcommand{\mathds}{\mathbb}[/math]

We have seen in Section~\ref{sec 9.3} that the distribution function for the sum of a large number [math]n[/math] of independent discrete random variables with mean [math]\mu[/math] and variance [math]\sigma^2[/math] tends to look like a normal density with mean [math]n\mu[/math] and variance [math]n\sigma^2[/math]. What is remarkable about this result is that it holds for any distribution with finite mean and variance. We shall see in this section that the same result also holds true for continuous random variables having a common density function.

Let us begin by looking at some examples to see whether such a result is even plausible.

Standardized Sums

Example Suppose we choose [math]n[/math] random numbers from the interval [math][0,1][/math] with uniform density. Let [math]X_1[/math], [math]X_2[/math], \dots, [math]X_n[/math] denote these choices, and [math]S_n = X_1 + X_2 +\cdots+ X_n[/math] their sum. We saw in Example that the density function for [math]S_n[/math] tends to have a normal shape, but is centered at [math]n/2[/math] and is flattened out. In order to compare the shapes of these density functions for different values of [math]n[/math], we proceed as in the previous section: we standardize [math]S_n[/math] by defining

[[math]] S_n^* = \frac {S_n - n\mu}{\sqrt n \sigma}\ . [[/math]]

Then we see that for all [math]n[/math] we have

[[math]] \begin{eqnarray*} E(S_n^*) & = & 0\ , \\ V(S_n^*) & = & 1\ . \end{eqnarray*} [[/math]]

The density function for [math]S_n^*[/math] is just a standardized version of the density function for [math]S_n[/math] (see Figure \ref{fig 9.7}).

Density function for [math]S^*_n[/math] (uniform case, [math]n = 2, 3, 4, 10[/math]).

Example Let us do the same thing, but now choose numbers from the interval [math][0,+\infty)[/math] with an exponential density with parameter [math]\lambda[/math]. Then (see Example)


[[math]] \begin{eqnarray*} \mu & = & E(X_i) = \frac 1\lambda\ , \\ \sigma^2 & = & V(X_j) = \frac 1{\lambda^2}\ . \end{eqnarray*} [[/math]]


Here we know the density function for [math]S_n[/math] explicitly (see Section \ref{sec 7.2}). We can use Corollary to calculate the density function for [math]S_n^*[/math]. We obtain


[[math]] \begin{eqnarray*} f_{S_n}(x) & = & \frac {\lambda e^{-\lambda x}(\lambda x)^{n - 1}}{(n - 1)!}\ , \\ f_{S_n^*}(x) & = & \frac {\sqrt n}\lambda f_{S_n} \left( \frac {\sqrt n x + n}\lambda \right)\ . \end{eqnarray*} [[/math]]

The graph of the density function for [math]S_n^*[/math] is shown in Figure \ref{fig 9.9}.

Density function for [math]S^*_n[/math] (exponential case, [math]n = 2, 3, 10, 30[/math], [math]\lambda = 1[/math]).

These examples make it seem plausible that the density function for the normalized random variable [math]S_n^*[/math] for large [math]n[/math] will look very much like the normal density with mean 0 and variance 1 in the continuous case as well as in the discrete case. The Central Limit Theorem makes this statement precise.

Central Limit Theorem

Theorem

(Central Limit Theorem) Let [math]S_n = X_1 + X_2 +\cdots+ X_n[/math] be the sum of [math]n[/math] independent continuous random variables with common density function [math]p[/math] having expected value [math]\mu[/math] and variance [math]\sigma^2[/math]. Let [math]S_n^* = (S_n - n\mu)/\sqrt n \sigma[/math]. Then we have, for all [math]a \lt b[/math],

[[math]] \lim_{n \to \infty} P(a \lt S_n^* \lt b) = \frac 1{\sqrt{2\pi}} \int_a^b e^{-x^2/2}\, dx\ . [[/math]]


We shall give a proof of this theorem in Section \ref{sec 10.3}. We will now look at some examples.

Example Suppose a surveyor wants to measure a known distance, say of 1 mile, using a transit and some method of triangulation. He knows that because of possible motion of the transit, atmospheric distortions, and human error, any one measurement is apt to be slightly in error. He plans to make several measurements and take an average. He assumes that his measurements are independent random variables with a common distribution of mean [math]\mu = 1[/math] and standard deviation [math]\sigma = .0002[/math] (so, if the errors are approximately normally distributed, then his measurements are within 1 foot of the correct distance about 65% of the time). What can he say about the average?


He can say that if [math]n[/math] is large, the average [math]S_n/n[/math] has a density function that is approximately normal, with mean [math]\mu = 1[/math] mile, and standard deviation [math]\sigma = .0002/\sqrt n[/math] miles. How many measurements should he make to be reasonably sure that his average lies within .0001 of the true value? The Chebyshev inequality says

[[math]] P\left(\left| \frac {S_n}n - \mu \right| \geq .0001 \right) \leq \frac {(.0002)^2}{n(10^{-8})} = \frac 4n\ , [[/math]]

so that we must have [math]n \ge 80[/math] before the probability that his error is less than .0001 exceeds .95.


We have already noticed that the estimate in the Chebyshev inequality is not always a good one, and here is a case in point. If we assume that [math]n[/math] is large enough so that the density for [math]S_n[/math] is approximately normal, then we have


[[math]] \begin{eqnarray*} P\left(\left| \frac {S_n}n - \mu \right| \lt .0001 \right) &=& P\bigl(-.5\sqrt{n} \lt S_n^* \lt +.5\sqrt{n}\bigr) \\ &\approx& \frac 1{\sqrt{2\pi}} \int_{-.5\sqrt{n}}^{+.5\sqrt{n}} e^{-x^2/2}\, dx\ , \end{eqnarray*} [[/math]]

and this last expression is greater than .95 if [math].5\sqrt{n} \ge 2.[/math] This says that it suffices to take [math]n = 16[/math] measurements for the same results. This second calculation is stronger, but depends on the assumption that [math]n = 16[/math] is large enough to establish the normal density as a good approximation to [math]S_n^*[/math], and hence to [math]S_n[/math]. The Central Limit Theorem here says nothing about how large [math]n[/math] has to be. In most cases involving sums of independent random variables, a good rule of thumb is that for [math]n \ge 30[/math], the approximation is a good one. In the present case, if we assume that the errors are approximately normally distributed, then the approximation is probably fairly good even for [math]n = 16[/math].


Estimating the Mean

Example Now suppose our surveyor is measuring an unknown distance with the same instruments under the same conditions. He takes 36 measurements and averages them. How sure can he be that his measurement lies within .0002 of the true value? Again using the normal approximation, we get

[[math]] \begin{eqnarray*} P\left(\left|\frac {S_n}n - \mu\right| \lt .0002 \right) &=& P\bigl(|S_n^*| \lt .5\sqrt n\bigr) \\ &\approx& \frac 2{\sqrt{2\pi}} \int_{-3}^3 e^{-x^2/2}\, dx \\ &\approx& .997\ . \end{eqnarray*} [[/math]]


This means that the surveyor can be 99.7 percent sure that his average is within .0002 of the true value. To improve his confidence, he can take more measurements, or require less accuracy, or improve the quality of his measurements (i.e., reduce the variance [math]\sigma^2[/math]). In each case, the Central Limit Theorem gives quantitative information about the confidence of a measurement process, assuming always that the normal approximation is valid.


Now suppose the surveyor does not know the mean or standard deviation of his measurements, but assumes that they are independent. How should he proceed?


Again, he makes several measurements of a known distance and averages them. As before, the average error is approximately normally distributed, but now with unknown mean and variance.

Sample Mean

If he knows the variance [math]\sigma^2[/math] of the error distribution is .0002, then he can estimate the mean [math]\mu[/math] by taking the average, or sample mean of, say, 36 measurements:

[[math]] \bar \mu = \frac {x_1 + x_2 +\cdots+ x_n}n\ , [[/math]]

where [math]n = 36[/math]. Then, as before, [math]E(\bar \mu) = \mu[/math]. Moreover, the preceding argument shows that

[[math]] P(|\bar \mu - \mu| \lt .0002) \approx .997\ . [[/math]]

The interval [math](\bar \mu - .0002, \bar \mu + .0002)[/math] is called the 99.7% confidence interval for [math]\mu[/math] (see Example).

Sample Variance

If he does not know the variance [math]\sigma^2[/math] of the error distribution, then he can estimate [math]\sigma^2[/math] by the sample variance:

[[math]] \bar \sigma^2 = \frac {(x_1 - \bar \mu)^2 + (x_2 - \bar \mu)^2 +\cdots+ (x_n - \bar \mu)^2}n\ , [[/math]]

where [math]n = 36[/math]. The Law of Large Numbers, applied to the random variables [math](X_i - \bar \mu)^2[/math], says that for large [math]n[/math], the sample variance [math]\bar \sigma^2[/math] lies close to the variance [math]\sigma^2[/math], so that the surveyor can use [math]\bar \sigma^2[/math] in place of [math]\sigma^2[/math] in the argument above.


Experience has shown that, in most practical problems of this type, the sample variance is a good estimate for the variance, and can be used in place of the variance to determine confidence levels for the sample mean. This means that we can rely on the Law of Large Numbers for estimating the variance, and the Central Limit Theorem for estimating the mean.


We can check this in some special cases. Suppose we know that the error distribution is normal, with unknown mean and variance. Then we can take a sample of [math]n[/math] measurements, find the sample mean [math]\bar \mu[/math] and sample variance [math]\bar \sigma^2[/math], and form

[[math]] T_n^* = \frac {S_n - n\bar\mu}{\sqrt{n}\bar\sigma}\ , [[/math]]

where [math]n = 36[/math]. We expect [math]T_n^*[/math] to be a good approximation for [math]S_n^*[/math] for large [math]n[/math].

[math]t[/math]-Density

The statistician W. S. Gosset[Notes 1] has shown that in this case [math]T_n^*[/math] has a density function that is not normal but rather a [math]t[/math]-density with [math]n[/math] degrees of freedom. (The number [math]n[/math] of degrees of freedom is simply a parameter which tells us which [math]t[/math]-density to use.) In this case we can use the [math]t[/math]-density in place of the normal density to determine confidence levels for [math]\mu[/math]. As [math]n[/math] increases, the [math]t[/math]-density approaches the normal density. Indeed, even for [math]n = 8[/math] the [math]t[/math]-density and normal density are practically the same (see Figure \ref{fig 9.12}).

Graph of [math]t-[/math]density for [math]n= 1, 3, 8[/math] and the normal density with [math]\mu = 0, \sigma = 1[/math].

Notes on computer problems:

  • [math]\ [/math]Simulation: Recall (see Corollary) that
    [[math]] X = F^{-1}(rnd) [[/math]]
    will simulate a random variable with density [math]f(x)[/math] and distribution
    [[math]] F(X) = \int_{-\infty}^x f(t)\, dt\ . [[/math]]
    In the case that [math]f(x)[/math] is a normal density function with mean [math]\mu[/math] and standard deviation [math]\sigma[/math], where neither [math]F[/math] nor [math]F^{-1}[/math] can be expressed in closed form, use instead
    [[math]] X = \sigma\sqrt {-2\log(rnd)} \cos 2\pi(rnd) + \mu\ . [[/math]]
  • [math]\ [/math]Bar graphs: you should aim for about 20 to 30 bars (of equal width) in your graph. You can achieve this by a good choice of the range [math][x{\rm min}, x{\rm min}][/math] and the number of bars (for instance, [math][\mu - 3\sigma, \mu + 3\sigma][/math] with 30 bars will work in many cases). Experiment!

General references

Doyle, Peter G. (2006). "Grinstead and Snell's Introduction to Probability" (PDF). Retrieved June 6, 2024.

Notes

  1. W. S. Gosset discovered the distribution we now call the [math]t[/math]-distribution while working for the Guinness Brewery in Dublin. He wrote under the pseudonym “Student.” The results discussed here first appeared in Student, “The Probable Error of a Mean,” Biometrika, vol. 6 (1908), pp. 1-24.