guide:523e6267ef: Difference between revisions

From Stochiki
No edit summary
 
mNo edit summary
 
(One intermediate revision by one other user not shown)
Line 1: Line 1:
<div class="d-none"><math>
\newcommand{\NA}{{\rm NA}}
\newcommand{\mat}[1]{{\bf#1}}
\newcommand{\exref}[1]{\ref{##1}}
\newcommand{\secstoprocess}{\all}
\newcommand{\NA}{{\rm NA}}
\newcommand{\mathds}{\mathbb}</math></div>


In the previous section we have seen how to simulate experiments with a whole
continuum of possible outcomes and have gained some experience in thinking
about such experiments.  Now we turn to the general problem of assigning
probabilities to the outcomes and events in such experiments.  We shall
restrict our attention here to those experiments whose sample space can be
taken as a suitably chosen subset of the line, the plane, or some other
Euclidean space.  We begin with some simple examples.
===Spinners===
<span id="exam 2.2.1"/>
'''Example'''
The spinner experiment described in [[guide:A070937c41#exam 2.1.1 |Example]] has the interval
<math>[0, 1)</math> as  the set of possible outcomes.  We would like to construct a probability model in
which  each outcome is equally likely to occur.  We saw that in such a model, it is necessary
to assign the probability 0 to each outcome.  This does not at all mean that the probability
of ''every'' event must be zero.  On the contrary, if we let the random variable <math>X</math>
denote the outcome, then the probability
<math display="block">
P(\,0 \leq X \leq 1)
</math>
that the head of the spinner comes to rest ''somewhere'' in the circle, should be
equal to 1.  Also, the probability that it comes to rest in the upper half of the circle
should be the same as for the lower half, so that
<math display="block">
P\biggl(0 \leq X  <  \frac12\biggr) = P\biggl(\frac12 \leq X  <
1\biggr) = \frac12\ .
</math>
More generally, in our model, we would like the equation
<math display="block">
P(c \leq X  <  d) = d - c
</math>
to be true for every choice of <math>c</math> and <math>d</math>.
If we let <math>E = [c, d]</math>, then we can write the above formula in the form
<math display="block">
P(E) = \int_E f(x)\,dx\ ,
</math>
where <math>f(x)</math> is the constant function with value 1. 
This should remind the reader of the corresponding formula in the discrete case
for the probability of an event:
<math display="block">
P(E) = \sum_{\omega \in E} m(\omega)\ .
</math>
The difference is that in the continuous case, the quantity being integrated, <math>f(x)</math>,
is not the probability of the outcome <math>x</math>.  (However, if one uses infinitesimals, one
can consider <math>f(x)\,dx</math> as the probability of the outcome <math>x</math>.) 
In the continuous case, we will use the following convention.  If the set of outcomes is a
set of real numbers, then the individual outcomes will be referred to by small Roman letters
such as <math>x</math>.  If the set of outcomes is a subset of <math>R^2</math>, then the individual
outcomes will be denoted by <math>(x, y)</math>.  In either case, it may be more convenient to refer to
an individual outcome by using <math>\omega</math>, as in Chapter [[guide:4f3a4e96c3|Discrete Probability Distributions]].
<div id="fig 2.16.5" class="d-flex justify-content-center">
[[File:guide_e6d15_PSfig2-16-5.png | 400px | thumb |Spinner experiment.  ]]
</div>
[[#fig 2.16.5|Figure]] shows the results of 1000 spins of the spinner.  The function
<math>f(x)</math> is also shown in the figure.  The reader will note that the area under <math>f(x)</math> and
above a given interval is approximately equal to the fraction of outcomes that fell in
that interval.  The function <math>f(x)</math> is called the ''density function'' of the random variable <math>X</math>.  The fact that the area under <math>f(x)</math> and above an
interval corresponds to a probability is  the defining property of density functions.  A
precise definition of density functions will be given shortly.
===Darts===
<span id="exam 2.2.2"/>
'''Example'''
A game of darts involves throwing a dart at a circular target of ''unit
radius.''  Suppose we throw a dart once so that it hits the target, and we
observe where it lands.
To describe the possible outcomes of this experiment, it is natural to take as
our sample space the set <math>\Omega</math> of all the points in the target.  It
is convenient to describe these points by their rectangular coordinates,
relative to a coordinate system with origin at the center of the target, so
that each pair <math>(x,y)</math> of coordinates with <math>x^2 + y^2 \leq 1</math> describes a
possible outcome of the experiment.  Then <math>\Omega = \{\,(x,y) : x^2 + y^2 \leq
1\,\}</math> is a subset of the Euclidean plane, and the event <math>E = \{\,(x,y) : y  >
0\,\}</math>, for example, corresponds to the statement that the dart lands in the
upper half of the target, and so forth.  Unless there is reason to believe
otherwise (and with experts at the game there may well be!), it is natural to
assume that the coordinates are chosen ''at random.''  (When doing this with
a computer, each coordinate is chosen uniformly from the interval <math>[-1, 1]</math>. 
If the resulting point does not lie inside the unit circle, the point is not counted.) 
Then the arguments used in the preceding example show that the probability of any
elementary event, consisting of a single outcome, must be zero, and suggest that the
probability of the event that the dart lands in any subset <math>E</math> of the target
should be determined by what fraction of the target area lies in <math>E</math>.  Thus,
<math display="block">
P(E) = \frac{\mbox{area\ of}\ E}{\mbox{area\ of\ target}} = \frac{\mbox{area\ of}\
E}\pi\ .
</math>
This can be written in the form
<math display="block">
P(E) = \int_E f(x)\,dx\ ,
</math>
where <math>f(x)</math> is the constant function with value <math>1/\pi</math>.
In particular, if <math>E = \{\,(x,y) : x^2 + y^2 \leq a^2\,\}</math> is the event that
the dart lands within distance <math>a  <  1</math> of the center of the target, then
<math display="block">
P(E) = \frac{\pi a^2}\pi = a^2\ .
</math>
For example, the probability that the dart lies within a distance 1/2 of the
center is 1/4.
<span id="exam 2.2.3"/>
'''Example'''
In the dart game considered above, suppose that, instead of
observing where the dart lands, we observe how far it lands from the center of
the target.
In this case, we take as our sample space the set <math>\Omega</math> of all circles with
centers at the center of the target.  It is convenient to describe these
circles by their radii, so that each circle is identified by its radius <math>r</math>, <math>0
\leq r \leq 1</math>.  In this way, we may regard <math>\Omega</math> as the subset <math>[0,1]</math> of
the real line.
What probabilities should we assign to the events <math>E</math> of <math>\Omega</math>?  If
<math display="block">
E = \{\,r : 0 \leq r \leq a\,\}\ ,
</math>
then <math>E</math> occurs if the
dart lands within a distance <math>a</math> of the center, that is, within the circle of
radius <math>a</math>, and we saw in the previous example that under our assumptions the
probability of this event is given by
<math display="block">
P([0,a]) = a^2\ .
</math>
More generally, if
<math display="block">
E = \{\,r : a \leq r \leq b\,\}\ ,
</math>
then by our basic assumptions,
<math display="block">
\begin{eqnarray*}
P(E) = P([a,b]) & = & P([0,b]) - P([0,a]) \\
            & = & b^2 - a^2 \\
            & = & (b - a)(b + a) \\
            & = & 2(b - a)\frac{(b + a)}2\ .
\end{eqnarray*}
</math>
Thus, <math>P(E) = </math>2(length of <math>E</math>)(midpoint of <math>E</math>).  Here we see that the
probability assigned to the interval <math>E</math> depends not only on its length but
also on its midpoint (i.e., not only on how long it is, but also on where it
is).  Roughly speaking, in this experiment, events of the form <math>E = [a,b]</math> are
more likely if they are near the rim of the target and less likely if they are
near the center.  (A common experience for beginners!  The conclusion might
well be different if the beginner is replaced by an expert.)
Again we can simulate this by computer. 
We divide the target area into ten concentric regions of equal thickness.
<div id="fig 2.15" class="d-flex justify-content-center">
[[File:guide_e6d15_PSfig2-15.png | 400px | thumb |Distribution of dart distances in 400 throws.  ]]
</div>
The computer program ''' Darts''' throws <math>n</math> darts and records what
fraction of the total falls in each of these concentric regions.  The
program ''' Areabargraph''' then plots a bar graph with the ''area'' of
the <math>i</math>th bar equal to the fraction of the total falling in the <math>i</math>th region.
Running the program for 1000 darts resulted in the bar graph of [[#fig 2.15|Figure]].
Note that here the heights of the bars are not all equal, but grow
approximately linearly with <math>r</math>.  In fact, the linear function <math>y = 2r</math> appears
to fit our bar graph quite well.  This suggests that the probability that the
dart falls within a distance <math>a</math> of the center should be given by the ''
area'' under the graph of the function <math>y = 2r</math> between 0 and <math>a</math>.  This area
is <math>a^2</math>, which agrees with the probability we have assigned above to this
event.
===Sample Space Coordinates===
These examples suggest that for continuous experiments of this sort we should
assign probabilities for the outcomes to fall in a given interval by means of the
area under a suitable function.
More generally, we suppose that suitable coordinates can be introduced into the
sample space <math>\Omega</math>, so that we can regard <math>\Omega</math> as a subset of
<math>''' R'''^n</math>.  We call such a sample space a ''continuous sample space.''  We let
<math>X</math> be a random variable which represents the outcome of the experiment.  Such a
random variable is called a ''continuous random variable.''  We then define a density function for <math>X</math> as follows. 
===Density Functions of Continuous Random Variables===
{{defncard|label=|id=|Let <math>X</math> be a continuous real-valued random variable.  A ''density function'' for
<math>X</math>  is a real-valued function <math>f</math> which satisfies
<math display="block">
P(a \le X \le b) = \int_a^b f(x)\,dx
</math>
for all <math>a,\ b \in ''' R'''</math>.}}
We note that it is ''not'' the case that all continuous real-valued random variables
possess density functions.  However, in this book, we will only consider continuous random
variables for which density functions exist.
In terms of the density <math>f(x)</math>, if <math>E</math> is a subset of
<math>{\mat R}</math>, then
<math display="block">
P(X \in E) = \int_Ef(x)\,dx\ .
</math>
The notation here assumes that <math>E</math> is a subset of <math>{\mat R}</math> for which <math>\int_E
f(x)\,dx</math> makes sense. 
<span id="exam 2.2.5"/>
'''Example'''
In the spinner experiment, we choose for our set of outcomes the
interval <math>0 \leq x  <  1</math>, and for our density function
<math display="block">
f(x) =  \left \{ \begin{array}{ll}
                    1, & \mbox{if $0 \leq x  <  1$,} \\
                    0, & \mbox{otherwise.}
                    \end{array}
            \right.
</math>
If <math>E</math> is the event that the head of the spinner falls in the upper half of the
circle, then <math>E = \{\,x : 0 \leq x \leq 1/2\,\}</math>, and so
<math display="block">
P(E) = \int_0^{1/2} 1\,dx = \frac12\ .
</math>
More generally, if <math>E</math> is the event that the head falls in the interval
<math>[a,b]</math>, then
<math display="block">
P(E) = \int_a^b 1\,dx = b - a\ .
</math>
'''Example'''
<span id="exam 2.2.6"/>
In the first dart game experiment, we choose for our sample space a disc of
unit radius in the plane and for our density function the function
<math display="block">
f(x,y) = \left \{ \begin{array}{ll}
              1/\pi, & \mbox{if $x^2 + y^2 \leq 1$,} \\
              0,    & \mbox{otherwise.}
                \end{array}
        \right.
</math>
The probability that the dart lands inside the subset <math>E</math> is then given by
<math display="block">
\begin{eqnarray*}
P(E) & = & \int\,\int_E \frac1\pi \,dx\,dy\\
    & = & \frac1\pi \cdot (\mbox{area\,\,\, of}\,\,\,E)\ .
\end{eqnarray*}
</math>
In these two examples, the density function is constant and does
not depend on the particular outcome.  It is often the case that experiments in which the
coordinates are chosen ''at random'' can be described by ''constant''
density functions, and, as in Section \ref{sec 1.2},
we call such density functions ''uniform'' or ''equiprobable.'' Not all experiments are of this type, however.
<span id="exam 2.2.7"/>
'''Example'''
In the second dart game experiment, we choose for our sample space the unit interval on the real line and for our density the function
<math display="block">
f(r) = \left \{ \begin{array}{ll}
                2r, & \mbox{if  $0  <  r  <  1$,} \\
                0,  & \mbox{otherwise.}
                \end{array}
        \right.
</math>
Then the probability that the dart lands at distance <math>r</math>, <math>a \leq r \leq b</math>,
from the center of the target is given by
<math display="block">
\begin{eqnarray*}
P([a,b]) & = & \int_a^b 2r\,dr\\
      & = & b^2 - a^2\ .
\end{eqnarray*}
</math>
Here again, since the density is small when
<math>r</math> is near 0 and large when <math>r</math> is near 1, we see that in this experiment the
dart is more likely to land near the rim of the target than near the center.
In terms of the bar graph of [[#exam 2.2.3 |Example]], the heights of the bars
approximate the density function, while the areas of the bars approximate the
probabilities of the subintervals (see [[#fig 2.15|Figure]]).
We see in this example that, unlike the case of discrete sample spaces, the
value <math>f(x)</math> of the density function for the outcome <math>x</math>
is ''not'' the probability of <math>x</math> occurring (we have seen that this
probability is always 0) and in general <math>f(x)</math> is ''not a probability
at all.''  In this example, if we take <math>\lambda = 2</math> then <math>f(3/4) = 3/2</math>,
which being bigger than 1, cannot be a probability.
Nevertheless, the density function <math>f</math> does contain all the
probability information about the experiment, since the probabilities of all
events can be derived from it.  In particular, the probability that the outcome
of the experiment falls in an interval <math>[a,b]</math> is given by
<math display="block">
P([a,b]) = \int_a^b f(x)\,dx\ ,
</math>
that is, by the ''area'' under the graph of the density function in the
interval <math>[a,b]</math>.  Thus, there is a close connection here between probabilities
and areas.  We have been guided by this close connection in making up our bar
graphs; each bar is chosen so that its ''area,'' and not its height,
represents the relative frequency of occurrence, and hence estimates the
probability of the outcome falling in the associated interval.
In the language of the calculus, we can say that the probability of occurrence
of an event of the form <math>[x, x + dx]</math>, where <math>dx</math> is small,
is approximately given by
<math display="block">
P([x, x+dx]) \approx f(x)dx\ ,
</math>
that is, by the area of the rectangle under the graph of <math>f</math>.  Note that as
<math>dx \to 0</math>, this probability <math>\to 0</math>, so that the probability
<math>P(\{x\})</math> of a single point is again 0, as in [[#exam 2.2.1 |Example]].
A glance at the graph of a density function tells us immediately
which events of an experiment are more likely.  Roughly speaking, we can say
that where the density is large the events are more likely, and where it is
small the events are less likely.  In [[guide:A070937c41#exam 2.1.4.5 |Example]] the density function
is largest at 1.  Thus, given the two intervals <math>[0, a]</math> and <math>[1, 1+a]</math>, where <math>a</math> is
a small positive real number, we see that <math>X</math> is more likely to take on a value in the
second interval than in the first.
===Cumulative Distribution Functions of Continuous Random Variables===
We have seen that density functions are useful when considering continuous random
variables.  There is another kind of function, closely related to these density functions,
which is also of great importance.  These functions are called ''cumulative
distribution'' functions.
{{defncard|label=|id=|Let <math>X</math> be a continuous real-valued random variable.  Then the cumulative distribution
function of <math>X</math> is defined by the equation
<math display="block">
F_X(x) = P(X \le x)\ .
</math>
}}
If <math>X</math> is a continuous real-valued random variable which possesses a density function,
then it also has a cumulative distribution function, and the following theorem shows that the
two functions are related  in a very nice way.
{{proofcard|Theorem|theorem-1|Let <math>X</math> be a continuous real-valued random variable with density function
<math>f(x)</math>.  Then the function defined by
<math display="block">
F(x) = \int_{-\infty}^x f(t)\,dt
</math>
is the cumulative distribution function of <math>X</math>.  Furthermore, we have
<math display="block">
{{d\ }\over{dx}} F(x) = f(x)\ .
</math>|By definition,
<math display="block">
F(x) = P(X \le x)\ .
</math>
Let <math>E = (-\infty, x]</math>.  Then
<math display="block">
P(X \le x) = P(X \in E)\ ,
</math>
which equals
<math display="block">
\int_{-\infty}^x f(t)\,dt\ .
</math>
Applying the Fundamental Theorem of Calculus to the first equation in the statement of
the theorem yields the second statement.}}
In many experiments, the density function of the relevant random variable is easy to
write down.  However, it is quite often the case that the cumulative distribution function
is easier to obtain than the density function.  (Of course, once we have the cumulative
distribution function, the density function can easily be obtained by differentiation, as
the above theorem shows.)  We now give some examples which exhibit this phenomenon.
<span id="exam 2.2.7.1"/>
'''Example'''
A real number is chosen at random from <math>[0, 1]</math> with uniform probability, and then this
number is squared.  Let <math>X</math> represent the result.  What is the cumulative distribution
function of <math>X</math>?  What is the density of <math>X</math>?
We begin by letting <math>U</math> represent the chosen real number.  Then <math>X = U^2</math>.  If <math>0 \le x
\le 1</math>, then we have
<math display="block">
\begin{eqnarray*}
F_X(x) & = & P(X \le x) \\
& = & P(U^2 \le x) \\
& = & P(U \le \sqrt x) \\
& = & \sqrt x\ .
\end{eqnarray*}
</math>
It is clear that <math>X</math> always takes on a value between 0 and 1, so the cumulative
distribution function of <math>X</math> is given by
<math display="block">
F_X(x) = \left \{ \begin{array}{ll}
                                0, & \mbox{if $x \le 0$}, \\
                        {\sqrt x}, & \mbox{if $0 \le x \le 1$}, \\
                                1, & \mbox{if $x \ge 1$}.
                  \end{array}
          \right.
</math>
From this we easily calculate that the density function of <math>X</math> is
<math display="block">
f_X(x) = \left \{ \begin{array}{ll}
                                    0, & \mbox{if $x \le 0$}, \\
                        1/(2{\sqrt x}), & \mbox{if $0 \le x \le 1$}, \\
                                    0, & \mbox{if  $x  >  1$}.
                  \end{array}
          \right.
</math>
Note that <math>F_X(x)</math> is continuous, but <math>f_X(x)</math> is not.  (See [[#fig 5.5|Figure]].)
<div id="fig 5.5" class="d-flex justify-content-center">
[[File:guide_e6d15_PSfig5-5.png | 400px | thumb | Distribution and density for <math>X = U^2</math>. ]]
</div>
When referring to a continuous random variable <math>X</math> (say with a uniform
density function), it is  customary to say that “<math>X</math> is uniformly
''distributed''
on the interval <math>[a, b]</math>.”  It is also customary to refer to the cumulative
distribution function of <math>X</math> as the distribution function of
<math>X</math>.  Thus, the word “distribution” is being used in several different ways in the
subject of probability.  (Recall that it also has a meaning when discussing discrete
random variables.)  When referring to the cumulative distribution function of a
continuous random variable <math>X</math>, we will always use the word “cumulative” as a
modifier, unless the use of another modifier, such as “normal” or “exponential,”
makes it clear.  Since the phrase “uniformly densitied on the interval
<math>[a, b]</math>” is not acceptable English, we will have to say “uniformly distributed”
instead.
<span id="exam 2.2.7.2"/>
'''Example'''
In [[guide:A070937c41#exam 2.1.4.5 |Example]], we considered a random variable, defined to be the
sum of two random real numbers chosen
uniformly from <math>[0, 1]</math>.  Let the random variables <math>X</math> and <math>Y</math> denote the two chosen real
numbers.  Define <math>Z = X + Y</math>.  We will now derive expressions for the cumulative distribution
function and the density function of
<math>Z</math>.
<div id="fig 2.15.5" class="d-flex justify-content-center">
[[File:guide_e6d15_PSfig2-15-5.png | 400px | thumb | Calculation of distribution function for [[#exam2.2.7.2|Example]]]]
</div>
<div id="fig 5.6" class="d-flex justify-content-center">
[[File:guide_e6d15_PSfig5-6.png | 400px | thumb | Distribution and density functions for [[#exam2.2.7.2|Example]]. ]]
</div>
Here we take for our sample space <math>\Omega</math> the unit square in <math>\mat{R}^2</math>
with uniform density.  A point <math>\omega \in \Omega</math> then consists of a pair <math>(x, y)</math>
of numbers chosen at random.  Then <math>0 \leq Z\leq 2</math>.  Let <math>E_z</math> denote the event
that <math>Z \le z</math>.  In [[#fig 2.15.5|Figure]], we show the set <math>E_{.8}</math>.  The event <math>E_z</math>,
for any <math>z</math> between 0 and 1, looks very similar to the shaded set in the figure.  For <math>1  <  z
\le 2</math>, the set <math>E_z</math> looks like the unit square with a triangle removed from the upper
right-hand corner.  We can now calculate the probability distribution <math>F_Z</math> of <math>Z</math>; it is
given by
<math display="block">
\begin{eqnarray*}
F_Z(z) & = & P(Z \le z) \\
      & = & \mbox {Area\ of\ }E_z \\
      & = & \left \{\begin{array}{ll}
                              0, & \mbox{if  $z  <  0$}, \\
                      (1/2)z^2, & \mbox{if  $0 \le z \le 1$}, \\
              1 - (1/2)(2-z)^2, & \mbox{if  $1 \le z \le 2$}, \\
                              1, & \mbox{if  $2  <  z$}.
                    \end{array}
            \right.
\end{eqnarray*}
</math>
The density function is obtained by differentiating this function:
<math display="block">
f_Z(z)  = \left \{\begin{array}{ll}
                      0, & \mbox{if  $z  <  0$}, \\
                      z, & \mbox{if  $0 \le z \le 1$}, \\
                  2 - z, & \mbox{if  $1 \le z \le 2$}, \\
                      0, & \mbox{if  $2  <  z$}.
                \end{array}
        \right.
</math>
The reader is referred to [[#fig 5.6|Figure]] for the graphs of these functions.
<span id="exam 2.2.7.3"/>
'''Example'''
In the dart game described in [[#exam 2.2.2 |Example]], what is the distribution of the distance of the dart from the center of the target?  What is its density?
<div id="fig 5.8" class="d-flex justify-content-center">
[[File:guide_e6d15_PSfig5-8.png | 400px | thumb |Calculation of <math>F_{z}</math> for  [[#exam 2.2.7.3 |Example]] ]]
</div>
Here, as before, our sample space <math>\Omega</math> is the unit disk in <math>\mat{R}^2</math>,
with coordinates <math>(X, Y)</math>.  Let <math>Z = \sqrt{X^2 + Y^2}</math> represent the
distance from the center of the target.  Let <math>E</math> be the event <math>\{Z \le z\}</math>.  Then the
distribution function <math>F_Z</math> of <math>Z</math> (see [[#fig 5.8|Figure]]) is given by
<math display="block">
\begin{eqnarray*}
F_Z(z) & = & P(Z \le z) \\
      & = & {{\mbox {Area\ of\ }E}\over \mbox {Area\ of\ target}}\ .
\end{eqnarray*}
</math>
Thus, we easily compute that
<math display="block">
F_Z(z) = \left \{ \begin{array}{ll}
                          0, & \mbox{if  $z \le 0$}, \\
                        z^2, & \mbox{if  $0 \le z \le 1$}, \\
                          1, & \mbox{if  $z  >  1$}.
                  \end{array}
          \right.
</math>
The density <math>f_Z(z)</math> is given again by the derivative of <math>F_Z(z)</math>:
<math display="block">
f_Z(z) = \left \{ \begin{array}{ll}
                        0, & \mbox{if  $z \le 0$}, \\
                        2z, & \mbox{if  $0 \le z \le 1$}, \\
                        0, & \mbox{if  $z  >  1$}.
                  \end{array}
          \right.
</math>
The reader is referred to [[#fig 5.9|Figure]] for the graphs of these functions.
We can verify this result by simulation, as follows: We choose values for <math>X</math>
and <math>Y</math> at random from <math>[0,1]</math> with uniform distribution, calculate <math>Z =
\sqrt{X^2 + Y^2}</math>, check whether <math>0 \leq Z \leq 1</math>, and present the results in a
bar graph (see [[#fig 5.10|Figure]]).
<div id="fig 5.9" class="d-flex justify-content-center">
[[File:guide_e6d15_PSfig5-9.png | 400px | thumb | Distribution and density for <math>Z =\sqrt{X^2 + Y^2}</math>. ]]
</div>
<div id="fig 5.10" class="d-flex justify-content-center">
[[File:guide_e6d15_PSfig5-10.png | 400px | thumb | Simulation results for [[#exam 2.2.7.3|Example]].]]
</div>
<span id="exam 2.2.7.4"/>
'''Example'''
Suppose Mr.\ and Mrs.\ Lockhorn agree to meet at the Hanover
Inn between 5:00 and 6:00 {\scriptsize P.M.} on Tuesday.  Suppose each
arrives at a time between 5:00 and 6:00 chosen at random with uniform probability.  What is the
distribution function for the length of time that the first to arrive has to
wait for the other?  What is the density function?
Here again we can take the unit square to represent the sample space, and <math>(X, Y)</math>
as the arrival times (after 5:00 {\scriptsize P.M.}) for the Lockhorns.  Let
<math>Z = |X - Y|</math>.  Then we have
<math>F_X(x) = x</math> and <math>F_Y(y) = y</math>.  Moreover (see [[#fig 5.11|Figure]]),
<math display="block">
\begin{eqnarray*}
F_Z(z) & = & P(Z \leq z) \\
      & = & P(|X - Y| \leq z) \\
      & = & \mbox {Area\ of\ }E\ .
\end{eqnarray*}
</math>
Thus, we have
<math display="block">
F_Z(z) = \left \{ \begin{array}{ll}
                        0, & \mbox{if  $z \le 0$}, \\
            1 - (1 - z)^2, & \mbox{if  $0 \le z \le 1$}, \\
                        1, & \mbox{if  $z  >  1$}.
                  \end{array}
          \right.
</math>
<div id="fig 5.11" class="d-flex justify-content-center">
[[File:guide_e6d15_PSfig5-11.png | 400px | thumb |  Calculation of <math>F_{Z}</math>.]]
</div> The density <math>f_Z(z)</math> is again obtained by differentiation:
<math display="block">
f_Z(z) = \left \{ \begin{array}{ll}
                        0, & \mbox{if  $z \le 0$}, \\
                  2(1-z), & \mbox{if  $0 \le z \le 1$}, \\
                        0, & \mbox{if  $z  >  1$}.
                  \end{array}
          \right.
</math>
<span id="exam 2.2.7.5"/>
'''Example'''
There are many occasions where we observe a sequence of occurrences which occur at
“random”  times.  For example, we might be observing emissions of a radioactive
isotope, or cars  passing a milepost on a highway, or light bulbs burning out.  In such cases, we might define a
random variable <math>X</math> to denote the time between successive occurrences.  Clearly,
<math>X</math> is a continuous random variable whose range consists of the non-negative real
numbers.  It is often the case that we can  model <math>X</math> by using the ''exponential
density''.  This density is given
by the formula
<math display="block">
f(t) = \left \{ \begin{array}{ll}
                        \lambda e^{-\lambda t}, & \mbox{if  $t \ge 0$}, \\
                                            0, & \mbox{if  $t  <  0$}.
                  \end{array}
        \right.
</math>
The number <math>\lambda</math> is a non-negative real number, and represents the reciprocal of the
average value of <math>X</math>.  (This will be shown in Chapter [[guide:E4fd10ce73|Expected Value and Variance]].)  Thus, if the average
time between occurrences is 30 minutes, then <math>\lambda = 1/30</math>.  A graph of this density
function with <math>\lambda = 1/30</math> is shown in [[#fig 2.16|Figure]].
<div id="fig 2.16.5" class="d-flex justify-content-center">
[[File:guide_e6d15_PSfig2-16.png | 400px | thumb | Spinner experiment. ]]
</div>
One can see from the figure that even though the average value is 30, occasionally much
larger values are taken on by <math>X</math>.
Suppose that we have bought a computer that contains a Warp 9 hard drive.  The salesperson says that the average time between breakdowns of this type of hard
drive is 30 months.  It is often  assumed that the length of time between breakdowns is
distributed according to the exponential density.  We will assume that this model applies
here, with
<math>\lambda = 1/30</math>.
Now suppose that we have been operating our computer for 15 months.  We assume that the
original hard drive is still running.  We ask how long we should expect the hard drive to
continue to run.  One  could reasonably expect that the hard drive will run, on the
average, another 15 months.  (One might also guess that it will run more than 15 months,
since the fact that it has already run for 15 months implies that we don't have a lemon.)
The time which we have to wait is a new random variable, which we will call
<math>Y</math>.  Obviously, <math>Y = X - 15</math>.  We can write a computer program to produce a sequence of
simulated <math>Y</math>-values.  To do this, we first produce a sequence of <math>X</math>'s, and discard those
values which are  less than or equal to 15 (these values correspond to the cases where the
hard drive has quit running before 15 months).  To simulate a value of
<math>X</math>, we compute the value of the expression
<math display="block">
\Bigl(-{1\over{\lambda}}\Bigr)\log(rnd)\ ,
</math>
where <math>rnd</math> represents a random real number between 0 and 1.  (That this expression has
the exponential density will be shown in Chapter [[guide:A618cf4c07|Distributions and Densities]].)  [[#fig 2.17|Figure]]
shows an area bar graph of 10,00 simulated <math>Y</math>-values.
The average value of <math>Y</math> in this simulation is 29.74, which is closer to the original
average life span of 30 months than to the value of 15 months which was guessed above. 
Also, the distribution of <math>Y</math> is seen to be close to the distribution of <math>X</math>.  It is in
fact the case that
<math>X</math> and <math>Y</math> have the same distribution.  This property is called the ''memoryless
property'', because the amount of time that we have to wait for an
occurrence does not depend on how long we have already waited.  The only continuous density
function with this property is the exponential density.
<div id="fig 2.17" class="d-flex justify-content-center">
[[File:guide_e6d15_PSfig2-17.png | 400px | thumb | Residual lifespan of a hard drive. ]]
</div>
===Assignment of Probabilities===
A fundamental question in practice is: How shall we choose the probability
density function in describing any given experiment?  The answer depends to a
great extent on the amount and kind of information available to us about the
experiment.  In some cases, we can see that the outcomes are equally likely.
In some cases, we can see that the experiment resembles another already
described by a known density.  In some cases, we can run the experiment a large
number of times and make a reasonable guess at the density on the basis of the
observed distribution of outcomes, as we did in Chapter [[guide:4f3a4e96c3|Discrete Probability Distributions]]. 
In general, the problem of choosing the right density function for a given
experiment is a central problem for the experimenter and is not always easy to
solve (see [[guide:A070937c41#exam 2.1.5 |Example]]). 
We shall not examine this question in detail here but instead shall assume that the
right density is already known for each of the experiments under study.
The introduction of suitable coordinates to describe a continuous sample space,
and a suitable density to describe its probabilities, is not always
so obvious, as our final example shows.
===Infinite Tree===
<span id="exam 2.2.12"/>
'''Example'''
Consider an experiment in which a fair coin is tossed repeatedly, without
stopping.  We have seen in [[guide:C9e774ade5#exam 1.5 |Example]]
that, for a coin tossed <math>n</math> times, the natural sample space is a binary tree
with <math>n</math> stages.  On this evidence we expect that for a coin tossed repeatedly,
the natural sample space is a binary tree with an infinite
number of stages, as indicated in [[#fig 2.23|Figure]].
It is surprising to learn that, although the <math>n</math>-stage tree is obviously a
finite sample space, the unlimited tree can be described as a continuous
sample space.  To see how this comes about, let us agree that a typical outcome
of the unlimited coin tossing experiment can be described by a sequence of the
form <math>\omega = \{\mbox{H H T H T T H}\dots\}</math>.  If we write 1 for H and 0 for
T, then <math>\omega = \{1\ 1\ 0\ 1\ 0\ 0\ 1\dots\}</math>.  In this way, each outcome is
described by a sequence of 0's and 1's.
<div id="fig 2.23" class="d-flex justify-content-center">
[[File:guide_e6d15_PSfig2-23.png | 400px | thumb | Tree for infinite number of tosses of a coin. ]]
</div>
Now suppose we think of this sequence of 0's and 1's as the binary expansion of
some real number <math>x = .1101001\cdots</math> lying between 0 and 1.  (A ''binary expansion'' is like a decimal expansion but based on 2 instead of 10.)
Then each outcome is described by a value of <math>x</math>, and in this way <math>x</math> becomes a
coordinate for the sample space, taking on all real values between 0 and 1.  (We note that
it is possible for two different sequences to correspond to the same real number; for example,
the sequences <math>\{\mbox{T H H H H H}\ldots\}</math> and <math>\{\mbox{H T T T T T}\ldots\}</math> both
correspond to the real number <math>1/2</math>.  We will not concern ourselves with this apparent problem
here.)
What probabilities should be assigned to the events of this sample space?
Consider, for example, the event <math>E</math> consisting of all outcomes for which the
first toss comes up heads and the second tails.  Every such outcome has the
form <math>.10****\cdots</math>, where <math>*</math> can be either 0 or 1.  Now if <math>x</math> is our
real-valued coordinate, then the value of <math>x</math> for every such outcome must lie
between <math>1/2 = .10000\cdots</math> and <math>3/4 = .11000\cdots</math>, and moreover, every
value of <math>x</math> between 1/2 and 3/4 has a binary expansion of the form
<math>.10****\cdots</math>.  This means that <math>\omega\in E</math> if and only if <math>1/2 \leq x  <
3/4</math>, and in this way we see that we can describe <math>E</math> by the interval
<math>[1/2,3/4)</math>.  More generally, every event consisting of outcomes for which the
results of the first <math>n</math> tosses are prescribed is described by a binary
interval of the form <math>[k/2^n,(k+1)/2^n)</math>.
We have already seen in [[guide:C9e774ade5|Section]] that in the experiment involving <math>n</math> tosses, the probability of any one outcome must be
exactly <math>1/2^n</math>.  It follows that in the unlimited toss experiment, the
probability of any event consisting of outcomes for which the results of the
first <math>n</math> tosses are prescribed must also be <math>1/2^n</math>.  But <math>1/2^n</math> is exactly
the length of the interval of <math>x</math>-values describing <math>E</math>!  Thus we see that,
just as with the spinner experiment, the probability of an event <math>E</math> is
determined by what fraction of the unit interval lies in <math>E</math>.
Consider again the statement: The probability is 1/2 that a fair coin will turn
up heads when tossed.  We have suggested that one interpretation of this
statement is that if we toss the coin indefinitely the proportion of heads will
approach 1/2.  That is, in our correspondence with binary sequences we expect
to get a binary sequence with the proportion of 1's tending to 1/2.  The event
<math>E</math> of binary sequences for which this is true is a proper subset of the set of all
possible binary sequences.  It does not contain, for example, the sequence
<math>011011011\ldots</math> (i.e., (011) repeated again and again).  The event <math>E</math> is
actually a very complicated subset of the binary sequences, but its probability
can be determined as a limit of probabilities for events with a finite number
of outcomes whose probabilities are given by finite tree measures.  When the
probability of <math>E</math> is computed in this way, its value is found to be 1.  This
remarkable result is known as the ''Strong Law of Large Numbers'' (or
''Law of Averages'') and is one justification for our frequency
concept of probability.  We shall prove a weak form of
this theorem in Chapter [[guide:1cf65e65b3|Law of Large Numbers]].
==General references==
{{cite web |url=https://math.dartmouth.edu/~prob/prob/prob.pdf |title=Grinstead and Snell’s Introduction to Probability |last=Doyle |first=Peter G.|date=2006 |access-date=June 6, 2024}}

Latest revision as of 01:57, 10 June 2024

[math] \newcommand{\NA}{{\rm NA}} \newcommand{\mat}[1]{{\bf#1}} \newcommand{\exref}[1]{\ref{##1}} \newcommand{\secstoprocess}{\all} \newcommand{\NA}{{\rm NA}} \newcommand{\mathds}{\mathbb}[/math]

In the previous section we have seen how to simulate experiments with a whole continuum of possible outcomes and have gained some experience in thinking about such experiments. Now we turn to the general problem of assigning probabilities to the outcomes and events in such experiments. We shall restrict our attention here to those experiments whose sample space can be taken as a suitably chosen subset of the line, the plane, or some other Euclidean space. We begin with some simple examples.

Spinners

Example The spinner experiment described in Example has the interval [math][0, 1)[/math] as the set of possible outcomes. We would like to construct a probability model in which each outcome is equally likely to occur. We saw that in such a model, it is necessary to assign the probability 0 to each outcome. This does not at all mean that the probability of every event must be zero. On the contrary, if we let the random variable [math]X[/math] denote the outcome, then the probability

[[math]] P(\,0 \leq X \leq 1) [[/math]]

that the head of the spinner comes to rest somewhere in the circle, should be equal to 1. Also, the probability that it comes to rest in the upper half of the circle should be the same as for the lower half, so that

[[math]] P\biggl(0 \leq X \lt \frac12\biggr) = P\biggl(\frac12 \leq X \lt 1\biggr) = \frac12\ . [[/math]]

More generally, in our model, we would like the equation

[[math]] P(c \leq X \lt d) = d - c [[/math]]

to be true for every choice of [math]c[/math] and [math]d[/math].


If we let [math]E = [c, d][/math], then we can write the above formula in the form

[[math]] P(E) = \int_E f(x)\,dx\ , [[/math]]

where [math]f(x)[/math] is the constant function with value 1. This should remind the reader of the corresponding formula in the discrete case for the probability of an event:

[[math]] P(E) = \sum_{\omega \in E} m(\omega)\ . [[/math]]

The difference is that in the continuous case, the quantity being integrated, [math]f(x)[/math], is not the probability of the outcome [math]x[/math]. (However, if one uses infinitesimals, one can consider [math]f(x)\,dx[/math] as the probability of the outcome [math]x[/math].)


In the continuous case, we will use the following convention. If the set of outcomes is a set of real numbers, then the individual outcomes will be referred to by small Roman letters such as [math]x[/math]. If the set of outcomes is a subset of [math]R^2[/math], then the individual outcomes will be denoted by [math](x, y)[/math]. In either case, it may be more convenient to refer to an individual outcome by using [math]\omega[/math], as in Chapter Discrete Probability Distributions.

Spinner experiment.


Figure shows the results of 1000 spins of the spinner. The function [math]f(x)[/math] is also shown in the figure. The reader will note that the area under [math]f(x)[/math] and above a given interval is approximately equal to the fraction of outcomes that fell in that interval. The function [math]f(x)[/math] is called the density function of the random variable [math]X[/math]. The fact that the area under [math]f(x)[/math] and above an interval corresponds to a probability is the defining property of density functions. A precise definition of density functions will be given shortly.


Darts

Example A game of darts involves throwing a dart at a circular target of unit radius. Suppose we throw a dart once so that it hits the target, and we observe where it lands. To describe the possible outcomes of this experiment, it is natural to take as our sample space the set [math]\Omega[/math] of all the points in the target. It is convenient to describe these points by their rectangular coordinates, relative to a coordinate system with origin at the center of the target, so that each pair [math](x,y)[/math] of coordinates with [math]x^2 + y^2 \leq 1[/math] describes a possible outcome of the experiment. Then [math]\Omega = \{\,(x,y) : x^2 + y^2 \leq 1\,\}[/math] is a subset of the Euclidean plane, and the event [math]E = \{\,(x,y) : y \gt 0\,\}[/math], for example, corresponds to the statement that the dart lands in the upper half of the target, and so forth. Unless there is reason to believe otherwise (and with experts at the game there may well be!), it is natural to assume that the coordinates are chosen at random. (When doing this with a computer, each coordinate is chosen uniformly from the interval [math][-1, 1][/math]. If the resulting point does not lie inside the unit circle, the point is not counted.) Then the arguments used in the preceding example show that the probability of any elementary event, consisting of a single outcome, must be zero, and suggest that the probability of the event that the dart lands in any subset [math]E[/math] of the target should be determined by what fraction of the target area lies in [math]E[/math]. Thus,

[[math]] P(E) = \frac{\mbox{area\ of}\ E}{\mbox{area\ of\ target}} = \frac{\mbox{area\ of}\ E}\pi\ . [[/math]]

This can be written in the form

[[math]] P(E) = \int_E f(x)\,dx\ , [[/math]]

where [math]f(x)[/math] is the constant function with value [math]1/\pi[/math]. In particular, if [math]E = \{\,(x,y) : x^2 + y^2 \leq a^2\,\}[/math] is the event that the dart lands within distance [math]a \lt 1[/math] of the center of the target, then

[[math]] P(E) = \frac{\pi a^2}\pi = a^2\ . [[/math]]

For example, the probability that the dart lies within a distance 1/2 of the center is 1/4.

Example In the dart game considered above, suppose that, instead of observing where the dart lands, we observe how far it lands from the center of the target.


In this case, we take as our sample space the set [math]\Omega[/math] of all circles with centers at the center of the target. It is convenient to describe these circles by their radii, so that each circle is identified by its radius [math]r[/math], [math]0 \leq r \leq 1[/math]. In this way, we may regard [math]\Omega[/math] as the subset [math][0,1][/math] of the real line.


What probabilities should we assign to the events [math]E[/math] of [math]\Omega[/math]? If

[[math]] E = \{\,r : 0 \leq r \leq a\,\}\ , [[/math]]

then [math]E[/math] occurs if the dart lands within a distance [math]a[/math] of the center, that is, within the circle of radius [math]a[/math], and we saw in the previous example that under our assumptions the probability of this event is given by

[[math]] P([0,a]) = a^2\ . [[/math]]

More generally, if

[[math]] E = \{\,r : a \leq r \leq b\,\}\ , [[/math]]

then by our basic assumptions,

[[math]] \begin{eqnarray*} P(E) = P([a,b]) & = & P([0,b]) - P([0,a]) \\ & = & b^2 - a^2 \\ & = & (b - a)(b + a) \\ & = & 2(b - a)\frac{(b + a)}2\ . \end{eqnarray*} [[/math]]


Thus, [math]P(E) = [/math]2(length of [math]E[/math])(midpoint of [math]E[/math]). Here we see that the probability assigned to the interval [math]E[/math] depends not only on its length but also on its midpoint (i.e., not only on how long it is, but also on where it is). Roughly speaking, in this experiment, events of the form [math]E = [a,b][/math] are more likely if they are near the rim of the target and less likely if they are near the center. (A common experience for beginners! The conclusion might well be different if the beginner is replaced by an expert.) Again we can simulate this by computer. We divide the target area into ten concentric regions of equal thickness.

Distribution of dart distances in 400 throws.


The computer program Darts throws [math]n[/math] darts and records what fraction of the total falls in each of these concentric regions. The program Areabargraph then plots a bar graph with the area of the [math]i[/math]th bar equal to the fraction of the total falling in the [math]i[/math]th region. Running the program for 1000 darts resulted in the bar graph of Figure.


Note that here the heights of the bars are not all equal, but grow approximately linearly with [math]r[/math]. In fact, the linear function [math]y = 2r[/math] appears to fit our bar graph quite well. This suggests that the probability that the dart falls within a distance [math]a[/math] of the center should be given by the area under the graph of the function [math]y = 2r[/math] between 0 and [math]a[/math]. This area is [math]a^2[/math], which agrees with the probability we have assigned above to this event.


Sample Space Coordinates

These examples suggest that for continuous experiments of this sort we should assign probabilities for the outcomes to fall in a given interval by means of the area under a suitable function.


More generally, we suppose that suitable coordinates can be introduced into the sample space [math]\Omega[/math], so that we can regard [math]\Omega[/math] as a subset of [math]''' R'''^n[/math]. We call such a sample space a continuous sample space. We let [math]X[/math] be a random variable which represents the outcome of the experiment. Such a random variable is called a continuous random variable. We then define a density function for [math]X[/math] as follows.

Density Functions of Continuous Random Variables

Definition

Let [math]X[/math] be a continuous real-valued random variable. A density function for [math]X[/math] is a real-valued function [math]f[/math] which satisfies

[[math]] P(a \le X \le b) = \int_a^b f(x)\,dx [[/math]]
for all [math]a,\ b \in ''' R'''[/math].

We note that it is not the case that all continuous real-valued random variables possess density functions. However, in this book, we will only consider continuous random variables for which density functions exist.


In terms of the density [math]f(x)[/math], if [math]E[/math] is a subset of [math]{\mat R}[/math], then

[[math]] P(X \in E) = \int_Ef(x)\,dx\ . [[/math]]

The notation here assumes that [math]E[/math] is a subset of [math]{\mat R}[/math] for which [math]\int_E f(x)\,dx[/math] makes sense.

Example In the spinner experiment, we choose for our set of outcomes the interval [math]0 \leq x \lt 1[/math], and for our density function

[[math]] f(x) = \left \{ \begin{array}{ll} 1, & \mbox{if $0 \leq x \lt 1$,} \\ 0, & \mbox{otherwise.} \end{array} \right. [[/math]]
If [math]E[/math] is the event that the head of the spinner falls in the upper half of the circle, then [math]E = \{\,x : 0 \leq x \leq 1/2\,\}[/math], and so

[[math]] P(E) = \int_0^{1/2} 1\,dx = \frac12\ . [[/math]]
More generally, if [math]E[/math] is the event that the head falls in the interval [math][a,b][/math], then

[[math]] P(E) = \int_a^b 1\,dx = b - a\ . [[/math]]

Example In the first dart game experiment, we choose for our sample space a disc of unit radius in the plane and for our density function the function

[[math]] f(x,y) = \left \{ \begin{array}{ll} 1/\pi, & \mbox{if $x^2 + y^2 \leq 1$,} \\ 0, & \mbox{otherwise.} \end{array} \right. [[/math]]
The probability that the dart lands inside the subset [math]E[/math] is then given by

[[math]] \begin{eqnarray*} P(E) & = & \int\,\int_E \frac1\pi \,dx\,dy\\ & = & \frac1\pi \cdot (\mbox{area\,\,\, of}\,\,\,E)\ . \end{eqnarray*} [[/math]]


In these two examples, the density function is constant and does not depend on the particular outcome. It is often the case that experiments in which the coordinates are chosen at random can be described by constant density functions, and, as in Section \ref{sec 1.2}, we call such density functions uniform or equiprobable. Not all experiments are of this type, however.

Example In the second dart game experiment, we choose for our sample space the unit interval on the real line and for our density the function

[[math]] f(r) = \left \{ \begin{array}{ll} 2r, & \mbox{if $0 \lt r \lt 1$,} \\ 0, & \mbox{otherwise.} \end{array} \right. [[/math]]
Then the probability that the dart lands at distance [math]r[/math], [math]a \leq r \leq b[/math], from the center of the target is given by

[[math]] \begin{eqnarray*} P([a,b]) & = & \int_a^b 2r\,dr\\ & = & b^2 - a^2\ . \end{eqnarray*} [[/math]]
Here again, since the density is small when [math]r[/math] is near 0 and large when [math]r[/math] is near 1, we see that in this experiment the dart is more likely to land near the rim of the target than near the center. In terms of the bar graph of Example, the heights of the bars approximate the density function, while the areas of the bars approximate the probabilities of the subintervals (see Figure).


We see in this example that, unlike the case of discrete sample spaces, the value [math]f(x)[/math] of the density function for the outcome [math]x[/math] is not the probability of [math]x[/math] occurring (we have seen that this probability is always 0) and in general [math]f(x)[/math] is not a probability at all. In this example, if we take [math]\lambda = 2[/math] then [math]f(3/4) = 3/2[/math], which being bigger than 1, cannot be a probability.


Nevertheless, the density function [math]f[/math] does contain all the probability information about the experiment, since the probabilities of all events can be derived from it. In particular, the probability that the outcome of the experiment falls in an interval [math][a,b][/math] is given by

[[math]] P([a,b]) = \int_a^b f(x)\,dx\ , [[/math]]
that is, by the area under the graph of the density function in the interval [math][a,b][/math]. Thus, there is a close connection here between probabilities and areas. We have been guided by this close connection in making up our bar graphs; each bar is chosen so that its area, and not its height, represents the relative frequency of occurrence, and hence estimates the probability of the outcome falling in the associated interval.


In the language of the calculus, we can say that the probability of occurrence of an event of the form [math][x, x + dx][/math], where [math]dx[/math] is small, is approximately given by

[[math]] P([x, x+dx]) \approx f(x)dx\ , [[/math]]
that is, by the area of the rectangle under the graph of [math]f[/math]. Note that as [math]dx \to 0[/math], this probability [math]\to 0[/math], so that the probability [math]P(\{x\})[/math] of a single point is again 0, as in Example.


A glance at the graph of a density function tells us immediately which events of an experiment are more likely. Roughly speaking, we can say that where the density is large the events are more likely, and where it is small the events are less likely. In Example the density function is largest at 1. Thus, given the two intervals [math][0, a][/math] and [math][1, 1+a][/math], where [math]a[/math] is a small positive real number, we see that [math]X[/math] is more likely to take on a value in the second interval than in the first.

Cumulative Distribution Functions of Continuous Random Variables

We have seen that density functions are useful when considering continuous random variables. There is another kind of function, closely related to these density functions, which is also of great importance. These functions are called cumulative distribution functions.

Definition

Let [math]X[/math] be a continuous real-valued random variable. Then the cumulative distribution function of [math]X[/math] is defined by the equation

[[math]] F_X(x) = P(X \le x)\ . [[/math]]

If [math]X[/math] is a continuous real-valued random variable which possesses a density function, then it also has a cumulative distribution function, and the following theorem shows that the two functions are related in a very nice way.

Theorem

Let [math]X[/math] be a continuous real-valued random variable with density function [math]f(x)[/math]. Then the function defined by

[[math]] F(x) = \int_{-\infty}^x f(t)\,dt [[/math]]
is the cumulative distribution function of [math]X[/math]. Furthermore, we have

[[math]] {{d\ }\over{dx}} F(x) = f(x)\ . [[/math]]

Show Proof

By definition,

[[math]] F(x) = P(X \le x)\ . [[/math]]
Let [math]E = (-\infty, x][/math]. Then

[[math]] P(X \le x) = P(X \in E)\ , [[/math]]
which equals

[[math]] \int_{-\infty}^x f(t)\,dt\ . [[/math]]

Applying the Fundamental Theorem of Calculus to the first equation in the statement of the theorem yields the second statement.


In many experiments, the density function of the relevant random variable is easy to write down. However, it is quite often the case that the cumulative distribution function is easier to obtain than the density function. (Of course, once we have the cumulative distribution function, the density function can easily be obtained by differentiation, as the above theorem shows.) We now give some examples which exhibit this phenomenon.

Example A real number is chosen at random from [math][0, 1][/math] with uniform probability, and then this number is squared. Let [math]X[/math] represent the result. What is the cumulative distribution function of [math]X[/math]? What is the density of [math]X[/math]?


We begin by letting [math]U[/math] represent the chosen real number. Then [math]X = U^2[/math]. If [math]0 \le x \le 1[/math], then we have

[[math]] \begin{eqnarray*} F_X(x) & = & P(X \le x) \\ & = & P(U^2 \le x) \\ & = & P(U \le \sqrt x) \\ & = & \sqrt x\ . \end{eqnarray*} [[/math]]
It is clear that [math]X[/math] always takes on a value between 0 and 1, so the cumulative distribution function of [math]X[/math] is given by

[[math]] F_X(x) = \left \{ \begin{array}{ll} 0, & \mbox{if $x \le 0$}, \\ {\sqrt x}, & \mbox{if $0 \le x \le 1$}, \\ 1, & \mbox{if $x \ge 1$}. \end{array} \right. [[/math]]
From this we easily calculate that the density function of [math]X[/math] is

[[math]] f_X(x) = \left \{ \begin{array}{ll} 0, & \mbox{if $x \le 0$}, \\ 1/(2{\sqrt x}), & \mbox{if $0 \le x \le 1$}, \\ 0, & \mbox{if $x \gt 1$}. \end{array} \right. [[/math]]
Note that [math]F_X(x)[/math] is continuous, but [math]f_X(x)[/math] is not. (See Figure.)

Distribution and density for [math]X = U^2[/math].

When referring to a continuous random variable [math]X[/math] (say with a uniform density function), it is customary to say that “[math]X[/math] is uniformly distributed on the interval [math][a, b][/math].” It is also customary to refer to the cumulative distribution function of [math]X[/math] as the distribution function of [math]X[/math]. Thus, the word “distribution” is being used in several different ways in the subject of probability. (Recall that it also has a meaning when discussing discrete random variables.) When referring to the cumulative distribution function of a continuous random variable [math]X[/math], we will always use the word “cumulative” as a modifier, unless the use of another modifier, such as “normal” or “exponential,” makes it clear. Since the phrase “uniformly densitied on the interval [math][a, b][/math]” is not acceptable English, we will have to say “uniformly distributed” instead.

Example In Example, we considered a random variable, defined to be the sum of two random real numbers chosen uniformly from [math][0, 1][/math]. Let the random variables [math]X[/math] and [math]Y[/math] denote the two chosen real numbers. Define [math]Z = X + Y[/math]. We will now derive expressions for the cumulative distribution function and the density function of

[math]Z[/math].

Calculation of distribution function for Example
Distribution and density functions for Example.


Here we take for our sample space [math]\Omega[/math] the unit square in [math]\mat{R}^2[/math] with uniform density. A point [math]\omega \in \Omega[/math] then consists of a pair [math](x, y)[/math] of numbers chosen at random. Then [math]0 \leq Z\leq 2[/math]. Let [math]E_z[/math] denote the event that [math]Z \le z[/math]. In Figure, we show the set [math]E_{.8}[/math]. The event [math]E_z[/math], for any [math]z[/math] between 0 and 1, looks very similar to the shaded set in the figure. For [math]1 \lt z \le 2[/math], the set [math]E_z[/math] looks like the unit square with a triangle removed from the upper right-hand corner. We can now calculate the probability distribution [math]F_Z[/math] of [math]Z[/math]; it is given by

[[math]] \begin{eqnarray*} F_Z(z) & = & P(Z \le z) \\ & = & \mbox {Area\ of\ }E_z \\ & = & \left \{\begin{array}{ll} 0, & \mbox{if $z \lt 0$}, \\ (1/2)z^2, & \mbox{if $0 \le z \le 1$}, \\ 1 - (1/2)(2-z)^2, & \mbox{if $1 \le z \le 2$}, \\ 1, & \mbox{if $2 \lt z$}. \end{array} \right. \end{eqnarray*} [[/math]]
The density function is obtained by differentiating this function:

[[math]] f_Z(z) = \left \{\begin{array}{ll} 0, & \mbox{if $z \lt 0$}, \\ z, & \mbox{if $0 \le z \le 1$}, \\ 2 - z, & \mbox{if $1 \le z \le 2$}, \\ 0, & \mbox{if $2 \lt z$}. \end{array} \right. [[/math]]
The reader is referred to Figure for the graphs of these functions.

Example In the dart game described in Example, what is the distribution of the distance of the dart from the center of the target? What is its density?

Calculation of [math]F_{z}[/math] for Example

Here, as before, our sample space [math]\Omega[/math] is the unit disk in [math]\mat{R}^2[/math], with coordinates [math](X, Y)[/math]. Let [math]Z = \sqrt{X^2 + Y^2}[/math] represent the distance from the center of the target. Let [math]E[/math] be the event [math]\{Z \le z\}[/math]. Then the distribution function [math]F_Z[/math] of [math]Z[/math] (see Figure) is given by

[[math]] \begin{eqnarray*} F_Z(z) & = & P(Z \le z) \\ & = & {{\mbox {Area\ of\ }E}\over \mbox {Area\ of\ target}}\ . \end{eqnarray*} [[/math]]
Thus, we easily compute that

[[math]] F_Z(z) = \left \{ \begin{array}{ll} 0, & \mbox{if $z \le 0$}, \\ z^2, & \mbox{if $0 \le z \le 1$}, \\ 1, & \mbox{if $z \gt 1$}. \end{array} \right. [[/math]]
The density [math]f_Z(z)[/math] is given again by the derivative of [math]F_Z(z)[/math]:

[[math]] f_Z(z) = \left \{ \begin{array}{ll} 0, & \mbox{if $z \le 0$}, \\ 2z, & \mbox{if $0 \le z \le 1$}, \\ 0, & \mbox{if $z \gt 1$}. \end{array} \right. [[/math]]
The reader is referred to Figure for the graphs of these functions.


We can verify this result by simulation, as follows: We choose values for [math]X[/math] and [math]Y[/math] at random from [math][0,1][/math] with uniform distribution, calculate [math]Z = \sqrt{X^2 + Y^2}[/math], check whether [math]0 \leq Z \leq 1[/math], and present the results in a bar graph (see Figure).

Distribution and density for [math]Z =\sqrt{X^2 + Y^2}[/math].
Simulation results for Example.

Example Suppose Mr.\ and Mrs.\ Lockhorn agree to meet at the Hanover Inn between 5:00 and 6:00 {\scriptsize P.M.} on Tuesday. Suppose each arrives at a time between 5:00 and 6:00 chosen at random with uniform probability. What is the distribution function for the length of time that the first to arrive has to wait for the other? What is the density function?


Here again we can take the unit square to represent the sample space, and [math](X, Y)[/math] as the arrival times (after 5:00 {\scriptsize P.M.}) for the Lockhorns. Let [math]Z = |X - Y|[/math]. Then we have [math]F_X(x) = x[/math] and [math]F_Y(y) = y[/math]. Moreover (see Figure),

[[math]] \begin{eqnarray*} F_Z(z) & = & P(Z \leq z) \\ & = & P(|X - Y| \leq z) \\ & = & \mbox {Area\ of\ }E\ . \end{eqnarray*} [[/math]]
Thus, we have

[[math]] F_Z(z) = \left \{ \begin{array}{ll} 0, & \mbox{if $z \le 0$}, \\ 1 - (1 - z)^2, & \mbox{if $0 \le z \le 1$}, \\ 1, & \mbox{if $z \gt 1$}. \end{array} \right. [[/math]]

Calculation of [math]F_{Z}[/math].
The density [math]f_Z(z)[/math] is again obtained by differentiation:

[[math]] f_Z(z) = \left \{ \begin{array}{ll} 0, & \mbox{if $z \le 0$}, \\ 2(1-z), & \mbox{if $0 \le z \le 1$}, \\ 0, & \mbox{if $z \gt 1$}. \end{array} \right. [[/math]]

Example There are many occasions where we observe a sequence of occurrences which occur at “random” times. For example, we might be observing emissions of a radioactive isotope, or cars passing a milepost on a highway, or light bulbs burning out. In such cases, we might define a random variable [math]X[/math] to denote the time between successive occurrences. Clearly, [math]X[/math] is a continuous random variable whose range consists of the non-negative real numbers. It is often the case that we can model [math]X[/math] by using the exponential density. This density is given by the formula

[[math]] f(t) = \left \{ \begin{array}{ll} \lambda e^{-\lambda t}, & \mbox{if $t \ge 0$}, \\ 0, & \mbox{if $t \lt 0$}. \end{array} \right. [[/math]]
The number [math]\lambda[/math] is a non-negative real number, and represents the reciprocal of the average value of [math]X[/math]. (This will be shown in Chapter Expected Value and Variance.) Thus, if the average time between occurrences is 30 minutes, then [math]\lambda = 1/30[/math]. A graph of this density function with [math]\lambda = 1/30[/math] is shown in Figure.

Spinner experiment.

One can see from the figure that even though the average value is 30, occasionally much larger values are taken on by [math]X[/math].


Suppose that we have bought a computer that contains a Warp 9 hard drive. The salesperson says that the average time between breakdowns of this type of hard drive is 30 months. It is often assumed that the length of time between breakdowns is distributed according to the exponential density. We will assume that this model applies here, with [math]\lambda = 1/30[/math].


Now suppose that we have been operating our computer for 15 months. We assume that the original hard drive is still running. We ask how long we should expect the hard drive to continue to run. One could reasonably expect that the hard drive will run, on the average, another 15 months. (One might also guess that it will run more than 15 months, since the fact that it has already run for 15 months implies that we don't have a lemon.) The time which we have to wait is a new random variable, which we will call [math]Y[/math]. Obviously, [math]Y = X - 15[/math]. We can write a computer program to produce a sequence of simulated [math]Y[/math]-values. To do this, we first produce a sequence of [math]X[/math]'s, and discard those values which are less than or equal to 15 (these values correspond to the cases where the hard drive has quit running before 15 months). To simulate a value of [math]X[/math], we compute the value of the expression

[[math]] \Bigl(-{1\over{\lambda}}\Bigr)\log(rnd)\ , [[/math]]
where [math]rnd[/math] represents a random real number between 0 and 1. (That this expression has the exponential density will be shown in Chapter Distributions and Densities.) Figure shows an area bar graph of 10,00 simulated [math]Y[/math]-values.


The average value of [math]Y[/math] in this simulation is 29.74, which is closer to the original average life span of 30 months than to the value of 15 months which was guessed above. Also, the distribution of [math]Y[/math] is seen to be close to the distribution of [math]X[/math]. It is in fact the case that [math]X[/math] and [math]Y[/math] have the same distribution. This property is called the memoryless property, because the amount of time that we have to wait for an occurrence does not depend on how long we have already waited. The only continuous density function with this property is the exponential density.

Residual lifespan of a hard drive.


Assignment of Probabilities

A fundamental question in practice is: How shall we choose the probability density function in describing any given experiment? The answer depends to a great extent on the amount and kind of information available to us about the experiment. In some cases, we can see that the outcomes are equally likely. In some cases, we can see that the experiment resembles another already described by a known density. In some cases, we can run the experiment a large number of times and make a reasonable guess at the density on the basis of the observed distribution of outcomes, as we did in Chapter Discrete Probability Distributions. In general, the problem of choosing the right density function for a given experiment is a central problem for the experimenter and is not always easy to solve (see Example). We shall not examine this question in detail here but instead shall assume that the right density is already known for each of the experiments under study. The introduction of suitable coordinates to describe a continuous sample space, and a suitable density to describe its probabilities, is not always so obvious, as our final example shows.

Infinite Tree

Example Consider an experiment in which a fair coin is tossed repeatedly, without stopping. We have seen in Example that, for a coin tossed [math]n[/math] times, the natural sample space is a binary tree with [math]n[/math] stages. On this evidence we expect that for a coin tossed repeatedly, the natural sample space is a binary tree with an infinite number of stages, as indicated in Figure. It is surprising to learn that, although the [math]n[/math]-stage tree is obviously a finite sample space, the unlimited tree can be described as a continuous sample space. To see how this comes about, let us agree that a typical outcome of the unlimited coin tossing experiment can be described by a sequence of the form [math]\omega = \{\mbox{H H T H T T H}\dots\}[/math]. If we write 1 for H and 0 for T, then [math]\omega = \{1\ 1\ 0\ 1\ 0\ 0\ 1\dots\}[/math]. In this way, each outcome is described by a sequence of 0's and 1's.

Tree for infinite number of tosses of a coin.


Now suppose we think of this sequence of 0's and 1's as the binary expansion of some real number [math]x = .1101001\cdots[/math] lying between 0 and 1. (A binary expansion is like a decimal expansion but based on 2 instead of 10.) Then each outcome is described by a value of [math]x[/math], and in this way [math]x[/math] becomes a coordinate for the sample space, taking on all real values between 0 and 1. (We note that it is possible for two different sequences to correspond to the same real number; for example, the sequences [math]\{\mbox{T H H H H H}\ldots\}[/math] and [math]\{\mbox{H T T T T T}\ldots\}[/math] both correspond to the real number [math]1/2[/math]. We will not concern ourselves with this apparent problem here.)


What probabilities should be assigned to the events of this sample space? Consider, for example, the event [math]E[/math] consisting of all outcomes for which the first toss comes up heads and the second tails. Every such outcome has the form [math].10****\cdots[/math], where [math]*[/math] can be either 0 or 1. Now if [math]x[/math] is our real-valued coordinate, then the value of [math]x[/math] for every such outcome must lie between [math]1/2 = .10000\cdots[/math] and [math]3/4 = .11000\cdots[/math], and moreover, every value of [math]x[/math] between 1/2 and 3/4 has a binary expansion of the form [math].10****\cdots[/math]. This means that [math]\omega\in E[/math] if and only if [math]1/2 \leq x \lt 3/4[/math], and in this way we see that we can describe [math]E[/math] by the interval [math][1/2,3/4)[/math]. More generally, every event consisting of outcomes for which the results of the first [math]n[/math] tosses are prescribed is described by a binary interval of the form [math][k/2^n,(k+1)/2^n)[/math].

We have already seen in Section that in the experiment involving [math]n[/math] tosses, the probability of any one outcome must be exactly [math]1/2^n[/math]. It follows that in the unlimited toss experiment, the probability of any event consisting of outcomes for which the results of the first [math]n[/math] tosses are prescribed must also be [math]1/2^n[/math]. But [math]1/2^n[/math] is exactly the length of the interval of [math]x[/math]-values describing [math]E[/math]! Thus we see that, just as with the spinner experiment, the probability of an event [math]E[/math] is determined by what fraction of the unit interval lies in [math]E[/math]. Consider again the statement: The probability is 1/2 that a fair coin will turn up heads when tossed. We have suggested that one interpretation of this statement is that if we toss the coin indefinitely the proportion of heads will approach 1/2. That is, in our correspondence with binary sequences we expect to get a binary sequence with the proportion of 1's tending to 1/2. The event [math]E[/math] of binary sequences for which this is true is a proper subset of the set of all possible binary sequences. It does not contain, for example, the sequence [math]011011011\ldots[/math] (i.e., (011) repeated again and again). The event [math]E[/math] is actually a very complicated subset of the binary sequences, but its probability can be determined as a limit of probabilities for events with a finite number of outcomes whose probabilities are given by finite tree measures. When the probability of [math]E[/math] is computed in this way, its value is found to be 1. This remarkable result is known as the Strong Law of Large Numbers (or Law of Averages) and is one justification for our frequency concept of probability. We shall prove a weak form of this theorem in Chapter Law of Large Numbers.

General references

Doyle, Peter G. (2006). "Grinstead and Snell's Introduction to Probability" (PDF). Retrieved June 6, 2024.