guide:269af6cf67: Difference between revisions
No edit summary |
mNo edit summary |
||
Line 1: | Line 1: | ||
==Normal Distribution== | ==Normal Distribution== | ||
The '''normal''' (or '''Gaussian''') '''distribution''' is a very common [[ | The '''normal''' (or '''Gaussian''') '''distribution''' is a very common [[guide:82d603b116#Continuous_probability_distribution|continuous probability distribution]]. Normal distributions are important in statistics and are often used in the natural and [[social science|social science]]s to represent real-valued [[guide:1b8642f694|random variable]]s whose distributions are not known.<ref>[http://www.encyclopedia.com/topic/Normal_Distribution.aspx#3 ''Normal Distribution''], Gale Encyclopedia of Psychology</ref><ref>{{harvtxt |Casella |Berger |2001 |p=102 }}</ref> | ||
The normal distribution is useful because of the [[ | The normal distribution is useful because of the [[central limit theorem|central limit theorem]]. In its most general form, under some conditions (which include finite [[guide:E4d753a3b5|variance]]), it states that averages of [[guide:1b8642f694|random variables]] independently drawn from independent distributions [[Convergence of random variables#Convergence in distribution|converge in distribution]] to the normal, that is, become normally distributed when the number of random variables is sufficiently large. Physical quantities that are expected to be the sum of many independent processes (such as [[measurement error|measurement error]]s) often have distributions that are nearly normal.<ref>Lyon, A. (2014). [http://aidanlyon.com/aidanlyon.com/media/publications/Lyon-normal_distributions.pdf Why are Normal Distributions Normal?], The British Journal for the Philosophy of Science.</ref> Moreover, many results and methods (such as [[propagation of uncertainty|propagation of uncertainty]] and [[least squares|least squares]] parameter fitting) can be derived analytically in explicit form when the relevant variables are normally distributed. | ||
The normal distribution is sometimes informally called the '''bell curve'''. However, many other distributions are bell-shaped (such as the [[ | The normal distribution is sometimes informally called the '''bell curve'''. However, many other distributions are bell-shaped (such as the [[Cauchy distribution|Cauchy]], [[Student's t-distribution|Student's t]], and [[logistic distribution|logistic]] distributions). The terms [[Gaussian function|Gaussian function]] and Gaussian bell curve are also ambiguous because they sometimes refer to multiples of the normal distribution that cannot be directly interpreted in terms of probabilities. | ||
The [[ | The [[guide:82d603b116#Continuous probability distribution|probability density function]] of the normal distribution is:<math display="block"> | ||
f(x \; | \; \mu, \sigma^2) = \frac{1}{\sqrt{2\sigma^2\pi} } \; e^{ -\frac{(x-\mu)^2}{2\sigma^2} } | f(x \; | \; \mu, \sigma^2) = \frac{1}{\sqrt{2\sigma^2\pi} } \; e^{ -\frac{(x-\mu)^2}{2\sigma^2} } | ||
</math> | </math> | ||
Where: | Where: | ||
* <math>\mu</math> is | * <math>\mu</math> is the [[guide:82d603b116#Expected_Value|mean]] or [[guide:82d603b116#Expected_Value|expectation]] of the distribution (and also its [[median|median]] and [[mode (statistics)|mode]]) | ||
* <math>\sigma</math> is [[ | * <math>\sigma</math> is [[guide:E4d753a3b5#Standard_deviation|standard deviation]] | ||
* <math>\sigma^2</math> is [[ | * <math>\sigma^2</math> is [[guide:E4d753a3b5|variance]] | ||
A random variable with a Gaussian distribution is said to be '''normally distributed''' and is called a '''normal deviate'''. | A random variable with a Gaussian distribution is said to be '''normally distributed''' and is called a '''normal deviate'''. | ||
Line 20: | Line 20: | ||
==== Standard normal distribution ==== | ==== Standard normal distribution ==== | ||
The simplest case of a normal distribution is known as the ''standard normal distribution''. This is a special case when <math>μ=0</math> and <math>σ=1</math>, and it is described by this [[ | The simplest case of a normal distribution is known as the ''standard normal distribution''. This is a special case when <math>μ=0</math> and <math>σ=1</math>, and it is described by this [[guide:82d603b116#Continuous probability distribution|probability density function]]: | ||
<math display="block">\phi(x) = \frac{e^{- \frac{\scriptscriptstyle 1}{\scriptscriptstyle 2} x^2}}{\sqrt{2\pi}}\, </math> | <math display="block">\phi(x) = \frac{e^{- \frac{\scriptscriptstyle 1}{\scriptscriptstyle 2} x^2}}{\sqrt{2\pi}}\, </math> | ||
The factor <math>1/\sqrt{2\pi}</math> in this expression ensures that the total area under the curve <math>\phi(x)</math> is equal to one.<ref>For the proof see [[ | The factor <math>1/\sqrt{2\pi}</math> in this expression ensures that the total area under the curve <math>\phi(x)</math> is equal to one.<ref>For the proof see [[Gaussian integral|Gaussian Integral]]</ref> The ½ in the exponent ensures that the distribution has unit variance (and therefore also unit standard deviation). This function is symmetric around <math>x=0</math>, where it attains its maximum value <math>1/\sqrt{2\pi}</math>; and has [[inflection point|inflection point]]s at +1 and −1. | ||
=== General normal distribution === | === General normal distribution === | ||
Line 36: | Line 36: | ||
=== Notation === | === Notation === | ||
The standard Gaussian distribution (with zero mean and unit variance) is often denoted with the Greek letter <math>\phi</math> ([[ | The standard Gaussian distribution (with zero mean and unit variance) is often denoted with the Greek letter <math>\phi</math> ([[phi (letter)|phi]]).<ref>{{harvtxt |Halperin |Hartley |Hoel |1965 |loc=item 7 }}</ref> The alternative form of the Greek phi letter, <math>\varphi</math>, is also used quite often. | ||
The normal distribution is also often denoted by <math>N(\mu,\sigma^2)</math>.<ref>{{harvtxt |McPherson |1990 |p=110 }}</ref> Thus when a random variable <math>X</math> is distributed normally with mean <math>\mu</math> and variance <math>\sigma^2</math>, we write | The normal distribution is also often denoted by <math>N(\mu,\sigma^2)</math>.<ref>{{harvtxt |McPherson |1990 |p=110 }}</ref> Thus when a random variable <math>X</math> is distributed normally with mean <math>\mu</math> and variance <math>\sigma^2</math>, we write | ||
Line 44: | Line 44: | ||
== Properties == | == Properties == | ||
The normal distribution is a subclass of the [[ | The normal distribution is a subclass of the [[elliptical distribution|elliptical distribution]]s. The normal distribution is [[Symmetric distribution|symmetric]] about its mean, and is non-zero over the entire real line. As such it may not be a suitable model for variables that are inherently positive or strongly skewed, such as the [[weight|weight]] of a person or the price of a [[share (finance)|share]]. Such variables may be better described by other distributions, such as the [[lognormal||log-normal distribution}} or the [[Pareto distribution|Pareto distribution]]. | ||
The value of the normal distribution is practically zero when the value <math>x</math> lies more than a few [[ | The value of the normal distribution is practically zero when the value <math>x</math> lies more than a few [[guide:E4d753a3b5#Standard_deviation|standard deviation]]s away from the mean. Therefore, it may not be an appropriate model when one expects a significant fraction of [[outlier|outlier]]s—values that lie many standard deviations away from the mean—and least squares and other [[statistical inference|statistical inference]] methods that are optimal for normally distributed variables often become highly unreliable when applied to such data. In those cases, a more [[heavy-tailed|heavy-tailed]] distribution should be assumed and the appropriate [[robust statistics|robust statistical inference]] methods applied. | ||
The Gaussian distribution belongs to the family of [[ | The Gaussian distribution belongs to the family of [[stable distribution|stable distribution]]s which are the attractors of sums of [[Independent and identically distributed random variables|independent, identically distributed]] distributions whether or not the mean or variance is finite. Except for the Gaussian which is a limiting case, all stable distributions have heavy tails and infinite variance. It is one of the few distributions that are stable and that have probability density functions that can be expressed analytically, the others being the [[Cauchy distribution|Cauchy distribution]] and the [[Lévy distribution|Lévy distribution]]. | ||
=== Symmetries and derivatives === | === Symmetries and derivatives === | ||
The normal distribution <math>f(x)</math>, with any mean <math>\mu</math> and any positive deviation <math>\sigma</math>, has the following properties: | The normal distribution <math>f(x)</math>, with any mean <math>\mu</math> and any positive deviation <math>\sigma</math>, has the following properties: | ||
* It is symmetric around the point <math>x = \mu </math>, which is at the same time the [[ | * It is symmetric around the point <math>x = \mu </math>, which is at the same time the [[mode (statistics)|mode]], the [[median|median]] and the [[guide:82d603b116#Expected_Value|mean]] of the distribution and it divides the data in half.<ref name="PR2.1.4">{{harvtxt |Patel |Read |1996 |loc=[2.1.4] }}</ref> | ||
* It is [[ | * It is [[unimodal|unimodal]]: its first [[derivative|derivative]] is positive for <math>x < \mu </math>, negative for <math> x> \mu </math>, and zero only at <math>x = \mu </math>. | ||
* The area under the curve and over the x-axis is unity. | * The area under the curve and over the x-axis is unity. | ||
* Its density has two [[ | * Its density has two [[inflection point|inflection point]]s (where the second derivative of <math>f</math> is zero and changes sign), located one standard deviation away from the mean, namely at <math>x = \mu - \sigma</math> and <math>x = \mu + \sigma </math>.<ref name="PR2.1.4" /> | ||
* Its density is [[ | * Its density is [[logarithmically concave function|log-concave]].<ref name="PR2.1.4" /> | ||
* Its density is infinitely [[ | * Its density is infinitely [[differentiable function|differentiable]], indeed [[supersmooth|supersmooth]] of order 2.<ref>{{harvtxt |Fan |1991 |p=1258 }}</ref> | ||
* Its second derivative <math>f^{''}(x)</math> is equal to its derivative with respect to its variance <math>\sigma^2</math>. | * Its second derivative <math>f^{''}(x)</math> is equal to its derivative with respect to its variance <math>\sigma^2</math>. | ||
Line 71: | Line 71: | ||
\end{cases} | \end{cases} | ||
</math> | </math> | ||
Here <math>n!!</math> denotes the [[ | Here <math>n!!</math> denotes the [[double factorial|double factorial]], that is, the product of every number from <math>n</math> to 1 that has the same parity as <math>n</math>. | ||
<center> | <center> | ||
Line 97: | Line 97: | ||
=== Moment generating functions === | === Moment generating functions === | ||
The [[ | The [[moment generating function|moment generating function]] of a real random variable <math>X</math> is the expected value of <math>e^{tX}</math>, as a function of the real parameter <math>t</math>. For a normal distribution with mean <math>\mu</math> and deviation <math>\sigma</math>, the moment generating function exists and is equal to | ||
<math display="block"> M(t) = \hat \phi(-it) = e^{ \mu t} e^{\frac12 \sigma^2 t^2 }</math> | <math display="block"> M(t) = \hat \phi(-it) = e^{ \mu t} e^{\frac12 \sigma^2 t^2 }</math> | ||
== Cumulative distribution function == | == Cumulative distribution function == | ||
The [[ | The [[guide:82d603b116#Cumulative_distribution_function|cumulative distribution function]] (CDF) of the standard normal distribution, usually denoted with the capital Greek letter <math>\Phi</math> ([[phi (letter)|phi]]), is the integral | ||
<math display="block">\Phi(x)\; = \;\frac{1}{\sqrt{2\pi}} \int_{-\infty}^x e^{-t^2/2} \, dt</math> | <math display="block">\Phi(x)\; = \;\frac{1}{\sqrt{2\pi}} \int_{-\infty}^x e^{-t^2/2} \, dt</math> | ||
In statistics one often uses the related [[ | In statistics one often uses the related [[error function|error function]], or erf(<math>x</math>), defined as the probability of a random variable with normal distribution of mean 0 and variance 1/2 falling in the range <math>[-x, x]</math>; that is | ||
<math display="block">\operatorname{erf}(x)\; =\; \frac{1}{\sqrt{\pi}} \int_{-x}^x e^{-t^2} \, dt</math> | <math display="block">\operatorname{erf}(x)\; =\; \frac{1}{\sqrt{\pi}} \int_{-x}^x e^{-t^2} \, dt</math> | ||
These integrals cannot be expressed in terms of elementary functions, and are often said to be [[ | These integrals cannot be expressed in terms of elementary functions, and are often said to be [[special function|special function]]s. However, many numerical approximations are known; see [[#Numerical approximations for the normal CDF|below]]. | ||
The two functions are closely related, namely | The two functions are closely related, namely | ||
Line 118: | Line 118: | ||
For a generic normal distribution <math>f</math> with mean <math>\mu</math> and deviation <math>\sigma</math>, the cumulative distribution function is <math display="block">F(x)\;=\;\Phi\left(\frac{x-\mu}{\sigma}\right)\;=\; \frac12\left[1 + \operatorname{erf}\left(\frac{x-\mu}{\sigma\sqrt{2}}\right)\right] </math> | For a generic normal distribution <math>f</math> with mean <math>\mu</math> and deviation <math>\sigma</math>, the cumulative distribution function is <math display="block">F(x)\;=\;\Phi\left(\frac{x-\mu}{\sigma}\right)\;=\; \frac12\left[1 + \operatorname{erf}\left(\frac{x-\mu}{\sigma\sqrt{2}}\right)\right] </math> | ||
The complement of the standard normal CDF, <math>Q(x) = 1 - \Phi(x)</math>, is often called the [[ | The complement of the standard normal CDF, <math>Q(x) = 1 - \Phi(x)</math>, is often called the [[Q-function|Q-function]], especially in engineering texts.<ref>{{cite web |url=http://cnx.org/content/m11537/1.2/ |last=Scott |first=Clayton |first2=Robert |last2=Nowak |title=The Q-function |work=Connexions |date=August 7, 2003 }}</ref><ref>{{cite web |url=http://www.eng.tau.ac.il/~jo/academic/Q.pdf |last=Barak |first=Ohad |title=Q Function and Error Function |publisher=Tel Aviv University |date=April 6, 2006 }}</ref> It gives the probability that the value of a standard normal random variable will exceed <math>x</math>. | ||
=== Standard deviation and tolerance intervals === | === Standard deviation and tolerance intervals === | ||
About 68% of values drawn from a normal distribution are within one standard deviation <math>\sigma</math> away from the mean; about 95% of the values lie within two standard deviations; and about 99.7% are within three standard deviations. This fact is known as the [[ | About 68% of values drawn from a normal distribution are within one standard deviation <math>\sigma</math> away from the mean; about 95% of the values lie within two standard deviations; and about 99.7% are within three standard deviations. This fact is known as the [[68–95–99.7 rule|68-95-99.7 (empirical) rule]], or the ''3-sigma rule''. | ||
More precisely, the probability that a normal deviate lies in [<math>\mu - n\sigma </math>, <math>\mu + n\sigma </math>] is given by | More precisely, the probability that a normal deviate lies in [<math>\mu - n\sigma </math>, <math>\mu + n\sigma </math>] is given by | ||
Line 132: | Line 132: | ||
=== Quantile function === | === Quantile function === | ||
The [[ | The [[quantile function|quantile function]] of a distribution is the inverse of the cumulative distribution function. The quantile function of the standard normal distribution is called the [[probit function|probit function]], and can be expressed in terms of the inverse [[error function|error function]]: | ||
<math display="block"> | <math display="block"> | ||
Line 146: | Line 146: | ||
</math> | </math> | ||
The [[ | The [[quantile|quantile]] <math> \Phi^{-1}(p) </math> of the standard normal distribution is commonly denoted as <math>z_p</math>. A normal random variable <math>X</math> will exceed <math>\mu + \sigma z_p</math> with probability <math>1-p</math>; and will lie outside the interval <math>\mu ± \sigma z_p </math> with probability <math>2(1-p)</math>. In particular, a normal random variable will lie outside the interval <math>\mu ± 1.96\sigma </math> in only 5% of cases. | ||
===Examples=== | ===Examples=== | ||
Line 227: | Line 227: | ||
==Gamma Distribution== | ==Gamma Distribution== | ||
The '''gamma distribution''' is a two-parameter family of continuous [[ | The '''gamma distribution''' is a two-parameter family of continuous [[guide:82d603b116|expected value]]s. The common [[exponential distribution|exponential distribution]] and [[chi-squared distribution|chi-squared distribution]] are special cases of the gamma distribution. | ||
The parametrization with <math>\alpha</math> and <math>\theta</math> appears to be more common in [[ | The parametrization with <math>\alpha</math> and <math>\theta</math> appears to be more common in [[econometrics|econometrics]] and certain other applied fields, where e.g. the gamma distribution is frequently used to model waiting times. For instance, in [[Accelerated life testing|life testing]], the waiting time until death is a random variable that is frequently modeled with a gamma distribution.<ref>See Hogg and Craig (1978, Remark 3.3.1) for an explicit motivation</ref> | ||
If <math>\alpha</math> is a positive integer, then the distribution represents an [[ | If <math>\alpha</math> is a positive integer, then the distribution represents an [[Erlang distribution|Erlang distribution]]; i.e., the sum of α independent [[exponentially distributed|exponentially distributed]] random variables, each of which has a mean of <math>\theta</math>. | ||
==Characterization using shape α and scale θ== | ==Characterization using shape α and scale θ== | ||
Line 242: | Line 242: | ||
<math display="block">f(x;k,\theta) = \frac{x^{\alpha-1}e^{-\frac{x}{\theta}}}{\theta^{\alpha}\Gamma(\alpha)} \quad \text{ for } x > 0 \text{ and } , \alpha \theta > 0.</math> | <math display="block">f(x;k,\theta) = \frac{x^{\alpha-1}e^{-\frac{x}{\theta}}}{\theta^{\alpha}\Gamma(\alpha)} \quad \text{ for } x > 0 \text{ and } , \alpha \theta > 0.</math> | ||
Here <math>\Gamma(\alpha)</math> is the [[ | Here <math>\Gamma(\alpha)</math> is the [[gamma function|gamma function]] evaluated at <math>α</math>. | ||
The '''cumulative distribution function''' is the regularized gamma function: | The '''cumulative distribution function''' is the regularized gamma function: | ||
Line 248: | Line 248: | ||
<math display="block"> F(x;\alpha,\theta) = \int_0^x f(u;\alpha,\theta)\,du = \frac{\gamma\left(\alpha, \frac{x}{\theta}\right)}{\Gamma(\alpha)}</math> | <math display="block"> F(x;\alpha,\theta) = \int_0^x f(u;\alpha,\theta)\,du = \frac{\gamma\left(\alpha, \frac{x}{\theta}\right)}{\Gamma(\alpha)}</math> | ||
where <math>\gamma\left(\alpha, \frac{x}{\theta}\right)</math> is the lower [[ | where <math>\gamma\left(\alpha, \frac{x}{\theta}\right)</math> is the lower [[incomplete gamma function|incomplete gamma function]]. | ||
==Properties== | ==Properties== | ||
Line 257: | Line 257: | ||
<math display="block"> \sum_{i=1}^N X_i \sim\mathrm{Gamma} \left( \sum_{i=1}^N \alpha_i, \theta \right)</math> | <math display="block"> \sum_{i=1}^N X_i \sim\mathrm{Gamma} \left( \sum_{i=1}^N \alpha_i, \theta \right)</math> | ||
provided all <math>X_i</math> are independent. This shows that the gamma distribution exhibits [[ | provided all <math>X_i</math> are independent. This shows that the gamma distribution exhibits [[Infinite divisibility (probability)|infinite divisibility]]. | ||
==Related distributions== | ==Related distributions== | ||
Line 265: | Line 265: | ||
! Distribution !! Relation | ! Distribution !! Relation | ||
|- | |- | ||
| Exponential || If <math>X \sim \textrm{Gamma}(1,1/\lambda) </math> then <math>X</math> has an [[ | | Exponential || If <math>X \sim \textrm{Gamma}(1,1/\lambda) </math> then <math>X</math> has an [[exponential distribution|exponential distribution]] with rate parameter <math>\lambda </math>. | ||
|- | |- | ||
| Chi-square || If <math>X \sim \textrm{Gamma}(\nu/2,2) </math> then <math>X</math> is identical to <math>\chi^2(\nu)</math>, the [[ | | Chi-square || If <math>X \sim \textrm{Gamma}(\nu/2,2) </math> then <math>X</math> is identical to <math>\chi^2(\nu)</math>, the [[chi-squared distribution|chi-squared distribution]] with <math>\nu</math> degrees of freedom. Conversely, if <math> Q \sim \chi^2(\nu)</math> and <math>c</math> is a positive constant, then <math>cQ \sim \textrm{Gamma}(\nu/2,2c) </math>. | ||
|- | |- | ||
| Waiting time of Poisson process || If <math>\alpha </math> is an [[ | | Waiting time of Poisson process || If <math>\alpha </math> is an [[integer|integer]], the gamma distribution is an [[Erlang distribution|Erlang distribution]] and is the probability distribution of the waiting time until the <math>k</math><sup>th</sup> in a one-dimensional [[Poisson process|Poisson process]] with intensity <math>\theta^{-1} </math>. | ||
|- | |- | ||
| Inverse Gamma || If <math>X \sim \textrm{Gamma}(\alpha,\theta) </math>, then <math>1/X \sim \textrm{Inv-Gamma}(\alpha, \theta^{-1})</math> (see [[ | | Inverse Gamma || If <math>X \sim \textrm{Gamma}(\alpha,\theta) </math>, then <math>1/X \sim \textrm{Inv-Gamma}(\alpha, \theta^{-1})</math> (see [[Inverse-gamma distribution|Inverse-gamma distribution]] for derivation). | ||
|- | |- | ||
| Beta || If <math>X \sim \textrm{Gamma}(\alpha,\theta) </math> and <math>Y \sim \textrm{Gamma}(\beta,\theta) </math> are independently distributed, then <math>X/(X+Y)</math> has a [[ | | Beta || If <math>X \sim \textrm{Gamma}(\alpha,\theta) </math> and <math>Y \sim \textrm{Gamma}(\beta,\theta) </math> are independently distributed, then <math>X/(X+Y)</math> has a [[beta distribution|Beta distribution]] with parameters <math>\alpha</math> and <math>\beta</math>. | ||
|- | |- | ||
| Gaussian/Normal || For large <math>\alpha</math> the gamma distribution converges to a Gaussian distribution with mean <math>\mu = \alpha \theta</math> and variance <math>\sigma^2 =\alpha \theta ^2</math>. | | Gaussian/Normal || For large <math>\alpha</math> the gamma distribution converges to a Gaussian distribution with mean <math>\mu = \alpha \theta</math> and variance <math>\sigma^2 =\alpha \theta ^2</math>. | ||
Line 280: | Line 280: | ||
==Uniform== | ==Uniform== | ||
The '''continuous uniform distribution''' or '''rectangular distribution''' is a family of [[ | The '''continuous uniform distribution''' or '''rectangular distribution''' is a family of [[Symmetric distribution|symmetric]] [[probability distributions|probability distributions]] such that for each member of the family, all intervals of the same length on the distribution's support are equally probable. The support is defined by the two parameters, <math>a</math> and <math>b</math>, which are its minimum and maximum values. The distribution is often abbreviated <math>U(a,b)</math>. | ||
==Characterization== | ==Characterization== | ||
===Probability density function=== | ===Probability density function=== | ||
The [[ | The [[guide:82d603b116#Continuous probability distribution|probability density function]] of the continuous uniform distribution is:<math display="block"> | ||
f(x)=\begin{cases} | f(x)=\begin{cases} | ||
\frac{1}{b - a} & \mathrm{for}\ a \le x \le b, \\ | \frac{1}{b - a} & \mathrm{for}\ a \le x \le b, \\ | ||
Line 295: | Line 295: | ||
===Cumulative distribution function=== | ===Cumulative distribution function=== | ||
The [[ | The [[guide:82d603b116#Cumulative_distribution_function|cumulative distribution function]] is:<math display="block"> | ||
F(x)= \begin{cases} | F(x)= \begin{cases} | ||
0 & \text{for }x < a \\[8pt] | 0 & \text{for }x < a \\[8pt] | ||
Line 306: | Line 306: | ||
===Moment-generating function=== | ===Moment-generating function=== | ||
The [[ | The [[moment_generating_function|moment generating function]] is:<ref>{{harvnb|Casella|Berger|2001|page=626}}</ref> | ||
<math display="block"> | <math display="block"> | ||
Line 312: | Line 312: | ||
</math> | </math> | ||
from which we may calculate the [[ | from which we may calculate the [[guide:E4d753a3b5#Moments|raw moments]] <math>m_k</math> with | ||
<math display="block">m_k=\frac{1}{k+1}\sum_{i=0}^k a^ib^{k-i}. \,\!</math> | <math display="block">m_k=\frac{1}{k+1}\sum_{i=0}^k a^ib^{k-i}. \,\!</math> | ||
Line 320: | Line 320: | ||
===Mean and Variance=== | ===Mean and Variance=== | ||
The mean (first [[ | The mean (first [[guide:E4d753a3b5#Moments|moment]]) of the distribution is: | ||
<math display="block">\operatorname{E}(X)=\frac{1}{2}(a+b).</math> | <math display="block">\operatorname{E}(X)=\frac{1}{2}(a+b).</math> | ||
The variance (second [[ | The variance (second [[guide:E4d753a3b5#Central_Moments|central moment]]) is: | ||
<math display="block">V(X)=\frac{1}{12}(b-a)^2</math> | <math display="block">V(X)=\frac{1}{12}(b-a)^2</math> | ||
===Order statistics=== | ===Order statistics=== | ||
Let <math>X_1,\ldots,X_n</math> be an [[ | Let <math>X_1,\ldots,X_n</math> be an [[iid|i.i.d]] sample from <math>U(0,1)</math>. Let <math>X_{(k)}</math> be the <math>k</math><sup>th</sup> [[order statistic|order statistic]] from this sample. Then the probability distribution of <math>X_{(k)}</math> is a [[Beta distribution|Beta distribution]] with parameters <math>k</math> and <math>n-k+1</math>. The expected value is | ||
<math display="block">\operatorname{E}(X_{(k)}) = {k \over n+1}.</math> | <math display="block">\operatorname{E}(X_{(k)}) = {k \over n+1}.</math> | ||
Line 349: | Line 349: | ||
Restricting <math>a=0</math> and <math>b=1</math>, the resulting distribution <math>U(0,1)</math> is called a '''standard uniform distribution'''. | Restricting <math>a=0</math> and <math>b=1</math>, the resulting distribution <math>U(0,1)</math> is called a '''standard uniform distribution'''. | ||
One interesting property of the standard uniform distribution is that if <math>U</math> has a standard uniform distribution, then so does 1-<math>U</math>. This property can be used for generating [[ | One interesting property of the standard uniform distribution is that if <math>U</math> has a standard uniform distribution, then so does 1-<math>U</math>. This property can be used for generating [[antithetic variates|antithetic variates]], among other things. | ||
==Related distributions== | ==Related distributions== | ||
Line 357: | Line 357: | ||
! Distribution !! Relation | ! Distribution !! Relation | ||
|- | |- | ||
| Exponential || If <math>X</math> has a standard uniform distribution, then by the [[ | | Exponential || If <math>X</math> has a standard uniform distribution, then by the [[inverse transform sampling|inverse transform sampling]] method, <math>Y = -\lambda^{-1}\ln(X)</math> has an [[exponential distribution|exponential distribution]] with (rate) parameter <math>\lambda </math>. | ||
|- | |- | ||
| Beta || If <math>X</math> has a standard uniform distribution, then <math>Y = X^n </math> has a [[ | | Beta || If <math>X</math> has a standard uniform distribution, then <math>Y = X^n </math> has a [[beta distribution|beta distribution]] with parameters <math>n^{-1}</math> and 1. (Note this implies that the standard uniform distribution is a special case of the beta distribution, with parameters 1 and 1.) | ||
|- | |- | ||
| Irwin-Hall || The [[ | | Irwin-Hall || The [[Irwin–Hall distribution|Irwin–Hall distribution]] is the sum of <math>n</math> [[iid | i.i.d]] <math>U(0,1)</math> distributions. | ||
|- | |- | ||
| Symmetric triangle || The sum of two independent, equally distributed, uniform distributions yields a symmetric [[ | | Symmetric triangle || The sum of two independent, equally distributed, uniform distributions yields a symmetric [[triangular distribution|triangular distribution]]. | ||
|- | |- | ||
| Triangle || The distance between two [[ | | Triangle || The distance between two [[iid|i.i.d.]] uniform random variables also has a [[triangular distribution|triangular distribution]], although not symmetric. | ||
|- | |- | ||
| Beta || The uniform distribution can be thought of as a [[ | | Beta || The uniform distribution can be thought of as a [[beta distribution|beta distribution]] with parameters (1,1). | ||
|} | |} | ||
The '''exponential distribution''' is the [[ | The '''exponential distribution''' is the [[guide:82d603b116|expected value]] that describes the time between events in a [[Poisson process|Poisson process]], i.e. a process in which events occur continuously and independently at a constant average rate. It is a particular case of the [[gamma distribution|gamma distribution]]. It is the continuous analogue of the [[geometric distribution|geometric distribution]], and it has the key property of being [[memoryless|memoryless]]. In addition to being used for the analysis of [[Poisson point process|Poisson processes,]] it is found in various other contexts. | ||
The exponential distribution is not the same as the class of [[ | The exponential distribution is not the same as the class of [[exponential family|exponential families]] of distributions, which is a large class of probability distributions that includes the exponential distribution as one of its members, but also includes the [[normal distribution|normal distribution]], [[binomial distribution|binomial distribution]], [[gamma distribution|gamma distribution]], [[Poisson|Poisson]], and many others. | ||
==Characterization== | ==Characterization== | ||
===Probability density function=== | ===Probability density function=== | ||
The [[ | The [[guide:82d603b116#Continuous probability distribution|probability density function]] (pdf) of an exponential distribution is | ||
<math display="block"> | <math display="block"> | ||
Line 385: | Line 385: | ||
\end{cases}</math> | \end{cases}</math> | ||
Here <math>\lambda > 0 </math> is the parameter of the distribution, often called the ''rate parameter''. The distribution is supported on the interval [0, ∞). If a [[ | Here <math>\lambda > 0 </math> is the parameter of the distribution, often called the ''rate parameter''. The distribution is supported on the interval [0, ∞). If a [[guide:1b8642f694|random variable]] <math>X</math> has this distribution, we write <math>X \sim \textrm{Exp}(\lambda) </math>. | ||
The exponential distribution exhibits [[ | The exponential distribution exhibits [[infinite divisibility (probability)|infinite divisibility]]. | ||
===Cumulative distribution function=== | ===Cumulative distribution function=== | ||
The [[ | The [[guide:82d603b116#Cumulative_distribution_function|cumulative distribution function]] is given by<math display="block">F(x;\lambda) = \begin{cases} | ||
1-e^{-\lambda x} & x \ge 0, \\ | 1-e^{-\lambda x} & x \ge 0, \\ | ||
0 & x < 0. | 0 & x < 0. | ||
Line 396: | Line 396: | ||
===Alternative parameterization=== | ===Alternative parameterization=== | ||
A commonly used alternative parametrization is to define the [[ | A commonly used alternative parametrization is to define the [[guide:82d603b116#Continuous probability distribution|probability density function]] (pdf) of an exponential distribution as<math display="block">f(x;\theta) = \begin{cases} | ||
\frac{1}{\theta} e^{-\frac{x}{\theta}} & x \ge 0, \\ | \frac{1}{\theta} e^{-\frac{x}{\theta}} & x \ge 0, \\ | ||
0 & x < 0. | 0 & x < 0. | ||
\end{cases}</math> | \end{cases}</math> | ||
where <math>\theta > 0 </math> is the [[ | where <math>\theta > 0 </math> is the [[mean|mean]], [[guide:E4d753a3b5#Standard_deviation|standard deviation]], and [[scale parameter|scale parameter]] of the distribution, the [[multiplicative inverse|reciprocal]] of the ''rate parameter'', <math>\lambda </math>, defined above. In this specification, <math>\theta </math> is a ''survival parameter'' in the sense that if a [[guide:1b8642f694|random variable]] <math>X</math> is the duration of time that a given biological or mechanical system manages to survive and <math>X \sim \textrm{Exp}(\theta) </math> then <math>\operatorname{E}[X] = \theta </math>. That is to say, the expected duration of survival of the system is <math>\theta</math> units of time. The parametrization involving the "rate" parameter arises in the context of events arriving at a rate <math>\lambda </math>, when the time between events (which might be modeled using an exponential distribution) has a mean of <math>\theta = \lambda^{-1} </math>. | ||
The alternative specification is sometimes more convenient than the one given above, and some authors will use it as a standard definition. Unfortunately this gives rise to a [[ | The alternative specification is sometimes more convenient than the one given above, and some authors will use it as a standard definition. Unfortunately this gives rise to a [[Mathematical notation|notational]] ambiguity. In general, the reader must check which of these two specifications is being used if an author writes <math>X \sim \textrm{Exp}(\lambda) </math>, since either the notation in the previous (using <math>\lambda </math>) or the notation in this section (here, using <math>\theta </math> to avoid confusion) could be intended. | ||
==Properties== | ==Properties== | ||
Line 409: | Line 409: | ||
===Mean, variance, moments and median=== | ===Mean, variance, moments and median=== | ||
The mean or [[ | The mean or [[guide:82d603b116#Expected_Value|expected value]] of an exponentially distributed random variable <math>X</math> with rate parameter λ is given by<math display="block">\operatorname{E}[X] = \frac{1}{\lambda} = \theta</math>, see above. In light of the examples given above, this makes sense: if you receive phone calls at an average rate of 2 per hour, then you can expect to wait half an hour for every call. | ||
The [[ | The [[guide:E4d753a3b5|variance]] of <math>X</math> is given by | ||
<math display="block">\operatorname{Var}[X] = \frac{1}{\lambda^2} = \theta^2</math> | <math display="block">\operatorname{Var}[X] = \frac{1}{\lambda^2} = \theta^2</math> | ||
,so the [[ | ,so the [[guide:E4d753a3b5#Standard_deviation|standard deviation]] is equal to the mean. | ||
The [[ | The [[guide:E4d753a3b5#Moments|moments]] of <math>X</math>, for <math>n = 1, 2, \ldots </math> are given by | ||
<math display="block">\operatorname{E}\left [X^n \right ] = \frac{n!}{\lambda^n} = \theta^n n!.</math> | <math display="block">\operatorname{E}\left [X^n \right ] = \frac{n!}{\lambda^n} = \theta^n n!.</math> | ||
The [[ | The [[median|median]] of <math>X</math> is given by | ||
<math display="block">\operatorname{m}[X] = \theta \ln(2) < \operatorname{E}[X]</math> | <math display="block">\operatorname{m}[X] = \theta \ln(2) < \operatorname{E}[X]</math> | ||
Line 429: | Line 429: | ||
<math display="block">|\operatorname{E}[X]- \operatorname{m}[X]| = \frac{1- \ln(2)}{\lambda}< \frac{1}{\lambda}</math> | <math display="block">|\operatorname{E}[X]- \operatorname{m}[X]| = \frac{1- \ln(2)}{\lambda}< \frac{1}{\lambda}</math> | ||
, in accordance with the [[ | , in accordance with the [[Chebyshev's inequality#An application: distance between the mean and the median|median-mean inequality]]. | ||
===Memorylessness=== | ===Memorylessness=== | ||
Line 436: | Line 436: | ||
<math display="block">\Pr \left (T > s + t | T > s \right ) = \Pr(T > t), \qquad \forall s, t \ge 0.</math> | <math display="block">\Pr \left (T > s + t | T > s \right ) = \Pr(T > t), \qquad \forall s, t \ge 0.</math> | ||
When <math>T</math> is interpreted as the waiting time for an event to occur relative to some initial time, this relation implies that, if <math>T</math> is conditioned on a failure to observe the event over some initial period of time <math>s</math>, the distribution of the remaining waiting time is the same as the original unconditional distribution. For example, if an event has not occurred after 30 seconds, the [[ | When <math>T</math> is interpreted as the waiting time for an event to occur relative to some initial time, this relation implies that, if <math>T</math> is conditioned on a failure to observe the event over some initial period of time <math>s</math>, the distribution of the remaining waiting time is the same as the original unconditional distribution. For example, if an event has not occurred after 30 seconds, the [[conditional probability|conditional probability]] that occurrence will take at least 10 more seconds is equal to the unconditional probability of observing the event more than 10 seconds relative to the initial time. | ||
The exponential distribution and the [[ | The exponential distribution and the [[geometric distribution|geometric distribution]] are the only memoryless probability distributions;consequently, the exponential distribution is the only continuous probability distribution that has a constant [[Failure rate|Failure rate]]. | ||
==Notes== | ==Notes== | ||
Line 445: | Line 445: | ||
==References== | ==References== | ||
* [[ | * [[Robert V. Hogg|R. V. Hogg]] and A. T. Craig (1978) ''Introduction to Mathematical Statistics'', 4th edition. New York: Macmillan. (See Section 3.3.)' | ||
* P. G. Moschopoulos (1985) ''The distribution of the sum of independent gamma random variables'', '''Annals of the Institute of Statistical Mathematics''', 37, 541–544 | * P. G. Moschopoulos (1985) ''The distribution of the sum of independent gamma random variables'', '''Annals of the Institute of Statistical Mathematics''', 37, 541–544 | ||
* A. M. Mathai (1982) ''Storage capacity of a dam with gamma type inputs'', '''Annals of the Institute of Statistical Mathematics''', 34, 591–597 | * A. M. Mathai (1982) ''Storage capacity of a dam with gamma type inputs'', '''Annals of the Institute of Statistical Mathematics''', 34, 591–597 |
Revision as of 21:30, 4 April 2024
Normal Distribution
The normal (or Gaussian) distribution is a very common continuous probability distribution. Normal distributions are important in statistics and are often used in the natural and social sciences to represent real-valued random variables whose distributions are not known.[1][2]
The normal distribution is useful because of the central limit theorem. In its most general form, under some conditions (which include finite variance), it states that averages of random variables independently drawn from independent distributions converge in distribution to the normal, that is, become normally distributed when the number of random variables is sufficiently large. Physical quantities that are expected to be the sum of many independent processes (such as measurement errors) often have distributions that are nearly normal.[3] Moreover, many results and methods (such as propagation of uncertainty and least squares parameter fitting) can be derived analytically in explicit form when the relevant variables are normally distributed.
The normal distribution is sometimes informally called the bell curve. However, many other distributions are bell-shaped (such as the Cauchy, Student's t, and logistic distributions). The terms Gaussian function and Gaussian bell curve are also ambiguous because they sometimes refer to multiples of the normal distribution that cannot be directly interpreted in terms of probabilities.
The probability density function of the normal distribution is:
Where:
- [math]\mu[/math] is the mean or expectation of the distribution (and also its median and mode)
- [math]\sigma[/math] is standard deviation
- [math]\sigma^2[/math] is variance
A random variable with a Gaussian distribution is said to be normally distributed and is called a normal deviate.
Definition
Standard normal distribution
The simplest case of a normal distribution is known as the standard normal distribution. This is a special case when [math]μ=0[/math] and [math]σ=1[/math], and it is described by this probability density function:
The factor [math]1/\sqrt{2\pi}[/math] in this expression ensures that the total area under the curve [math]\phi(x)[/math] is equal to one.[4] The ½ in the exponent ensures that the distribution has unit variance (and therefore also unit standard deviation). This function is symmetric around [math]x=0[/math], where it attains its maximum value [math]1/\sqrt{2\pi}[/math]; and has inflection points at +1 and −1.
General normal distribution
Every normal distribution is a version of the standard normal distribution whose domain has been stretched by a factor [math]\sigma[/math] (the standard deviation) and then translated by [math]\mu[/math] (the mean value):
The probability density must be scaled by [math]1/\sigma[/math] so that the integral is still 1.
If [math]Z[/math] is a standard normal deviate, then [math]X = Z\sigma + \mu [/math] will have a normal distribution with expected value [math]\mu[/math] and standard deviation [math]\sigma[/math]. Conversely, if [math]X[/math] is a general normal deviate, then [math]Z = X-\mu/\sigma [/math] will have a standard normal distribution.
Notation
The standard Gaussian distribution (with zero mean and unit variance) is often denoted with the Greek letter [math]\phi[/math] (phi).[5] The alternative form of the Greek phi letter, [math]\varphi[/math], is also used quite often.
The normal distribution is also often denoted by [math]N(\mu,\sigma^2)[/math].[6] Thus when a random variable [math]X[/math] is distributed normally with mean [math]\mu[/math] and variance [math]\sigma^2[/math], we write
Properties
The normal distribution is a subclass of the elliptical distributions. The normal distribution is symmetric about its mean, and is non-zero over the entire real line. As such it may not be a suitable model for variables that are inherently positive or strongly skewed, such as the weight of a person or the price of a share. Such variables may be better described by other distributions, such as the [[lognormal||log-normal distribution}} or the Pareto distribution.
The value of the normal distribution is practically zero when the value [math]x[/math] lies more than a few standard deviations away from the mean. Therefore, it may not be an appropriate model when one expects a significant fraction of outliers—values that lie many standard deviations away from the mean—and least squares and other statistical inference methods that are optimal for normally distributed variables often become highly unreliable when applied to such data. In those cases, a more heavy-tailed distribution should be assumed and the appropriate robust statistical inference methods applied.
The Gaussian distribution belongs to the family of stable distributions which are the attractors of sums of independent, identically distributed distributions whether or not the mean or variance is finite. Except for the Gaussian which is a limiting case, all stable distributions have heavy tails and infinite variance. It is one of the few distributions that are stable and that have probability density functions that can be expressed analytically, the others being the Cauchy distribution and the Lévy distribution.
Symmetries and derivatives
The normal distribution [math]f(x)[/math], with any mean [math]\mu[/math] and any positive deviation [math]\sigma[/math], has the following properties:
- It is symmetric around the point [math]x = \mu [/math], which is at the same time the mode, the median and the mean of the distribution and it divides the data in half.[7]
- It is unimodal: its first derivative is positive for [math]x \lt \mu [/math], negative for [math] x\gt \mu [/math], and zero only at [math]x = \mu [/math].
- The area under the curve and over the x-axis is unity.
- Its density has two inflection points (where the second derivative of [math]f[/math] is zero and changes sign), located one standard deviation away from the mean, namely at [math]x = \mu - \sigma[/math] and [math]x = \mu + \sigma [/math].[7]
- Its density is log-concave.[7]
- Its density is infinitely differentiable, indeed supersmooth of order 2.[8]
- Its second derivative [math]f^{''}(x)[/math] is equal to its derivative with respect to its variance [math]\sigma^2[/math].
Moments
If [math]X[/math] has a normal distribution, the moments exist and are finite for any [math]p[/math] whose real part is greater than −1. For any non-negative integer
[math]p[/math], the plain central moments are
Here [math]n!![/math] denotes the double factorial, that is, the product of every number from [math]n[/math] to 1 that has the same parity as [math]n[/math].
Order | Non-central moment | Central moment |
---|---|---|
1 | [math]\mu[/math] | 0 |
2 | [math]\mu[/math]2 + [math]\sigma[/math]2 | [math]\sigma[/math] 2 |
3 | [math]\mu[/math]3 + 3μσ2 | 0 |
4 | [math]\mu[/math]4 + 6[math]\mu[/math]2[math]\sigma[/math]2 + 3[math]\sigma[/math]4 | 3[math]\sigma[/math] 4 |
Moment generating functions
The moment generating function of a real random variable [math]X[/math] is the expected value of [math]e^{tX}[/math], as a function of the real parameter [math]t[/math]. For a normal distribution with mean [math]\mu[/math] and deviation [math]\sigma[/math], the moment generating function exists and is equal to
Cumulative distribution function
The cumulative distribution function (CDF) of the standard normal distribution, usually denoted with the capital Greek letter [math]\Phi[/math] (phi), is the integral
In statistics one often uses the related error function, or erf([math]x[/math]), defined as the probability of a random variable with normal distribution of mean 0 and variance 1/2 falling in the range [math][-x, x][/math]; that is
These integrals cannot be expressed in terms of elementary functions, and are often said to be special functions. However, many numerical approximations are known; see below.
The two functions are closely related, namely
For a generic normal distribution [math]f[/math] with mean [math]\mu[/math] and deviation [math]\sigma[/math], the cumulative distribution function is
The complement of the standard normal CDF, [math]Q(x) = 1 - \Phi(x)[/math], is often called the Q-function, especially in engineering texts.[9][10] It gives the probability that the value of a standard normal random variable will exceed [math]x[/math].
Standard deviation and tolerance intervals
About 68% of values drawn from a normal distribution are within one standard deviation [math]\sigma[/math] away from the mean; about 95% of the values lie within two standard deviations; and about 99.7% are within three standard deviations. This fact is known as the 68-95-99.7 (empirical) rule, or the 3-sigma rule.
More precisely, the probability that a normal deviate lies in [[math]\mu - n\sigma [/math], [math]\mu + n\sigma [/math]] is given by
Quantile function
The quantile function of a distribution is the inverse of the cumulative distribution function. The quantile function of the standard normal distribution is called the probit function, and can be expressed in terms of the inverse error function:
For a normal random variable with mean [math]\mu[/math] and variance [math]\sigma^2[/math], the quantile function is
The quantile [math] \Phi^{-1}(p) [/math] of the standard normal distribution is commonly denoted as [math]z_p[/math]. A normal random variable [math]X[/math] will exceed [math]\mu + \sigma z_p[/math] with probability [math]1-p[/math]; and will lie outside the interval [math]\mu ± \sigma z_p [/math] with probability [math]2(1-p)[/math]. In particular, a normal random variable will lie outside the interval [math]\mu ± 1.96\sigma [/math] in only 5% of cases.
Examples
A professor's exam scores are approximately distributed normally with mean 80 and standard deviation 5. Using the cumulative standard normal table, we can answer the questions below.
What is the probability that a student scores an 82 or less?
What is the probability that a student scores a 90 or more?
What is the probability that a student scores a 74 or less?
What is the probability that a student scores between 74 and 82?
What is the probability that an average of three scores is 82 or less?
Gamma Distribution
The gamma distribution is a two-parameter family of continuous expected values. The common exponential distribution and chi-squared distribution are special cases of the gamma distribution.
The parametrization with [math]\alpha[/math] and [math]\theta[/math] appears to be more common in econometrics and certain other applied fields, where e.g. the gamma distribution is frequently used to model waiting times. For instance, in life testing, the waiting time until death is a random variable that is frequently modeled with a gamma distribution.[11]
If [math]\alpha[/math] is a positive integer, then the distribution represents an Erlang distribution; i.e., the sum of α independent exponentially distributed random variables, each of which has a mean of [math]\theta[/math].
Characterization using shape α and scale θ
A random variable [math]X[/math] that is gamma-distributed with shape [math]\alpha[/math] and scale [math]\theta[/math] is denoted by
The probability density function using the shape-scale parametrization is
Here [math]\Gamma(\alpha)[/math] is the gamma function evaluated at [math]α[/math].
The cumulative distribution function is the regularized gamma function:
where [math]\gamma\left(\alpha, \frac{x}{\theta}\right)[/math] is the lower incomplete gamma function.
Properties
Summation
If [math]X_i[/math] has a [math]\textrm{Gamma}(\alpha_i, \theta)[/math] distribution for [math]i =1,\ldots, N[/math] (i.e., all distributions have the same scale parameter [math]\theta_i[/math]), then
provided all [math]X_i[/math] are independent. This shows that the gamma distribution exhibits infinite divisibility.
Related distributions
Distribution | Relation |
---|---|
Exponential | If [math]X \sim \textrm{Gamma}(1,1/\lambda) [/math] then [math]X[/math] has an exponential distribution with rate parameter [math]\lambda [/math]. |
Chi-square | If [math]X \sim \textrm{Gamma}(\nu/2,2) [/math] then [math]X[/math] is identical to [math]\chi^2(\nu)[/math], the chi-squared distribution with [math]\nu[/math] degrees of freedom. Conversely, if [math] Q \sim \chi^2(\nu)[/math] and [math]c[/math] is a positive constant, then [math]cQ \sim \textrm{Gamma}(\nu/2,2c) [/math]. |
Waiting time of Poisson process | If [math]\alpha [/math] is an integer, the gamma distribution is an Erlang distribution and is the probability distribution of the waiting time until the [math]k[/math]th in a one-dimensional Poisson process with intensity [math]\theta^{-1} [/math]. |
Inverse Gamma | If [math]X \sim \textrm{Gamma}(\alpha,\theta) [/math], then [math]1/X \sim \textrm{Inv-Gamma}(\alpha, \theta^{-1})[/math] (see Inverse-gamma distribution for derivation). |
Beta | If [math]X \sim \textrm{Gamma}(\alpha,\theta) [/math] and [math]Y \sim \textrm{Gamma}(\beta,\theta) [/math] are independently distributed, then [math]X/(X+Y)[/math] has a Beta distribution with parameters [math]\alpha[/math] and [math]\beta[/math]. |
Gaussian/Normal | For large [math]\alpha[/math] the gamma distribution converges to a Gaussian distribution with mean [math]\mu = \alpha \theta[/math] and variance [math]\sigma^2 =\alpha \theta ^2[/math]. |
Uniform
The continuous uniform distribution or rectangular distribution is a family of symmetric probability distributions such that for each member of the family, all intervals of the same length on the distribution's support are equally probable. The support is defined by the two parameters, [math]a[/math] and [math]b[/math], which are its minimum and maximum values. The distribution is often abbreviated [math]U(a,b)[/math].
Characterization
Probability density function
The probability density function of the continuous uniform distribution is:
The values of [math]f(x)[/math] at the two boundaries [math]a[/math] and [math]b[/math] are usually unimportant because they do not alter the values of the integrals of [math]f(x) [/math] over any interval, nor of [math]x f(x) [/math] or any higher moment.
Cumulative distribution function
The cumulative distribution function is:
Its inverse is:
Moment-generating function
The moment generating function is:[12]
from which we may calculate the raw moments [math]m_k[/math] with
Properties
Mean and Variance
The mean (first moment) of the distribution is:
The variance (second central moment) is:
Order statistics
Let [math]X_1,\ldots,X_n[/math] be an i.i.d sample from [math]U(0,1)[/math]. Let [math]X_{(k)}[/math] be the [math]k[/math]th order statistic from this sample. Then the probability distribution of [math]X_{(k)}[/math] is a Beta distribution with parameters [math]k[/math] and [math]n-k+1[/math]. The expected value is
The variances are
Uniformity
The probability that a uniformly distributed random variable falls within any interval of fixed length is independent of the location of the interval itself (but it is dependent on the interval size), so long as the interval is contained in the distribution's support.
To see this, if [math]X \sim U(a,b) [/math] and [math][x,x+d][/math] is a subinterval of [math][a,b][/math] with fixed [math]d \gt 0 [/math], then
which is independent of [math]x[/math]. This fact motivates the distribution's name.
Standard uniform
Restricting [math]a=0[/math] and [math]b=1[/math], the resulting distribution [math]U(0,1)[/math] is called a standard uniform distribution.
One interesting property of the standard uniform distribution is that if [math]U[/math] has a standard uniform distribution, then so does 1-[math]U[/math]. This property can be used for generating antithetic variates, among other things.
Related distributions
Distribution | Relation |
---|---|
Exponential | If [math]X[/math] has a standard uniform distribution, then by the inverse transform sampling method, [math]Y = -\lambda^{-1}\ln(X)[/math] has an exponential distribution with (rate) parameter [math]\lambda [/math]. |
Beta | If [math]X[/math] has a standard uniform distribution, then [math]Y = X^n [/math] has a beta distribution with parameters [math]n^{-1}[/math] and 1. (Note this implies that the standard uniform distribution is a special case of the beta distribution, with parameters 1 and 1.) |
Irwin-Hall | The Irwin–Hall distribution is the sum of [math]n[/math] i.i.d [math]U(0,1)[/math] distributions. |
Symmetric triangle | The sum of two independent, equally distributed, uniform distributions yields a symmetric triangular distribution. |
Triangle | The distance between two i.i.d. uniform random variables also has a triangular distribution, although not symmetric. |
Beta | The uniform distribution can be thought of as a beta distribution with parameters (1,1). |
The exponential distribution is the expected value that describes the time between events in a Poisson process, i.e. a process in which events occur continuously and independently at a constant average rate. It is a particular case of the gamma distribution. It is the continuous analogue of the geometric distribution, and it has the key property of being memoryless. In addition to being used for the analysis of Poisson processes, it is found in various other contexts.
The exponential distribution is not the same as the class of exponential families of distributions, which is a large class of probability distributions that includes the exponential distribution as one of its members, but also includes the normal distribution, binomial distribution, gamma distribution, Poisson, and many others.
Characterization
Probability density function
The probability density function (pdf) of an exponential distribution is
Here [math]\lambda \gt 0 [/math] is the parameter of the distribution, often called the rate parameter. The distribution is supported on the interval [0, ∞). If a random variable [math]X[/math] has this distribution, we write [math]X \sim \textrm{Exp}(\lambda) [/math].
The exponential distribution exhibits infinite divisibility.
Cumulative distribution function
The cumulative distribution function is given by
Alternative parameterization
A commonly used alternative parametrization is to define the probability density function (pdf) of an exponential distribution as
where [math]\theta \gt 0 [/math] is the mean, standard deviation, and scale parameter of the distribution, the reciprocal of the rate parameter, [math]\lambda [/math], defined above. In this specification, [math]\theta [/math] is a survival parameter in the sense that if a random variable [math]X[/math] is the duration of time that a given biological or mechanical system manages to survive and [math]X \sim \textrm{Exp}(\theta) [/math] then [math]\operatorname{E}[X] = \theta [/math]. That is to say, the expected duration of survival of the system is [math]\theta[/math] units of time. The parametrization involving the "rate" parameter arises in the context of events arriving at a rate [math]\lambda [/math], when the time between events (which might be modeled using an exponential distribution) has a mean of [math]\theta = \lambda^{-1} [/math].
The alternative specification is sometimes more convenient than the one given above, and some authors will use it as a standard definition. Unfortunately this gives rise to a notational ambiguity. In general, the reader must check which of these two specifications is being used if an author writes [math]X \sim \textrm{Exp}(\lambda) [/math], since either the notation in the previous (using [math]\lambda [/math]) or the notation in this section (here, using [math]\theta [/math] to avoid confusion) could be intended.
Properties
Mean, variance, moments and median
The mean or expected value of an exponentially distributed random variable [math]X[/math] with rate parameter λ is given by
, see above. In light of the examples given above, this makes sense: if you receive phone calls at an average rate of 2 per hour, then you can expect to wait half an hour for every call.
The variance of [math]X[/math] is given by
,so the standard deviation is equal to the mean.
The moments of [math]X[/math], for [math]n = 1, 2, \ldots [/math] are given by
The median of [math]X[/math] is given by
, where ln refers to the natural logarithm. Thus the absolute difference between the mean and median is
, in accordance with the median-mean inequality.
Memorylessness
An exponentially distributed random variable [math]T[/math] obeys the relation
When [math]T[/math] is interpreted as the waiting time for an event to occur relative to some initial time, this relation implies that, if [math]T[/math] is conditioned on a failure to observe the event over some initial period of time [math]s[/math], the distribution of the remaining waiting time is the same as the original unconditional distribution. For example, if an event has not occurred after 30 seconds, the conditional probability that occurrence will take at least 10 more seconds is equal to the unconditional probability of observing the event more than 10 seconds relative to the initial time.
The exponential distribution and the geometric distribution are the only memoryless probability distributions;consequently, the exponential distribution is the only continuous probability distribution that has a constant Failure rate.
Notes
- Normal Distribution, Gale Encyclopedia of Psychology
- Casella & Berger (2001, p. 102)
- Lyon, A. (2014). Why are Normal Distributions Normal?, The British Journal for the Philosophy of Science.
- For the proof see Gaussian Integral
- Halperin, Hartley & Hoel (1965, item 7)
- McPherson (1990, p. 110)
- 7.0 7.1 7.2 Patel & Read (1996, [2.1.4])
- Fan (1991, p. 1258)
- Scott, Clayton; Nowak, Robert (August 7, 2003). "The Q-function". Connexions.
- Barak, Ohad (April 6, 2006). "Q Function and Error Function" (PDF). Tel Aviv University.
- See Hogg and Craig (1978, Remark 3.3.1) for an explicit motivation
- Casella & Berger 2001, p. 626
References
- R. V. Hogg and A. T. Craig (1978) Introduction to Mathematical Statistics, 4th edition. New York: Macmillan. (See Section 3.3.)'
- P. G. Moschopoulos (1985) The distribution of the sum of independent gamma random variables, Annals of the Institute of Statistical Mathematics, 37, 541–544
- A. M. Mathai (1982) Storage capacity of a dam with gamma type inputs, Annals of the Institute of Statistical Mathematics, 34, 591–597
- Wikipedia contributors. "Uniform distribution". Wikipedia. Wikipedia. Retrieved 28 January 2022.
- Wikipedia contributors. "Gamma distribution". Wikipedia. Wikipedia. Retrieved 28 January 2022.
- Wikipedia contributors. "Exponential distribution". Wikipedia. Wikipedia. Retrieved 28 January 2022.
- Casella, George; Berger, Roger L. (2001). Statistical Inference (2nd ed.). Duxbury. ISBN 0-534-24312-6.CS1 maint: ref=harv (link)
- "Recommended Standards for Statistical Symbols and Notation. COPSS Committee on Symbols and Notation" (1965). The American Statistician 19 (3): 12–14. doi: .
- McPherson, Glen (1990). Statistics in Scientific Investigation: Its Basis, Application and Interpretation. Springer-Verlag. ISBN 0-387-97137-8.CS1 maint: ref=harv (link)
- Cover, Thomas M.; Thomas, Joy A. (2006). Elements of Information Theory. John Wiley and Sons.CS1 maint: ref=harv (link)
- Fan, Jianqing (1991). "On the optimal rates of convergence for nonparametric deconvolution problems". The Annals of Statistics 19 (3): 1257–1272. doi: .
- Wikipedia contributors. "Normal distribution". Wikipedia. Wikipedia. Retrieved 28 January 2022.