Escolar Documentos
Profissional Documentos
Cultura Documentos
IQR
Q1
Q3
Q1 1.5 IQR
Q3 + 1.5 IQR
Median
4
2.698
0.6745
0.6745
2.698
mode
24.65%
4
50%
0
24.65%
1
50% 50%
15.73%
4
68.27%
0
median
15.73%
1
mean
FURTHER DETAILS
3 Formal denition
(This denition may be extended to any probability distribution using the measure-theoretic denition of probability.)
A]
=
dP
=
f d
version 3.61012 nanoseconds = 1 hour).
X 1 A
A
There is a probability density function f with f(5 hours) = for any measurable set A A .
2 hour1 . The integral of f over any window of time (not
only innitesimal windows but also large windows) is the
probability that the bacterium dies in that window.
3.1 Discussion
A probability density function is most commonly associated with absolutely continuous univariate distributions.
A random variable X has density fX, where fX is a nonnegative Lebesgue-integrable function, if:
Pr[a X b] =
fX (x) dx.
4 Further details
Hence, if FX is the cumulative distribution function of X, Unlike a probability, a probability density function can
take on values greater than one; for example, the uniform
then:
distribution on the interval [0, ] has probability density
f(x) = 2 for 0 x and f(x) = 0 elsewhere.
x
The standard normal distribution has probability density
FX (x) =
fX (u) du,
d
FX (x).
fX (x) =
dx
2
1
ex /2 .
f (x) =
2
3
Not every probability distribution has a density function:
the distributions of discrete random variables do not; nor
does the Cantor distribution, even though it has no discrete component, i.e., does not assign positive probability
to any individual point.
d
F (x) = f (x).
dx
6 Families of densities
If a probability distribution admits a density, then the It is common for probability density functions (and
probability of every one-point set {a} is zero; the same probability mass functions) to be parametrizedthat is,
holds for nite and countable sets.
to be characterized by unspecied parameters. For exTwo probability densities f and g represent the same ample, the normal distribution is parametrized in terms
2
probability distribution precisely if they dier only on a of the mean and the variance, denoted by and respectively,
giving
the
family
of
densities
set of Lebesgue measure zero.
In the eld of statistical physics, a non-formal reformu1 x 2
lation of the relation above between the derivative of the
1
f (x; , 2 ) = e 2 ( ) .
cumulative distribution function and the probability den 2
sity function is generally used as the denition of the
probability density function. This alternate denition is It is important to keep in mind the dierence between the
domain of a family of densities and the parameters of the
the following:
family. Dierent values of the parameters describe difIf dt is an innitely small number, the probability that X ferent distributions of dierent random variables on the
is included within the interval (t, t + dt) is equal to f(t) dt, same sample space (the same set of all possible values of
or:
the variable); this sample space is the domain of the family of random variables that this family of distributions
describes. A given set of parameters describes a single
Pr(t < X < t + dt) = f (t) dt.
distribution within the family sharing the functional form
of the density. From the perspective of a given distribution, the parameters are constants, and terms in a den5 Link between discrete and con- sity function that contain only parameters, but not variables, are part of the normalization factor of a distribution
tinuous distributions
(the multiplicative factor that ensures that the area under
the densitythe probability of something in the domain
It is possible to represent certain discrete random vari- occurring equals 1). This normalization factor is outables as well as random variables involving both a con- side the kernel of the distribution.
tinuous and a discrete part with a generalized probability
Since the parameters are constants, reparametrizing a
density function, by using the Dirac delta function. For
density in terms of dierent parameters, to give a charexample, let us consider a binary discrete random variacterization of a dierent random variable in the famable having the Rademacher distributionthat is, taking
ily, means simply substituting the new parameter values
1 or 1 for values, with probability each. The density
into the formula in place of the old ones. Changing the
of probability associated with this variable is:
domain of a probability density, however, is trickier and
requires more work: see the section below on change of
variables.
1
f (t) = ((t + 1) + (t 1)).
2
More generally, if a discrete variable can take n dierent
values among real numbers, then the associated probability density function is:
f (t) =
i=1
pi (t xi ),
Pr (X1 , , XN
7.4 Example
D) =
fX1 , ,XN (x1 , , xN ) dx1positive
dxNx. and y is
D
If F(x1 , , xn) = Pr(X1 x1 , , Xn xn) is the
Pr
(X
>
0,
Y
>
0)
=
fX,Y (x, y) dx dy.
cumulative distribution function of the vector (X1 , ,
0
0
Xn), then the joint probability density function can be
computed as a partial derivative
F
f (x) =
x1 xn
n
fXi (xi ) =
7.2
d 1
f (x1 , , xn ) dx1 dxi1 dxi+1 dxn .fY (y) = (g (y)) fX (g 1 (y)).
dy
Independence
7.3
Corollary
d
dx
d
If the joint probability density function of a vector of n
fY (y) = fX (x) = (x) fX (x) = (g 1 (y)) fX (g 1 (y)) =
random variables can be factored into a product of n funcdy
dy
dy
tions of one variable
For functions which are not monotonic the probability
density function for y is
fX1 , ,Xn (x1 , , xn ) = f1 (x1 ) fn (xn ),
(where each is not necessarily a density) then the n variables in the set are all independent from each other, and
the marginal probability density function of each of them
is given by
fXi (xi ) =
fi (xi )
.
fi (x) dx
d
g 1 (y) fX (g 1 (y))
k
dy k
n(y)
k=1
5
fgX of the new random variable Y = g(X). However,
rather than computing
fY (y) =
E(g(X)) =
E(g(X)) =
The values of the two integrals are the same in all cases in
Not to be confused with Mixture distribution
which both X and g(X) actually have probability density
functions. It is not necessary that g be a one-to-one function. In some cases the latter integral is computed much The probability density function of the sum of two
more easily than the former.
independent random variables U and V, each of which
has a probability density function, is the convolution of
their separate density functions:
8.1
Multiple variables
y=g(x1 , ,xn )
f (x1 , , xn )
dV
n
g
2
j=1 xj (x1 , , xn )
fU +V (x) =
12 BIBLIOGRAPHY
Then, the joint density p(Y,Z) can be computed by a This leads to:
change of variables from U,V to Y,Z, and Y can be derived by marginalizing out Z from the joint density.
p(Y ) =
pU (Y Z) pV (Z) |Z| dZ
2 2
1
1 2
1
1
e 2 Y Z e 2 Z |Z| dZ
2
2
2
2
1
1 (Y +1)Z
=
e 2
|Z| dZ
2
1 1 (Y 2 +1)Z 2
=2
e 2
Z dZ
2
0
1 (Y 2 +1)u
e
du
=
0
]
1
(Y 2 +1)u
=
e
(Y 2 + 1)
u=0
1
=
(Y 2 + 1)
=
U =YZ
V =Z
The Jacobian matrix J(U, V |Y, Z) of this transformation
is
U
Y
V
Y
U
Z
V
Z
Z
=
0
Y
= |Z|.
1
Thus:
u = 12 Z 2
is paUstandard
distribution.
p(Y, Z) = p(U, V ) J(U, V |Y, Z) = p(U ) p(V ) J(U, V |Y,This
Z) =
(Y Z) pVCauchy
(Z) |Z|.
And the distribution of Y can be computed by
marginalizing out Z:
p(Y ) =
pU (Y Z) pV (Z) |Z| dZ
Note that this method crucially requires that the transformation from U,V to Y,Z be bijective. The above transformation meets this because Z can be mapped directly back
to V, and for a given V the quotient U/V is monotonic.
This is similarly the case for the sum U+V, dierence
U-V and product UV.
Exactly the same method can be used to compute the distribution of other functions of multiple independent random variables.
10.2
Given two standard normal variables U and V, the quotient can be computed as follows. First, the variables have
the following density functions:
1
p(U ) = e
2
2
U2
V2
1
p(V ) = e 2
2
We transform as described above:
11 See also
12 Bibliography
Pierre Simon de Laplace (1812). Analytical Theory
of Probability.
The rst major treatise blending
calculus with probability theory,
originally in French: Thorie Analytique des Probabilits.
Andrei Nikolajevich Kolmogorov (1950). Foundations of the Theory of Probability.
The modern measure-theoretic
foundation of probability theory; the original German version
(Grundbegrie der Wahrscheinlichkeitsrechnung) appeared in
1933.
Patrick Billingsley (1979). Probability and Measure.
New York, Toronto, London: John Wiley and Sons.
ISBN 0-471-00710-2.
David Stirzaker (2003).
ISBN 0-521-42028-8.
Elementary Probability.
13
External links
14
14
14.1
14.2
Images
14.3
Content license