Main Page | See live article | Alphabetical index

Cumulative distribution function

The cumulative distribution function (abbreviated cdf) describes the probability distribution of a real-valued random variable, X, completely. For every real number x, the cdf is given by

the probability that the variable X takes on a value less than or equal to x. The probability that X lies in the interval (a,b) is therefore F(b) - F(a) if a <= b. It is conventional to use a capital F for a cumulative distribution function, in contrast to the lower-case f used for probability density functions and probability mass functions.

Examples

As an example, suppose X is uniformly distributed on the unit interval [0,1]. Then the cdf is given by

F(x) = 0, if x < 0
F(x) = x, if 0 <= x <= 1
F(x) = 1, if x > 1

For a different example, suppose X takes only the values 0 and 1, with equal probability. Then the cdf is given by

F(x) = 0, if x < 0
F(x) = 1/2, if 0 <= x < 1
F(x) = 1, if x >= 1

Properties

Every cumulative distribution function F is monotone increasing and continuous from the right. Furthermore, we have limx→-∞ F(x) = 0 and limx→∞ F(x) = 1. Every function with these four properties is a cdf.

If X is a discrete random variable, then it attains values x1, x2, ... with probability p1, p2 etc., and the cdf of X will be discontinuous at the points xi and constant in between.

If the cdf F of X is continuous, then X is a continuous random variable; if furthermore F is absolutely continuous, then there exists a Lebesgue-integrable function f(x) such that

for all real numbers a and b. (The first of the two equalities displayed above would not be correct in general if we had not said that the distribution is continuous. Continuity of the distribution implies that P(X = a) = P(X = b) = 0, so the difference between "" and "" ceases to be important in this context.) The function f is equal to the derivative of F almost everywhere, and it is called the probability density function of the distribution of X.

The Kolmogorov-Smirnov test is based on cumulative distribution functions and can be used to test to see whether two empirical distributions are different or whether an empirical distribution is different from an ideal distribution. The closely related Kuiper's test (pronounced in Dutch the way an Cowper might be pronounced in English) is useful if the domain of the distribution is cyclic as in day of the week. For instance we might use Kuiper's test to see if the number of tornadoes varies during the year or if sales of a product vary by day of the week or day of the month.

See also Descriptive statistics, Probability distribution.