The joint distribution function is a function that completely characterizes the probability distribution of a random vector.
Table of contents
It is also called joint cumulative distribution function (abbreviated as joint cdf).
Let us start with the simple case in which we have two random variables and .
Their joint cdf is defined aswhere and are two real numbers.
indicates a probability;
the comma inside the parentheses stands for AND.
In other words, the joint cdf gives the probability that two conditions are simultaneously true:
the random variable takes a value less than or equal to ;
the random variable takes a value less than or equal to .
Suppose that there are only four possible cases:
Further assume that each of these cases has probability equal to 1/4.
Let us compute, as an example, the following value of the joint distribution function:
The two conditions that need to be simultaneously true are:
There are two cases in which they are satisfied:
Therefore, we have
In the previous example we have shown a special case.
In general, the formula for the joint cdf of two discrete random variables and is:where:
is the support of the vector , that is, the set of all the values of that have a strictly positive probability of being observed;
we sum the probabilities over the setthat contains all the couples belonging to the support and such that and .
The probabilities in the sum are often written using the so-called joint probability mass function
The sum in the formula above can be easily computed with the help of a table.
Here is an example.
In this table, there are nine possible couples and they all have the same probability (1/9).
In order to compute the joint cumulative distribution function, all we need to do is to shade all the probabilities to the left of (included) and above (included).
Then, the value of is equal to the sum of the probabilities in the shaded area.
The computation of the double integral can be broken down in two steps:
first compute the inner integralwhich, in general, is a function of and ;
then calculate the outer integral
Let us make an example.
Let the joint pdf be
When and , we have
This is only one of the possible cases. We also have the two cases:
or , in which case
and , in which case
The two marginal distribution functions of and are
They can be derived from the joint cumulative distribution function as follows:where the exact meaning of the notation is
This can be demonstrated as follows:because the condition is always met and, as a consequence, the condition is satisfied whenever is true.
The proof for is analogous.
In general, we cannot derive the joint cdf from the marginals, unless we know the so-called copula function, which links the two marginals.
However, there is an important exception, discussed in the next section.
When and are independent, then the joint cdf is equal to the product of the marginals:
See the lecture on independent random variables for a proof, a discussion and some examples.
Until now, we have discussed the case of two random variables. However, the joint cdf is defined for any collection of random variables forming a random vector.
Definition The joint distribution function of a random vector is a function such that:where the entries of and are denoted by and respectively, for .
More details about joint distribution functions can be found in the lecture entitled Random vectors.
Previous entry: Integrable random variable
Next entry: Joint probability density function
Please cite as:
Taboga, Marco (2021). "Joint distribution function", Lectures on probability theory and mathematical statistics. Kindle Direct Publishing. Online appendix. https://www.statlect.com/glossary/joint-distribution-function.
Most of the learning materials found on this website are now available in a traditional textbook format.