The joint probability mass function is a function that completely characterizes the distribution of a discrete random vector. When evaluated at a given point, it gives the probability that the realization of the random vector will be equal to that point.
The term joint probability function is often used as a synonym. Sometimes, the abbreviation joint pmf is used.
The following is a formal definition.
Definition
Let
be a
discrete random vector. Its joint probability mass function is a function
such
that
where
is the probability
that the random vector
takes the value
.
This is a straightforward multi-variate generalization of the definition of the probability mass function of a discrete variable (uni-variate case).
Suppose
is a
discrete random vector and that its
support (the set of
values it can take)
is:
If the three values have the same probability, then the joint probability mass
function
is:
When the two components of
are denoted by
and
,
the joint pmf can also be written using the following alternative
notation:
The joint pmf can be used to derive the marginal probability mass functions of the single entries of the random vector.
Given
,
the marginal of
is
In order to get the entire marginal, we need to compute
separately for each
belonging to the support of
.
Each of the computations involves a sum over all the possible values of
(i.e., over the support
).
Similarly, the marginal of
is
Let us derive the marginal pmf of
from the joint pmf in the previous example.
The supports of
and
are
We
haveand
Thus, the marginal probability mass function of
is
If a random vector has two entries
and
,
then its joint pmf can be written in tabular form:
each row corresponds to one of the possible values of
;
each column corresponds to one of the possible values of
;
each cell contains the probability of a couple of values
.
In the next example it will become clear why the tabular form is very convenient.
Let us put in tabular form the joint pmf used in the previous examples.
x2=0 | x2=1 | Marginal of X1 | |
---|---|---|---|
x1=0 | 0 | 1/3 | 1/3 |
x1=1 | 1/3 | 1/3 | 2/3 |
Marginal of X2 | 1/3 | 2/3 |
We can easily obtain the marginals by summing the probabilities by column and by row.
The joint pmf can also be used to derive the conditional probability mass function of the single entries of the random vector.
This is carefully explained and illustrated with examples in the glossary entry on conditional pmfs.
For a thorough discussion of joint pmfs, go to the lecture entitled Random vectors, where discrete random vectors are introduced and you can also find some solved exercises involving joint pmfs.
Previous entry: Joint probability density function
Next entry: Log likelihood
Please cite as:
Taboga, Marco (2021). "Joint probability mass function", Lectures on probability theory and mathematical statistics. Kindle Direct Publishing. Online appendix. https://www.statlect.com/glossary/joint-probability-mass-function.
Most of the learning materials found on this website are now available in a traditional textbook format.