One of the central topics in probability theory and statistics is the study of
sequences of random variables, that is, of
sequences
whose generic element
is a random variable.
Table of contents
A sequence of random variables is also often called a random sequence or a stochastic process.
There are several reasons why random sequences are important.
In statistical
inference,
is often an estimate of an unknown quantity.
The properties of
depend on the sample size
,
that is, on the number of observations used to compute the estimate.
Usually, we are able to analyze the properties of
only asymptotically, as
tends to infinity.
In this case,
is a sequence of estimates and we analyze the properties of the limit of
,
in the hope that a large sample (the one we observe) and an infinite sample
(the one we analyze by taking the limit of
)
have a similar behavior.
Examples of asymptotic results are:
the law of large numbers;
In many applications a random variable is observed repeatedly through time (for example, the price of a stock is observed every day).
In this case
is the sequence of observations of the random variable and
is a time-index (in the stock price example,
is the price observed in the
-th
period).
Often, we need to analyze a random variable
,
but for some reasons
is too complex to analyze directly.
What we usually do in this case is to approximate
by simpler random variables
that are easier to study.
The approximating random variables are arranged into a sequence
and they become better and better approximations of
as
increases.
For example, this is what we did when we introduced the Lebesgue integral.
Let
be a sequence of real numbers and
a sequence of random variables.
If the real number
is a realization
of the random variable
for every
,
then we say that the sequence of real numbers
is a realization of the sequence of random variables
.
We
write
Let
be a sample space.
Let
be a sequence of random variables.
We say that
is a sequence of random variables defined on the sample space
if and only if all the random variables
belonging to the sequence
are functions from
to
.
Let
be a sequence of random variables defined on a sample space
.
A finite subset of
is any finite set of random variables belonging to the sequence.
We say that
is an independent sequence of random variables (or a sequence
of independent random variables) if and only if every finite subset of
is a set of mutually independent random
variables.
Let
be a sequence of random variables.
Denote by
the distribution function
of a generic element of the sequence
.
We say that
is a sequence of identically distributed random variables if
and only if any two elements of the sequence have the same distribution
function:
Let
be a sequence of random variables defined on a sample space
.
We say that
is a sequence of independent and identically distributed random
variables (or an IID sequence of random variables) if and only if
is both a sequence of independent random variables and a
sequence of identically distributed random variables.
Let
be a sequence of random variables defined on a sample space
.
Take a first group of
successive terms of the sequence
,
...,
.
Now take a second group of
successive terms of the sequence
,
...,
.
The second group is located
positions after the first group.
Denote the joint
distribution function of the first group of terms
byand
the joint distribution function of the second group of terms
by
The sequence
is said to be stationary (or strictly
stationary) if and only
if
for
any
and for any vector
.
In other words, a sequence is strictly stationary if and only if the two
random vectors
and
have the same distribution (for any
,
and
).
Strict stationarity is a weaker requirement than the IID
assumption: if
is an IID sequence, then it is also strictly stationary, while the converse is
not necessarily true.
Let
be a random sequence defined on a sample space
.
We say that
is a covariance stationary sequence (or weakly stationary
sequence) if and only
if
where
and
are, of course, integers.
Property (1) means that all the random variables belonging to the sequence
have the same mean.
Property (2) means that the covariance between a
term
of
the sequence and the term that is located
positions before it
(
)
is always the same, irrespective of how
has been chosen.
In other words,
depends only on
and not on
.
Since
,
Property (2) implies that all the random variables in the sequence have the
same
variance:
Note that strictly stationarity implies weak
stationarity only if the mean
and all the covariances
exist and are finite.
Obviously, covariance stationarity does not imply strict stationarity: the former imposes restrictions only on the first and second moments, while the latter imposes restrictions on the whole distribution.
Let
be a sequence of random variables defined on a sample space
.
A sequence
is mixing if any two groups of terms of the sequence that are far apart from
each other are approximately independent (and the further the closer to being
independent).
Take a first group of
successive terms of the sequence
,
...,
.
Now take a second group of
successive terms of the sequence
,
...,
.
The second group is located
positions after the first group.
The two groups of terms are independent if and only if
for
any two functions
and
.
As explained in the lecture on mutual
independence, this is just the definition of independence between the two
random vectors
and
The above condition can be written
as
If this condition is true asymptotically (i.e., when
),
then we say that the sequence
is mixing.
Definition
We say that a sequence of random variables
is mixing (or strongly mixing) if and only
if
for
any two functions
and
and for any
and
.
In other words, a sequence is strongly mixing if and only if the two random
vectors
and
tend to become more and more independent by increasing
(for any
and
).
This is a milder requirement than the requirement of independence (see Independent sequences above):
if
is an independent sequence, all its terms are independent from one another;
if
is a mixing sequence, its terms can be dependent, but they become less and
less dependent as the distance between their locations in the sequence
increases.
Of course, an independent sequence is also a mixing sequence, while the converse is not necessarily true.
In this section we discuss ergodicity. Roughly speaking, ergodicity is a weak concept of independence for sequences of random variables.
In the subsections above we have discussed other two concepts of independence for sequences of random variables:
independent sequences are sequences of random variables whose terms are mutually independent;
mixing sequences are sequences of random variables whose terms can be dependent but become less and less dependent as their distance increases (by distance we mean how far apart they are located in the sequence).
Requiring that a random sequence be mixing is weaker than requiring that a sequence be independent: in fact, an independent sequence is also mixing, but the converse is not true.
Requiring that a sequence be ergodic is even weaker than requiring that a sequence be mixing. In fact, mixing implies ergodicity, but not vice versa.
This is probably all you need to know if you are not studying asymptotic theory at an advanced level because ergodicity is quite a complicated topic and the definition of ergodicity is fairly abstract. Nevertheless, we give here a quick definition of ergodicity for the sake of completeness.
Denote by
the set of all possible sequences of real numbers.
When
is a sequence of real numbers, denote by
the subsequence obtained by dropping the first term of
,
that
is,
We say that a subset
is a shift invariant set if and only if
belongs to
whenever
belongs to
.
Definition
A set
is shift invariant if and only
if
Shift invariance is used to define ergodicity.
Definition
A sequence of random variables
is said to be an ergodic sequence if an only
if
whenever
is a shift invariant set.
As we explained in the lecture entitled Limit of a sequence, whenever we want to assess whether a sequence is convergent to a limit, we need to define a distance function (or metric) to measure the distance between the terms of the sequence.
Intuitively, a sequence converges to a limit if, by dropping a sufficiently high number of initial terms of the sequence, the remaining terms can be made as close to each other as we wish.
The problem is how to define "close to each other".
As we have explained, the concept of "close to each other" can be made fully rigorous by using the notion of a metric. Therefore, discussing convergence of a sequence of random variables boils down to discussing what metrics can be used to measure the distance between two random variables.
In other lectures, we introduce several different notions of convergence of a sequence of random variables: to each different notion corresponds a different way of measuring the distance between two random variables.
The notions of convergence (also called modes of convergence) are:
This lecture was focused on sequences of random variables. For sequences of random vectors, please go to this lecture.
Please cite as:
Taboga, Marco (2021). "Sequence of random variables", Lectures on probability theory and mathematical statistics. Kindle Direct Publishing. Online appendix. https://www.statlect.com/asymptotic-theory/sequences-of-random-variables.
Most of the learning materials found on this website are now available in a traditional textbook format.