Skip to content

Expectation

Integration, Lebesgue Convergence Theorem

Definition: Simple Random Variables

A random variable X is said to be simple if there exists a finite family (Ak)1kn of pairwise disjoint events and real nonzero numbers (αk)1kn such that:

X=k=1nαk1Ak

We denote by L0,step the set of all simple random variables.

Simple RV Simple RV

Warning

Clearly, L0,step is a subset of L0. However, the decomposition of a simple random variable X into (Ak)1kn and (αk)1kn is not unique. Indeed, let A and B be two elements of F such that AB. It follows that X=1B=1BA+1A.

Lemma

The spaces L0 as well as L0,step are vector spaces.

Proof

The proof is left as an exercise.

Let X=1knαk1Ak and Y=1lmβl1Bl be two simple random variables. It is possible to find a finite family (Cr)1jr of pairwise disjoint elements in F and two finite families (αj~)1jr and (β~j)1jr of nonzero numbers such that:

X=1jrα~j1Cj,Y=1jrβ~j1Cj

Indeed, let Ckl=AkBl, which constitutes a finite family of pairwise disjoint elements in F. It follows that:

X=1kn1lmαk1Ckl,Y=1kn1lmβl1Ckl

In other words, any two simple random variables can be decomposed on the same common family of pairwise disjoint elements.

Definition: Expectation 1.0

We define the expectation of a simple random variable X=1knαk1Ak with respect to P as:

E^[X]:=knαkP[Ak]

Expectation Simple RV Expecation Simple RV

Exercise

Show that the definition of expectation is a well-defined operator on L0,step. Indeed, the decomposition of X is not unique.

Proposition

On L0,step, the following properties hold:

  • Monotonicity: E^[X]E^[Y] whenever XY.
  • Linearity: E^ is a linear operator on L0,step.

Proof

Let X and Y be two simple random variables. Since those two simple random variables can be decomposed on a common set of events we can write:

X=knαk1Ak,Y=kmβk1Ak

If XY, it follows that αk=X(ω)Y(ω)=βk for every state ω in Ak. Hence:

E^[X]=knαkP[Ak]knβkP[Ak]=E^[Y]

For real numbers a and b, it holds:

E^[aX+bY]=kn(aαk+bβk)P[Ak]=aknαkP[Ak]+bknβkP[Ak]=aE^[X]+bE^[Y]

This proposition is important in so far that ir shows that the expectation is a linear operator which is monotone. This monotonicity property allows to extends naturally from below the expectation to the class of positive random variables.

  • First Approximation


    Expectation 1 Expectation 1

  • Second Approximation


    Expectation 2 Expectation 2

Definition: Expectation 2.0

For any positive extended random variable X in L¯+0, we define the expectation of which as

E[X]:=sup{E^[Y]:YX,YL+0,step}

A random variable X is called integrable if E[X+]< and E[X]<. The set of integrable random variables is denoted by L1 and the expectation of which is defined as

E[X]:=E[X+]E[X]

Remark

  • Show as an exercise that for a positive extended random variable X where P[X=]>0, then it follows that E[X]=;
  • Clearly L0,stepL1;
  • Also, by definition and monotonicity of E^, for every XL0,step, it holds that E[X]=E^[X]. In other terms, E is an extension of E^ to the space L¯+0. We therefore remove the hat on the top of the expectation symbol everywhere.

Lemma

For every X and Y in L¯+0 and a,b positive numbers, it holds:

  • E[X]E[Y] whenever XY.
  • E[aX+bY]=aE[X]+bE[Y].

Proof

The proof is left as an exercise.

Theorem: Lebegue's Monotone Convergence Theorem

Let (Xn) be an increasing sequence of positive random variables. Denote by X=limXn=supXn the resulting extended positive random variable limit of the sequence. Then it holds

E[X]=E[limXn]=limE[Xn]

Proof

By monotonicity, we clearly have E[Xn]E[X] for every n, therefore supE[Xn]E[X]. Reciprocally, suppose that E[X]< and pick ε>0 and some simple positive random variable Y such that YX and E[X]εE[Y]. For 0<c<1 define the sets An={XncY}. Since Xn is increasing to X, it follows that An is an increasing sequence of events. Furthermore, since cYYX and cY<X on {X>0}, it follows that An=Ω. By non-negativity of Xn and monotonicity, it follows that

cE[1AnY]E[1AnXn]E[Xn]

and so

csupE[1AnY]supE[Xn]

Since Y=lkαl1Bl for α1,,αkR+ and B1,,BkF, it follows that

E[1AnY]=lkαlP[AnBl].

However, since P is a probability measure, and An is increasing to Ω, it follows from the lower semi-continuity of probability measures that P[AnBl]P[ΩBl]=P[Bl], and so

supE[1AnY]=lkαlsupP[AnBl]=αlP[Bl]=E[Y].

Consequently

E[X]limE[Xn]=supE[Xn]cE[Y]cE[X]cε

which by letting c converging to 1 and ε to 0 yields the result. The case where E[X]= is similar and left to the reader.

As the previous figure suggests, it is actually possible to construct by hand a sequential approximation of positive random variables by simple ones.

Proposition: Approximation by Simple Random Variables

For any positive random variabel X, there exists an increasing sequence of simple positive random variables (Xn) such that Xn(ω)X(ω) and uniformly on each set {XM} where MR.

Proof

Let Akn={(k1)/2nX<k/2n} for k=1,,n2n and define

Xn:=k=1n2nk12n1Akn+n1{X>n}

From the definition, it follows that XnX for every n and X(ω)2nXn(ω) for every ω{Xn}. This, along with the monotonicity, concludes the proof.

Proposition

For X and Y in L1, a real number a and two disjoint events A,B in F. The following assertions hold:

  1. 1AX, X+Y, aX and |X+Y| are integrable.
  2. E[(1A+1B)X]=E[1AX]+E[1BY].
  3. E[X+Y]=E[X]+E[Y] and E[aX]=aE[X].
  4. E[X]E[Y] whenever XY.
  5. If X0 and E[X]=0, then P[X=0]=1.
  6. If P[XY]=0, then E[X]=E[Y].
  7. If Z is a random variable such that |Z|X, then Z is integrable.

Remark

In particular, L1 is a vector space and the expectation operator E:L1R is a monotone, positive, and linear functional.

Proof

  1. It holds |X+Y||X|+|Y|. According to Lemma \ref{lem:linearityL0+}, it follows that E[|X+Y|]E[|X|+|Y|]=E[|X|]+E[|Y|]<, showing that X+Y and |X+Y| are integrable. The argumentation for 1AX and aX follows the same line.

  2. It holds (1A+1B)X=(1A+1B)X+(1A+1B)X. From the linearity on L+0, it follows that E[(1A+1B)X±]=E[1AX±]+E[1BX±], showing that E[(1A+1B)X]=E[1AX]+E[1BX].

  3. Without loss of generality, assume that a0. Here again, it follows from aX=aX+aX and from the linearity on L+0 that E[aX±]=aE[X±]. Also, since X+Y=(X++Y+)(X+Y)=(X+Y)+(X+Y), it follows that (X++Y+)+(X+Y)=(X+Y)+(X+Y)+. However, again from the linearity on L0+, it holds E[(X++Y+)+(X+Y)]=E[X+]+E[Y+]+E[(X+Y)] and E[(X+Y)+(X+Y)+]=E[X]+E[Y]+E[(X+Y)+], showing that E[X+Y]=E[(X+Y)+]E[(X+Y)]=E[X+]E[X]+E[Y+]E[Y]=E[X]+E[Y].

  4. If XY, it follows that 0YX. According to the proposition stating the approximation from below, let (Zn) be an increasing sequence of positive simple random variables such that ZnYX. It follows from the monotone convergence Theorem that 0E[Zn]supE[Zn]=E[YX]. Applying the previous point, we get E[YX]=E[Y]E[X], yielding the assertion.

  5. Let An={X1/n} which is an increasing sequence of events such that An={X>0}. It follows that 1An1/n1AnXX since X is positive. Monotonicity from the previous point yields P[An]/nE[1AnX]E[X]=0, showing that P[An]=0 for every n. By the lower semi-continuity property of measures, it follows that P[A]=supP[An]=0, showing that P[X>0]=0.

  6. Suppose that P[X0]=0 and define Xn=|X|1{X=0}+(|X|n)1An where An={|X|1/n}. On the one hand, by definition, An is an increasing sequence such that An={|X|0}. Hence, 0Xn|X|, which by the monotone convergence Theorem implies that E[Xn]E[|X|]. On the other hand, An{X0}, which by monotonicity of the measure implies that P[An]=0 for every n. Hence, E[Xn]nP[An]=0 for every n. We conclude that E[|X|]=0, which implies that E[X]=0.

  7. Follows directly from the linearity on L+0.

Remark

Note that for XL¯0 with XL1, and YL1, the same argumentation as above yields that:

<E[XY]=E[X+XY]=E[X+]E[X]E[Y]=E[X]E[Y]

We finish this section with two of the most important assertions of integration theory.

Theorem: Fatou's Lemma and Lebegue's Dominated Convergence

Let (Xn) be a sequence in L0.

  • Fatou's Lemma: Suppose that XnY for some YL1. Then it holds

    E[lim infXn]lim infE[Xn].
  • Dominated Convergence Theorem: Suppose that |Xn|Y for some YL1 and Xn(ω)X(ω) for any state ω. Then it holds

    E[X]=limE[Xn].

Proof

By linearity, up to the variable change XnY, we can assume that Xn is positive since E[lim infXnY]=E[lim infXn]Y and E[XnY]=E[Xn]Y for every n. Let Yn=infknXn, which is an increasing sequence of positive random variables that converges to lim infXn=supninfknXk. Notice also that YnXk for every kn, and therefore by monotonicity of the expectation E[Yn]infknE[Xk]. We conclude Fatou's lemma with the monotone convergence theorem as follows:

E[lim infXn]=limE[Yn]=supE[Yn]supninfknE[Xk]=lim infE[Xn].

A simple sign change shows that Fatou's lemma holds in the other direction. That is, if XnY for some YL1, then it holds

lim supE[Xn]E[lim supXn].

Now the dominated convergence theorem assumptions yield that YXnY for some YL1. Hence, since X=limXn=lim infXn=lim supXn, it follows that

lim supE[Xn]E[lim supXn]=E[X]=E[lim infXn]lim infE[Xn].

However, lim infE[Xn]lim supE[Xn], showing that E[Xn] converges, and

E[X]=lim infE[Xn]=lim supE[Xn]=limE[Xn].

which ends the proof.

Example: Defining a Probability Measure from a Density

The concept of density is quite often used in statistics as it defines new measures. Let us formalize it using dominated convergence.

On a probability space (Ω,F,P), consider a positive integrable random variable Z such that E[Z]=1. We define the set function

Q:F[0,1]AQ[A]=E[Z1A]

which is clearly well defined and mapping to [0,1] since Z is positive and E[Z1A]E[Z]=1.

It follows that Q defined as such is a new probability measure. Indeed

  • Q[]=E[Z1]=E[0]=0, Q[Ω]=E[Z1Ω]=E[Z]=1;
  • σ-additivity: Let (An) be a sequence of disjoint events. It follows that

    1knAk=kn1Ak1An=1An

    By monotone convergence

    Q[An]=limknQ[Ak]=limknE[Z1Ak]=limE[Zkn1Ak]=E[Z1An]=E[Z1An]=Q[An]

It can be shown using step functions that integration under P and Q are related through the formula

EQ[X]=EP[ZX]

for any bounded random variable X or any X with sufficient integrability under Q.

Another particular property of the probability measure Q so defined is that it is absolutely continuous with respect to P in the sense that

P[A]=0impliesQ[A]=0