2.1. Set Theory#

2.1.1. Set theory basics#

To start our coverage here, we need to know some elementary basics about sets. I am assuming you are already relatively familiar with sets, so this should be largely review:

Definition 2.1 (Set union)

Suppose that A,B are sets. The union of A and B is the set:

AB{ω:ωA or ωB}.

Definition 2.2 (Set intersection)

Suppose that A,B are sets. The intersection of A and B is the set:

AB{ω:ωA and ωB}.

Definition 2.3 (Subset)

Suppose that Ω is a set. A is a subset of Ω if for every ωA, then ωΩ. We denote this as AΩ.

This definition is what is usually used for subsets, but it turns out this is actually called an improper subset. The distinction is extremely small, but notice here that AA (a set is a subset of itself). For this reason, we have a further term, a proper subset:

Definition 2.4 (Proper subset)

Suppose that Ω is a set. A is a proper subset of Ω if for every ωA, then ωΩ, but AΩ. We denote this as AΩ.

So, in the proper subset definition here, Ω necessarily contains elements ωΩ where ωA. Further, we can flip these two definitions around, and obtain their equivalent superset definitions: ΩA and ΩA, respectively.

We can use this to define the set complement:

Definition 2.5 (Set complement)

Suppose that AΩ. The complement of A in Ω is:

Ac(Ω){ωΩ:ωA}.

Definition 2.6 (Set difference)

Suppose that A,BΩ, and AB. The set difference BA is:

BA{ωB:ωA}.

This can get extremely cumbersome, and when we are talking about sets, the it’s usually really clear what the superset is, so we’ll generally abbreviate this Ac.

Definition 2.7 (Disjoint sets)

Suppose that A,B are sets. A and B are called disjoint if AB=. If A and B are disjoint, their union AB is abbreviated AB, to emphasize that it is a disjoint union.

So the idea here is that A and B simply contain nothing in common, and AB is just notation which emphasizes that it is a union of disjoint sets. Sometimes we’ll be explicit when we’re talking about disjoint sets, and other times we’ll use the abbreviated version above (AB=).

All of these terms can be extended to countable (or uncountable) sequences of sets, too:

Definition 2.8 (Arbitrary union)

Suppose that (Ai)iI are a sequence of sets. Their union is:

iIAi{ω:there exists i where ωAi}.

Basically, the union is the set that contains all of the elements of the arbitrary sequence of sets. Likewise, we can define an arbitrary intersection:

Definition 2.9 (Arbitrary union)

Suppose that A,B are sets. The union of A and B is the set:

iIAi{ω:ωAi for all iI}.

Here, the union is the set that contains the elements which are common to all of the arbitrary sequence of sets. We can use this to obtain arbitrary sequences of mutually disjoint sets, too:

Definition 2.10 (Mutually disjoint sets)

Suppose that (Ai)iI are a sequence of sets, where I is an indexing set (countable or uncountable). (Ai)iI are called mutually disjoint if for all i,jI, AiAj=.

A lot of times when you’re proving things about sets, it can really help intuitively to set the proof up first using pictures. For the basic set operations, these pictures would look like this:

../../_images/basic_ops.png

Fig. 2.1 Here, the blue circle represents the space Ω upon which three sets A, B, and C are defined. In this setup, notice that CB. (B) shows the set union of A with B, (C) shows the set intersection of A with B, (D) shows the complement of A, and (E) shows the set difference of B and C.#

As a quick primer, take a chance to prove some things about set differences, and prove De Morgan’s laws:

Example 2.1 (Set difference as intersection)

Suppose that A,BΩ, and AB. Show that BABAc.

Hint: reason out using the definitions of the set differences, the set complement, and the set intersection, and first think about it using a picture like Fig. 2.1 above.

Note: this one should be pretty easy, so if you’re struggling, try starting with a book on set theory before tackling this book.

Example 2.2 (De Morgan’s Laws I)

Suppose that Ω is a set, and A,BΩ. Then:

  1. (AB)c=AcBc, and

  2. (AB)c=AcBc.

Hint: Do this by showing that if an element is in each resulting set on the left, that it follows boolean properties (either or ) A and/or B, respectively.

When you conceptualize sets, you should know these two properties inside and out! As a quick follow-up exercise, show that these laws extend to finitely many operations, too:

Example 2.3 (De Morgan’s Laws II)

Suppose that Ω is a set, and AmΩ, for all mN. Then if nN:

  1. (m[n]Am)c=m[n]Amc, and

  2. (m[n]Am)c=m[n]Amc.

Hint: Use De Morgan’s Laws I.

These also extend to arbitrary indexing sets:

Example 2.4 (De Morgan’s Laws III)

Suppose that Ω is a set, and AiΩ, for all iI, where I is an arbitrary indexing set (countable or uncountable). Then:

  1. (nNAm)c=m[n]Amc, and

  2. (nNAm)c=m[n]Amc.

Hint: Use a similar proofing approach to De Morgan’s Laws I.

2.1.2. Establishing units upon which we can build random variables#

To begin to delineate how we can come to terms with this seeming inconsistency when we handle discrete outcomes (like coin flips) with continuous outcomes (like throwing a ball), we need to create a sort of homologous system on which we can ascribe probabilities, that works whether or not the units are discrete or continous. To do this, we’re going to define families of sets (they are sets that contain sets) with increasing complexity (each family of sets you see below is also described by the subsection that precedes it, but not necessarily the reverse).

2.1.3. π-systems#

The first, and simplest, family of sets that we’re going to see so far is called a π-system. A π-system is defined on a set Ω, which can be anything (such as the real line).

Definition 2.11 (π-system)

Suppose that Ω is a set. P is called a π-system on Ω if it is a non-empty collection of subsets P of Ω, where if P1,P2P, then P1P2P.

In words, a π-system is a non-empty family of sets P that is closed under finitely many intersections. Here, a set being closed under an operation means that, if you were to consider this operation on elements of that set, the result would still be in the set. That it is closed under finitely many interesections means that if you had sets any collection of n-many sets (nN is finite) where PmP for m[n] (stated another way, (Pm)m[n]P), that m[n]PmP, too.

Example 2.5 (Example of a simple π-system)

Let Ω={1,2,3}. Show that a π-system on Ω is P={,{1},{1,2},{1,3}} is a π-system.

One note that is going to be important later on is that a π-system on Ω does not, necessarily, contain the entire set Ω itself.

As we build up, π-systems are going to be extremely useful because, as you can see above, the statement for a π-system is pretty easy to wrap your head around: closed under finitely many intersections. When we prove things later on about much, much more complicated families of sets, there will be some extremely nice results about π-systems that will come in handy. In effect, what we are going to do later on is we are going to make statements that hold on π-systems, and then use extremely basic extensions about the particularities that differentiate π-systems from these more complicated families of sets to make much more general statements taht are true about the complicated family of sets. So, for now, you can put π-systems in the back of your head, but they will come up pretty soon.

Example 2.6 (π-systems are closed under finitely many intersections)

Show that if P is a π-system on a set Ω, that if (Pm)m[n]P, that m[n]PmP, too.

2.1.4. Algebras#

The next simplest family of sets are called algebras. An algebra can be thought of as an extension of a π-system, in that it is also defined on a set Ω, and it includes closure under finitely many intersections like a π-system, but it adds a few interesting properties:

Definition 2.12 (Algebra)

Suppose that Ω is a set. A is called an algebra on Ω if it is a non-empty collection of subsets AF, s.t. AΩ, where:

  1. Contains the entire set: ΩA,

  2. Closure under complement: If AA, then AcA, and

  3. Closure under finitely many unions: If A1,A2A, then A1A2A.

It isn’t too tough to determine that an algebra is a π-system, simply by using De’Morgan’s laws. We’ll use this to get your brain in shape for set theory types of proofs, which will grow in difficulty as we go:

Theorem 2.1 (Every algebra is a π-system)

Suppose the set Ω, where A is an algebra on Ω. A is also a π-system on Ω.

Proof. Suppose that A1,A2A.

For A to be a π-system, we need to show that A1A2A.

Notice that since A is closed under complements, that A1c,A2cA as well.

Since A is closed under finitely many unions, then A1cA2cA.

Finally, since A is closed under complements, then (A1cA2c)cA.

By De Morgan’s Law, (A1cA2c)c=A1A2, so A1A2=(A1cA2c)cA, as desired.

So there’s our first set theory proof. Not bad, right?

In the language we described π-systems, algebras can be understood similarly: they add that the entire set itself Ω must be contained, and they also add closure under complement and finitely many unions.

2.1.5. σ-algebras#

The first ingredient for us to build this homologous system that we’ve been working towards is called a σ-algebra. A σ-algebra is defined on a set Ω, which can be anything (such as the real line). σ-algebras are going to take the property of closure under finitely many unions from algebras, and take it a step further:

Definition 2.13 (σ-algebra)

Suppose that Ω is a set. F is called a σ-algebra on Ω if it is a non-empty collection of events FF, s.t. FΩ, where:

  1. Contain event space: ΩF,

  2. Closure under complement: If FF, then FcF,

  3. Closure under countable unions: If {Fn}nNF is a countable sequence of events, then nNFnF.

You’ll notice that we called these “subsets” of F “events” instead of just subsets (the language we used for π-systems and algebras): the reason for this will become a little more apparent in a section or two, so just bear with us for a few pages. We’ll often call the entire set Ω the “event space”, because it is the set in which our “events” F live (FΩ). F collects these events in a very particular way. It contains the entire “event space” Ω, it is closed under complements, and if a countable set of events are in F, then the union of these events are also in F (closure under countable unions). Using similar insight to the insight you used for algebras in Theorem 2.1, you can also conclude that σ-algebras are closed under countable intersections, as well, which is a property we’ll use a bit more later on:

Property 2.1 (σ-algebras are closed under countable intersections)

Suppose that Ω is a set, and F is a σ-algebra on Ω. Show that if {Fi}iNF is a countable sequence of events, nNFnF.

2.1.5.1. Properties of σ-algebras#

The idea of a σ-algebra is very simple to conceptualize; a set which contains sets which obey these 3 rules we described above. However, as we will see here, these 3 laws imply that σ-algebras also follow many other critical properties. Together, the properties about σ-algebras (and tools that we can define on them, such as measures, which we will learn about later), will allow us to define many higher-order operations like integration.

If we have a pair of σ-algebras (each of which is a collection of subsets of the event space), and we take the subsets of the event space that are in their intersection, what is the result? It turns out that the result is also a σ-algebra:

Property 2.2 (intersection of σ-algebras are σ-algebras)

Let {Fi}iI be a collection of σ-algebras on Ω, where I is an arbitrary indexing set (which may be countable or uncountable). Then iIFi is a σ-algebra.

Notice that the indexing set here is, unlike some of the statements we’ve seen so far, potentially uncountable. This means that the family of σ-algebras are closed under intersections, without any qualifications needed about the number of intersections we might take being finite nor countable. That it is closed under arbitrarily many intersections will be critical for several useful properties about σ-algebras that you will use later on.

When proving things about σ-algebras, you always need to show exactly the properties in the definition. This means that every proof where we seek to show something is a σ-algebra is going to have 3 parts, one for each delineation in the definition:

Proof. Denote F=iIFi. To see that it is a σ-algebra, note:

1. Contains Ω: Note that by definition, since Fi are σ-algebras, then ΩFi, for all iI.

Then since ΩFi for all iI, ΩF.

2. Closed under complement: Let AjF.

Then by definition of the intersection, AjFi, for all iI.

Then since each Fi are σ-algebras, then AjcFi, since Fi are closed under intersection.

Then since AjcFi for all iI, AjcF, by definition of the intersection.

3. Closed under countable unions: Let {An}nNF be a countable set of elements of F.

Then by definition of the intersection, {An}nNFi, for all iI, by definition of the intersection operation.

Then since Fi is a σ-algebra and hence closed under countable unions, nNAnFi for all iI.

Then nNAnF, by definition of the intersection.

If we have an event space Ω, and a collection of sets that we are interested in A, can we make a σ-algebra from the collection of sets we are interested in? The answer is yes, and it is called the σ-generated algebra:

Definition 2.14 (σ-generated algebra)

Let Ω be an event space, and let A be a collection of sets, where AAAΩ. Then the smallest σ-algebra containing A is denoted σ(A), and is defined:

σ(A)=Fi{Fi:Fi is a σ-algebra on Ω and AFi}Fi

Notice that by Property 2.2, that σ(A) is also a σ-algebra.

We’ll often want to consider σ-algebras generated by a particular type of sets, called partitions:

Definition 2.15 (Partition)

Suppose the event space Ω. A family of sets A is called a partition if:

  1. It does not contain the empty set: A,

  2. The sets are disjoint: If Ωi,ΩjA, then ΩiΩj=, and

  3. The sets in A exhaust Ω: ΩnAΩi=Ω.

Can we describe the σ-algebra generated by these partitions easily? It turns out, we can:

Lemma 2.1 (σ-algebra generated by a partition)

Let Ω be an event space with σ-algebra F, where {Ωi}iIF is a partition of Ω. Then:

σ(Ωi:iI)G{jJ:JIΩj}

Proof. To start, we need to show that G is a σ-algebra. We can do this in 3 steps: 1. Contains the event space: Let J=I, so ΩnAΩiG.

But since {Ωi}iI is a partition of ω, ΩnAΩi=ΩG.

2. Closed under complements: Suppose that AG.

Then by construction, there exists JI s.t. A=jJΩj.

Take T=IJ. By construction, TI, and further, TJ=.

Define B=jTΩj. Since TI, then BG.

Notice that AB=, and AB=Ω, so B=ΩA, and BAc.

Then AcG.

3. Closed under countable unions: Suppose that AnG, for nN.

Then there exists Jn, s.t. An=jJnΩj, by construction.

Further, note that J=nNJnI.

Then A=nJAn has an indexing set JI, so AI.

To see that σ(Ωi:iI) is equivalent to G, we can do this in parts:

1. Gσ(Ωi:iI): Note that G is a σ-algebra containing each Ωi, for all iI, by taking Ji={i}.

Then σ(Ωi:iI)G, since σ(Ωi:iI) is the intersection of σ-algebras containing Ωis by definition Definition 2.14, and G is one such σ-algebra.

2. Gσ(Ωi:iI): Suppose that AG.

Then by construction, there exists JI s.t. A=jJΩj.

Then since σ(Ωi:iI) is the intersection of all σ-algebras containing {Ωi:iI}, σ(Ωi:iI) must contain all of the countable unions of elements of {Ωi:iI}, of which A is one.

Then Aσ(Ωi:iI).

Another interesting property is that σ-generated algebras preserve the subset operation. While this might seem rather intuitive, it isn’t immediately obvious, but this lemma will be extremely helpful later on:

Lemma 2.2 (Generated algebras preserve subsets)

Let Ω be an event space, and let C,F be families of sets on Ω (for every CC and FF, C,FΩ). If CF, then σ(C)σ(F).

Proof. By definition, σ(F) is the σ-algebra at the intersection of all σ-algebras containing F.

Since CF, then Cσ(F), since σ(F) is the intersection of all σ-algebras containing F (and consequently, also contain C) by definition of σ(F) Definition 2.14.

Then since σ(C) is the intersection of every σ-algebra containing C, it is the intersection of σ(F) (since Cσ(F)) with other σ-algebras that contain C but might not contain F.

Since the intersection operation of a set σ(F) with other sets will necessarily be at most σ(C), then σ(C)σ(F).

2.1.5.2. Borel σ-algebras#

Next, we have one of the most important results about σ-algebras that we will use repeatedly throughout this book when looking at probabilities: the Borel σ-algebra. The Borel σ-algebra will allow us to break up intervals of the (or the entire) real line into palatable open intervals. When we begin to define probability measures, we will generally define them using Borel σ-algebras.

Definition 2.16 (Borel σ-algebra)

Suppose that Ω is a set. The borel σ-algebra is defined as:

B(Ω)σ({AΩ:A is open})

So, the Borel σ-algebra is the σ-algebra generated by the family of sets containing all open sets in Ω.

There are a variety of special Borel σ-algebras that we will deal with in this book, for which we’ll use special notation to simplify things. They are:

Remark 2.1 (Special Borel σ-algebras)

  1. If Ω=R, then RB(R),

  2. If Ω=Rd, then RdB(Rd),

  3. If Ω=[α,β], is a closed interval s.g. ΩR, then R[α,β]B([α,β]),

  4. If Ω=(α,β), is a closed interval s.g. ΩR, then R[α,β]B((α,β))

Next, we’ll introduce some sets which generate R. These will be fundamental for many proofs you’ll find later on:

Example 2.7 (Generators of R)

Show that the families of sets PR={(,x]:xR} and PQ={(,x):xQ} are generators of R.

Hint: Begin by showing that PR and PQ are π-systems on R.

Then, show that B(R)σ(π(P)) (using that any open subset GR is in σ(π(R)), and σ(P)B(R)) (using that any subset of P can be expressed using the closure properties of σ-algebras on open subsets of R).

2.1.6. Set-Theoretic Limits and extrema#

Just like we have extrama of sequences, we can conceptualize extrama of sets, too.

Definition 2.17 (Supremum of sets)

Suppose that Ω is an event space, and FnΩ for nN is a sequence of sets. Then:

supnNFnnNFn.

The idea here is that the supremum of a sequence of sets is just the union of the whole sequence of sets. To draw an analogy to the definition of a supremum of a sequence, you can understand this as the supremum of a sequence of sets is the smallest possible set (in terms of the number of elements in it) which contains all of the elements of the entire sequence. Intuitively, this idea is just the union. Likewise, we can conceptualize the infimum:

Definition 2.18 (Infimum of sets)

Suppose that Ω is an event space, and FnΩ for nN is a sequence of sets. Then:

infnNFnnNFn.

Likewise, here the idea is that the infinmum of a sequence of sets is just the intersection of the whole sequence of sets. Extending our analogy, the infimum of a sequence of sets is the largest possible set which contains elements common to the entire sequence of sets. This idea is captured by an intersection. Try to see what happens when, further, the sequence of sets is captured by a σ-algebra:

Example 2.8 (σ-algebras are closed under extrema)

Suppose that Ω is an event space with σ-algebra F, and (Fn)nNF is a sequence of sets in the σ-algebra. Show that supnNFn and infnNFn are also in the σ-algebra.

Hint: these proofs should be short and sweet, and borrow from the definition of σ-algebras and De Morgan’s Law III.

Likewise, we can conceptualize set-theoretic versions of the limits you have come to appreciate in real analysis:

Definition 2.19 (Set-theoretic limits of sets)

Suppose that Ω is an event space, and FnΩ for nN is a sequence of sets. Then:

lim infnFninfnNsupmnFnnNmnFn,lim supnFninfnNsupmnFnnNmnFn,

and further, if lim supnNFn=lim infnNFn, and we say that:

limnFn=lim infnFn=lim supnFn

σ-algebras are also closed under set-theoretic limits, too:

Example 2.9 (σ-algebra is closed under set-theoretic limits)

Suppose that Ω is an event space with σ-algebra F, and {Fn}nNF is a sequence of sets in the σ-algebra. Show that lim supnFn and lim infnFn are also in the σ-algebra.

Hint: use your result from Example 2.8 and borrow from the definition of σ-algebras.

We can further clarify some special situations where the set-theoretic limit, lim, of a sequence of sets exists, using monotone sequences of sets:

Definition 2.20 (Monotone sequence of sets)

Suppose that Ω is an event space, and FnΩ for nN is a sequence of sets. Then:

  1. {Fn} is called monotone non-increasing if for every nN, Fn+1Fn, and

  2. {Fn} is called monotone non-decreasing if for every nN, FnFn+1.

The ideas here are that, like for sequences of numbers, monotonicity here refers to the composition of elements in the sets. Monotone non-increasing sets are those where every sequential set is contained in the preceding set, and monotone non-decreasing sets are those where every sequential set contains the preceding set.

If we have a sequence of sets which is monotone, the set-theoretic limits exist:

Lemma 2.3 (Monotone sequences of sets have limits)

Suppose that Ω is an event space, and FnΩ for nN is a monotone (non-increasing or non-decreasing) sequence of sets. Then:

limnFn=lim infnFn=lim supnFn.

Proof. We have two cases, the case where {Fn} is monotone non-increasing or non-decreasing: 1. Monotone non-increasing: If {Fn} is monotone non-increasing, then for any nN:

mNFm=mnFm,Fn=mnFm,

by definition of Fn+1Fn. Then:

lim infnFn=supnNinfmnFm=nNmnFm=nNmNFm,mNFm=mnFm=nNFn,double  is redundant=nNmnFm,Fn=mnFm=infnNsupmnFm=lim supnFn.

2. Monotone non-decreasing:

Repeat the same argument, noting that if {Fn} is monotone non-increasing, then FnFn+1, so for every nN:

Fn=mnFm,mNFm=mnFm.

Then in both cases, lim infnFn=lim supnFn, and we can define limnFn to be this set.