Basics of Measure Theoretic Probability + Brownian Motion
Reference sheet for Brownian motion. Convergence theorems, measure theory, normal dist and Gaussian processes, intro to Brownian motion. IN PROGRESS.
Measure Theory Primer
Probability space a triple (Ω,F,P) where Ω a set of outcomes, F a set of events (and a σ algebra) and P:F→[0,1] a function assigning probabilities to events.
P a probability measure, i.e.,
P(A)≥P(∅)=0 for all A∈F
If Ai∈F a countable sequence of disjoint sets, then P(∪iAi)=∑i(Ai)
Total mass is 1.
It is easy to show that the intersection of σ algebras is itself a σ algebra. Then for a set Ω and a collection A of subsets of Ω, there exists a smallest σ algebra containing A, denoted σ(A).
In particular, the Borel σ algebra B(R) is the smallest σ algebra on R.
Let (S,S) be an arbitrary measurable space. A function X:Ω→S is said to be a measurable map from (Ω,F) to (S,S) if
X−1(B)={w:X(w)∈B}∈F
for all B∈S.
If (S,S)=(Rd,B(Rd)) and d>1, X is a random vector. For d=1, X is a random variable!
Univariate Normal Distribution
To understand Brownian motion as a Gaussian process (GP), we need to be clear on the normal and multivariate normal (MVN) distributions. A random variable Z is said to have the standard normal distribution if its probability density function (PDF) ϕ is:
ϕ(z)=2π1e−z2/2
for −∞<z<∞.
Integrating, we have its cumulative distribution function (CDF) Φ as
Φ(z)=∫−∞zϕ(t)dt=∫−∞z2π1e−t2/2dt,
which we usually leave written as is.
If Z∼N(0,1), then a random variable X=μ+σZ for constants μ,σ is distributed ∼N(μ,σ2).
The CDF and PDF of X would be
F(x)=Φ(σx−μ),f(x)=ϕ(σx−μ)σ1
For a normally distributed r.v. X, we can also normalize to get back to the standard normal. That is, if X∼N(μ,σ2), the standardized version is
σX−μ∼N(0,1).
Joint Distributions
The distribution of a random variable X tells us about the probability of X falling into any subset of the real line. The joint distribution of random variables X,Y tells us about the probability of (X,Y) falling into a subset of the plane.
Marginal distribution of X tells us dist. of X ignoring Y. Conditional distribution of X tells us the dist. of X after observing Y=y.
For a continuous joint distribution, the CDF
FX,Y(x,y)=P(X≤x,Y≤y)
must be differentiable w.r.t. x,y such that the joint PDF
fX,Y(x,y)=dxdyd2FX,Y(x,y)
exists. We define the covariance between two random variables X,Y as
Cov(X,Y)=E((X−EX)(Y−EY)).
We can think of this as measuring what happens when X and / or Y deviate from their expected values. For example, if large X implies small Y, the covariance would be negative. Note that
Zero covariance means that two random variables are uncorrelated. We define the correlation as the covariance normalized between −1 and 1:
Corr(X,Y)=Var(X)Var(Y)Cov(X,Y).
Multivariate Normal Distribution
The k-dimensional random vector X=(X1,...,Xk) has a multivariate normal distribution if every linear combination of the Xj has a normal distribution.
The multivariate normal distribution is fully specified by the mean of the components and the covariance matrix.
A Gaussian process (GP) is a stochastic process such that every finite collection of those random variables has a multivariate normal distribution.
Brownian Motion
Standard Brownian motion (SBM) is a stochastic process W with
continuous paths
stationary, independent increments
W(t)∼N(0,T) for all t≥0.
Continuous paths
What does it mean for something to have “continuous paths”?
Recall that a stochastic process is a collection of random variables on a common probability space (Ω,F,P) for Ω the sample space, F a σ-algebra, and P a probability measure.
The sample space consists of all possible outcomes/trajectories. We can think of these as sequences. For X a stochastic process, we can view it in two ways. For fixed time t,
Xt(w):=X(t,w):Ω→S,
where X maps a realization w to whatever value it takes at t in the state space S. This is a random variable.
On the other hand, for fixed outcome w∈Ω, we have
Xw(t):=X(t,w):T→S,
where X maps a time t to a value in the state space S. This is called a sample function or realization. In the case where T is time, we call Xw(t) a sample path.
So, in Brownian motion, when we say “continuous paths,” we mean for w∈Ω, that the stochastic process W(⋅,w) is continuous with probability 1:
P(w∈Ω:W(⋅,w)∈C)=1.
Stationary & independent increments
Stationary: We want the distribution of W(t)−W(s) to depend only on t−s:
Independent: W(t1)−W(t0),W(t2)−W(t1),...,W(tn)−W(tn−1) are independent from each other.
W(t)−W(d)∼W(t−s)−W(s−s)=W(t−s)∼N(0,t−s).
The above shows us that the increments are normally distributed, too.