Probability Theory
Last updated on Oct 29, 2021
Probability
Probability Space
A probability space is a triple
is the sample space. is the -algebra on . is a probability measure.
The sample space
What is a
Sigma Algebra
A nonempty set (of subsets of
- If
, then - If
, then
The smallest
-algebra is and the largest one is (in cardinality terms).
Suppose
Probability Measure
A probability measure
is -additive: is are pairwise disjoint events ( for ), then
Properties
Some properties of probability measures
- For
, - For
, if then - For
, - For
, if then
Conditional Probability
Let
Two events
Law of Total Probability
Theorem (Law of Total Probability)
Let
Bayes Theorem
Theorem (Bayes Theorem)
Let
For a single event
Random Variables
Definition
A random variable
The measurability condition states that the inverse image is a measurable set of
i.e. . This is essential since probabilities are defined only on .
In words, a random variable it’s a mapping from events to real numbers such that each interval on the real line can be mapped back into an element of the sigma algebra (it can be the empty set).
Distribution Function
Let
Properties
is monotone non-decreasing is right continuous and
The random variables
Density Function
Let
Moments
Expected Value
The expected value of a random variable, when it exists, is given by
The empirical expectation (or sample average) is given by
Variance and Covariance
The covariance of two random variables
The variance of a random variable
Properties
Let
Let
If
Note that the converse does not hold:
.
Sample Variance
The sample variance is given by
Finite Sample Bias Theorem
Theorem: The expected sample variance
Proof:
Inequalities
-
Triangle Inequality: if
, then -
Markov’s Inequality: if
, then -
Chebyshev’s Inequality: if
, then -
Cauchy-Schwarz’s Inequality:
-
Minkowski Inequality:
-
Jensen’s Inequality: if
is concave (e.g. logarithmic function), then Similarly, if is convex (e.g. exponential function), then
Law of Iterated Expectations
Theorem (Law of Iterated Expectations)
Law of Total Variance
Theorem (Law of Total Variance)
Since variances are always non-negative, the law of total variance
implies
Distributions
Normal Distribution
We say that a random variable
Multinomial Normal Distribution
We say that the k -vector Z has a multivariate standard normal
distribution, written
Properties
- The expectation and covariance matrix of
are $\mathbb E = \mu Var =\Sigma$. - If
are multivariate normal, and are uncorrelated if and only if they are independent. - If
and , then . - If
, then , chi-square with degrees of freedom. - If
with , then where . - If
and are independent then , student t with k degrees of freedom.
Normal Distribution Relatives
These distributions are relatives of the normal distribution
where
The
distribution is approximately standard normal but has heavier tails. The approximation is good for :