Probability theory

Posted on July 26, 2017

probability theory, expectation, variance, definition of a random variable


This is mostly for my own review. I’ll be updating it periodically. Eventually I’d like it to be a one-stop shop for brushing up on theoretical probability.


Basics

A state space is a set of outcomes.

A random variable is a deterministic function from the state space to numbers.

A probability function assigns probabilities to those numbers


Expectation

Expectation is a number, the center of mass of a probability function. It’s a convex sum over all possible values of , each weighted by its probability .

The expectation of a constant is the same constant.

Expectation is linear.


Variance

Variance is a number, the weighted average distance of any from the mean . The weights are and the distance metric is squared , or rather, squared Euclidean distance.

The variance of a constant is zero.

Variance is not linear, but what follows is a useful property:

Sometimes we write for variance. The square root is called standard deviation


Covariance

Covariance is similar to variance. It’s the weighted average product of the distance of from its mean, with the distance of from its mean.

Covariance is symmetric.

The covariance of and a constant is zero.

The covariance of and a linear function of is slope times variance.


Correlation

Correlation is covariance normalized to lie between zero and one.

The correlation of and a linear function of is one.

Correlation is a measure of linear relationship. It does not capture non-linear effects. As such, independence is a stricter condition than zero correlation. Independence implies zero covariance, but zero covariance does not imply independence. However, non-zero covariance implies dependence, specifically, at least some linear dependence.


Estimation

The sample mean is unbiased.