Skip to content

Prerequisites

Probability Basics

Probability is like quantum states, just like a particle can be in different states with different probabilities, we can have a random variable X that can take different values. For example:

X={0with probability 1p1with probability p

This is called a binary random variable (like spin up/down).

A concrete example: Let p=0.3, the probability of getting 1. So the probability of 0 is 1p=0.7. If X follows this distribution, we write:

X(0.7,0.3)

meaning p(X=0)=0.7 and p(X=1)=0.3.

This can be written as pX=(1p,p).

Expected value

Just like expectation values in QM, we can calculate the average/expected values:

E(X)=0×(1p)+1×p=p

For example with p=0.3:

E(X)=0.3

which means that, on average, X=0.3.

Independent events

When we have multiple random variables that don't affect each other, we call them independent and identically distributed (i.i.d.). When there is a statement like "Let Xi be i.i.d. binary random variables", this means that we are doing multiple independent trials of the same experiment.

Probability mass function (p.m.f.)

In probability theory, the probability mass function (p.m.f.) formally describes the probability distribution of a discrete random variable. For a random variable X that can take values in a set X, the p.m.f. is denoted as pX(x) and gives the probability that X takes the specific value x:

pX(x)=P(X=x)
  • X is the set of all possible values for the random variable X.
  • X is the random variable itself.
  • x is a specific outcome or value from the set X.

For example, if X is a binary random variable:

pX(0)=1ppX(1)=p

The p.m.f. provides a complete description of the probabilities associated with the possible values of the random variable.

Probability laws

There are some fundamental probability laws we need to be familiar with:

  • Axiom of Nonnegativity: 0pX(x)1 for all x in X.
  • Axiom of Unitarity: pX(x)=1, summing over all possible values of x.

These ensure that the p.m.f. defines a valid probability distribution.

Independence

When we have multiple random variables that don't affect each other, we call them independent. For independent variables X and Y:

P(X=x,Y=y)=P(X=x)P(Y=x)

This factorisation property is key for working with joint distributions.

Union bound

The union bound, also known as Boole's inequality, states that for any finite or countable set of events {A1,A2,...,An}, the probability of the union is less than or equal to the sum of their individual probabilities:

P(A1  A2  ...  An)iP(Ai)

In simpler terms, the probability of at least one event occurring is less than or equal to the sum of the probabilities of each event occurring individually. It's useful in information theory for error analysis, bounding failure probabilities in communication systems, analysing the probability of decoding errors and deriving achievability bounds. This bound is especially valuable because it's simple to apply, often gives useful results even though it's not always tight, and is valid regardless of whether the events are independent.

Markov's inequality

Markov's inequality states that for a non-negative random variable X and a positive real number a:

Pr(Xa)E(X)a

where Pr(Xa) is the probability that X is greater than or equal to a, and E(X) is the expected value of X.

This inequality is particularly useful because it provides an upper bound on the probability that a random variable takes values far from its mean. It also works for any non-negative random variable, regardless of its distribution, and it only requires knowledge of the expected value. For example, if we have a random variable with mean 10, the probability of it being 50 is at most 1050=0.2 or 20%.

In information theory, Markov's inequality is often used to analyse convergence properties of random processes, prove concentration inequalities, derive bounds on error probabilities in coding theorems and study the behavior of information measures.