Conditional distributions help us formalize our intuitive ideas about whether two random variables are independent of each other. Let $X$ and $Y$ be two random variables, and suppose we are given the value of $X$. Does that change our opinion about $Y$? If the answer is yes, then we will say that $X$ and $Y$ are dependent. If the answer is no regardless of the given value of $X$, then we will say that $X$ and $Y$ are independent.

Let's start with some examples and then move to precise definitions and results.

```
dist1
```

You can see at once that if $X = 3$ then $Y$ can only be 0, whereas if $X = 2$ then $Y$ can be either 0 or 1. Knowing the value of $X$ changes the distribution of $Y$. That's dependence.

Here is an example in which you can't quickly determine dependence or independence by just looking at the possible values.

```
dist2
```

But you can tell by looking at the conditional distributions of $X$ given $Y$. Two of them are the same, but the third is different. Knowing the value of $Y$ affects the chances for $X$.

```
dist2.conditional_dist('X', 'Y')
```

It follows (and you should try to prove this), that at least some of the conditional distributions of $Y$ given the different values of $X$ will also be different from each other and from the marginal of $Y$.

In this example, all three conditional distributions of $Y$ given the three different values of $X$ are different from each other.

```
dist2.conditional_dist('Y', 'X')
```

Here is a joint distribution table in which you can't immediately tell whether there is dependence.

```
dist3
```

But look what happens when you condition $X$ on $Y$.

```
dist3.conditional_dist('X', 'Y')
```

All the rows are the same. That is, all the conditional distributions of $X$ given different values of $Y$ are the same, and hence are the same as the marginal of $X$ too.

Given the value of $Y$, the probabilities for $X$ don't change at all. That's independence.

You could have drawn the same conclusion by conditioning $Y$ on $X$:

```
dist3.conditional_dist('Y', 'X')
```

### Independence of Two Events

The concept of independence seems intuitive, but it is possible to run into trouble by not being careful about its definition. So let's define it formally.

There are two equivalent definitions of the independence of two events. The first encapsulates the main idea of independence, and the second is useful for calculation.

Two events $A$ and $B$ are *independent* if $P(B \mid A) = P(B)$. Equivalently, $A$ and $B$ are independent if $P(AB) = P(A)P(B)$.

### Independence of Two Random Variables

What we have observed in the examples of this section can be turned into a formal definition of independence.

Two random variables $X$ and $Y$ are *independent* if for every value $x$ of $X$ and $y$ of $Y$,

That is, no matter what the given $x$ is, the conditional distribution of $Y$ given $X=x$ is the same as if we didn't know that $X=x$.

Equivalently (this needs a proof, which consists of a routine application of definitions), for every $y$ the conditional distribution of $X$ given $Y=y$ is the same as if we didn't know that $Y=y$.

An equivalent definition in terms of the independence of events is that for any values of $x$ and $y$, the events $\{ X=x\}$ and $\{Y=y\}$ are independent.

That is, $X$ and $Y$ are independent if for any values $x$ of $X$ and $y$ of $Y$,

$$ P(X = x, Y = y) ~ = ~ P(X=x)P(Y=y) $$Independence simplifies the conditional probabilities in the multiplication rule.

It is a fact that if $X$ and $Y$ are independent random variables, then any event determined by $X$ is independent of any event determined by $Y$. For example, if $X$ and $Y$ are independent and $x$ is a number, then $\{X=x\}$ is independent of $\{Y>x\}$. Also, any function of $X$ is independent of any function of $Y$.

You can prove these facts by partitioning and then using the definition of independence. The proofs are routine but somewhat labor intensive. You are welcome to just accept the facts if you don't want to prove them.

### Mutual Independence

Events $A_1, A_2, \ldots A_n$ are *mutually independent* (or *independent* for short) if given that any subset of the events has occurred, the conditional chances of all other subsets remain unchanged.

That's quite a mouthful. In practical terms it means that it doesn't matter which of the events you know have happened; chances involving the remaining events are unchanged.

In terms of random variables, $X_1, X_2, \ldots , X_n$ are independent if given the values of any subset, chances of events determined by the remaining variables are unchanged.

In practice, this just formalizes statements such as "results of different tosses of a coin are independent" or "draws made at random with replacement are independent".

Try not to become inhibited by the formalism. Notice how the theory not only supports intuition but also develops it. You can expect your probabilistic intuition to be much sharper at the end of this course than it is now!

### IID Random Variables

If random variables are mutually independent and identically distributed, they are called "i.i.d." That's one of the most famous acronyms in probability theory. You can think of i.i.d. random variables as draws with replacement from a population, or as the results of independent replications of the same experiment.

Calculations involving i.i.d. random variables are often straightforward. For example, suppose the distribution of $X$ is given by

$$ P(X = i) = p_i, ~~~ i = 1, 2, \ldots, n $$where $\sum_{i=1}^n p_i = 1$. Now let $X$ and $Y$ be i.i.d. What is $P(X = Y)$? We'll answer this question by using the fundamental method, now in random variable notation.

$$ \begin{align*} P(X = Y) ~ &= ~ \sum_{i=1}^n P(X = i, Y = i) ~~~ \text{(partitioning)} \\ &= ~ \sum_{i=1}^n P(X = i)P(Y = i) ~~~ \text{(independence)} \\ &= ~ \sum_{i=1}^n p_i \cdot p_i ~~~ \text{(identical distributions)} \\ &= ~ \sum_{i=1}^n p_i^2 \end{align*} $$The last expression is easy to calculate if you know the numerical values of all the $p_i$.