Conditional distributions help us formalize our intuitive ideas about whether two random variables are independent of each other. Let and be two random variables, and suppose we are given the value of . Does that change our opinion about ? If the answer is yes, then we will say that and are dependent. If the answer is no regardless of the given value of , then we will say that and are independent.
Let’s start with some examples and then move to precise definitions and results.
4.5.1Dependence¶
Here is the joint distribution of two random variables and . From this, what can we say about whether and are dependent or independent?
dist1You can see at once that if then can only be 0, whereas if then can be either 0 or 1. Knowing the value of changes the distribution of . That’s dependence.
Here is an example in which you can’t quickly determine dependence or independence by just looking at the possible values.
dist2But you can tell by looking at the conditional distributions of given . Two of them are the same, but the third is different. Knowing the value of affects the chances for .
dist2.conditional_dist('X', 'Y')It follows (and you should try to prove this), that at least some of the conditional distributions of given the different values of will also be different from each other and from the marginal of .
In this example, all three conditional distributions of given the three different values of are different from each other.
dist2.conditional_dist('Y', 'X')4.5.2Independence¶
Here is a joint distribution table in which you can’t immediately tell whether there is dependence.
dist3But look what happens when you condition on .
dist3.conditional_dist('X', 'Y')All the rows are the same. That is, all the conditional distributions of given different values of are the same, and hence are the same as the marginal of too.
Given the value of , the probabilities for don’t change at all. That’s independence.
You could have drawn the same conclusion by conditioning on :
dist3.conditional_dist('Y', 'X')🎥 See More
4.5.3Independence of Two Events¶
The concept of independence seems intuitive, but it is possible to run into trouble by not being careful about its definition. So let’s define it formally.
There are two equivalent definitions of the independence of two events. The first encapsulates the main idea of independence, and the second is useful for calculation.
Two events and are independent if . Equivalently, and are independent if .
4.5.4Independence of Two Random Variables¶
What we have observed in the examples of this section can be turned into a formal definition of independence.
Two random variables and are independent if for every value of and of ,
That is, no matter what the given is, the conditional distribution of given is the same as if we didn’t know that .
Equivalently (this needs a proof, which consists of a routine application of definitions), for every the conditional distribution of given is the same as if we didn’t know that .
An equivalent definition in terms of the independence of events is that for any values of and , the events and are independent.
That is, and are independent if for any values of and of ,
Independence simplifies the conditional probabilities in the multiplication rule.
It is a fact that if and are independent random variables, then any event determined by is independent of any event determined by . For example, if and are independent and is a number, then is independent of . Also, any function of is independent of any function of .
You can prove these facts by partitioning and then using the definition of independence. The proofs are routine but somewhat labor intensive. You are welcome to just accept the facts if you don’t want to prove them.
4.5.5Mutual Independence¶
Events are mutually independent (or independent for short) if given that any subset of the events has occurred, the conditional chances of all other subsets remain unchanged.
That’s quite a mouthful. In practical terms it means that it doesn’t matter which of the events you know have happened; chances involving the remaining events are unchanged.
In terms of random variables, are independent if given the values of any subset, chances of events determined by the remaining variables are unchanged.
In practice, this just formalizes statements such as “results of different tosses of a coin are independent” or “draws made at random with replacement are independent”.
Try not to become inhibited by the formalism. Notice how the theory not only supports intuition but also develops it. You can expect your probabilistic intuition to be much sharper at the end of this course than it is now!
4.5.6IID Random Variables¶
If random variables are mutually independent and identically distributed, they are called “i.i.d.” That’s one of the most famous acronyms in probability theory. You can think of i.i.d. random variables as draws with replacement from a population, or as the results of independent replications of the same experiment.
Calculations involving i.i.d. random variables are often straightforward. For example, suppose the distribution of is given by
where . Now let and be i.i.d. What is ? We’ll answer this question by using the fundamental method, now in random variable notation.
The last expression is easy to calculate if you know the numerical values of all the .