There is nothing you cannot illustrate by drawing two circles.
Sample Space : just a set, usually denoted as . Elements of are called outcome. Certain nice subsets of are called events. (In practice, all subsets that you can cook up are 'nice'.)
Probability : given an event , is a real number, called the 'probability' that happens. It needs to satisfy the following conditions
I will call a probability measure on .
Independence : we say two events are independent, if .
Conditional Probability : suppose we want to know “given that happens, what is the probability will happen?”
The product rule
Bayes Formula Suppose you know and , then you can know
Random Variable A random variable is a (measurable) function .
The distribution of a random variable, is a probability measure on , such that given an interval , we define This is like we are pushing forward the probability measure on to .
Probability Density Suppose is a probability measure on , then sometimes we can find a function , such that In this case, we call the density of (with respect to ).
This generalizes to more than one variables.
Joint density, conditional density Suppose we have a probability density on , denoted as with the meaning Suppose we know that is near , we want to know that given this information, how is distributed, then we have This is a function of , we just need to 'renormalize' it so that the integral of over is . This gives
First thing first, find out what is and .
Be aware when someone say: “let me randomly choose ….”
Here is an interesting example: Bertrand Paradox
Another hard example: let be a random symmetric matrix of size , where each entry is iid Gausian. Question: how does eigenvalues of this matrix distribute?