User Tools

Site Tools


math121b:04-15

2020-04-15, Wednesday

Concepts

There is nothing you cannot illustrate by drawing two circles.

Sample Space : just a set, usually denoted as Ω\Omega. Elements of Ω\Omega are called outcome. Certain nice subsets of Ω\Omega are called events. (In practice, all subsets that you can cook up are 'nice'.)

Probability : given an event AΩA \In \Omega, P(A)\P(A) is a real number, called the 'probability' that AA happens. It needs to satisfy the following conditions

  • P(A)[0,1]\P(A) \in [0, 1], and P(Ω)=1,P()=0\P(\Omega) = 1, \P(\emptyset) = 0
  • If A1,A2A_1, A_2 are events, that are mutually exclusive, i.e. A1A2=A_1 \cap A_2 = \emptyset, then P(A1A2)=P(A1)+P(A2)\P(A_1 \cup A_2) = \P(A_1) + \P(A_2) . More generally, if you are given countably many mutually exclusive events A1,A2,A_1, A_2, \cdots, then P(iAi)=iP(Ai)\P(\cup_i A_i ) = \sum_i \P(A_i).

I will call P\P a probability measure on Ω\Omega.

Independence : we say two events A,BA, B are independent, if P(AB)=P(A)P(B)\P(A \cap B) = \P(A) \P(B).

Conditional Probability : suppose we want to know “given that AA happens, what is the probability BB will happen?” P(AB):=P(AB)P(B) \P(A | B) := \frac{ \P(A \cap B) } {\P(B) }

The product rule P(AB)=P(AB)P(B)=P(BA)P(A) \P(A \cap B) = \P(A | B) \P(B) = \P(B | A) \P(A)

Bayes Formula Suppose you know P(A),P(B)\P(A),\P(B) and P(BA)\P(B | A), then you can know P(AB)\P(A | B) P(AB)=P(BA)P(A)P(B). \P(A | B) = \frac{\P(B | A) \P(A)}{\P(B)}.

Random Variable A random variable is a (measurable) function X:ΩRX: \Omega \to \R.

The distribution of a random variable, is a probability measure on R\R, such that given an interval (a,b)R(a,b) \In \R, we define PX((a,b))=P({ωΩX(ω)(a,b)}).\P_X( (a,b)) = \P( \{ \omega \in \Omega | X(\omega) \in (a,b) \}). This is like we are pushing forward the probability measure on Ω\Omega to R\R.

Probability Density Suppose P\P is a probability measure on R\R, then sometimes we can find a function ρ(x)\rho(x), such that P((a,b))=abρ(x)dx\P( (a,b) ) = \int_a^b \rho(x) dx In this case, we call ρ(x)\rho(x) the density of P\P (with respect to dxdx).

This generalizes to more than one variables.

Joint density, conditional density Suppose we have a probability density on R2\R^2, denoted as ρXY(x,y)\rho_{XY}(x,y) with the meaning P(X(a,b),Y(c,d))=x:aby:cdρXY(x,y)dxdy. \P( X \in (a,b), Y \in (c,d)) = \int_{x:a}^b \int_{y:c}^d \rho_{XY}(x,y) dx dy. Suppose we know that YY is near y0y_0, we want to know that given this information, how is XX distributed, then we have ρXY(xy=y0)ρXY(x,y0) \rho_{X|Y}(x|y=y_0) \propto \rho_{XY}(x, y_0) This is a function of xx, we just need to 'renormalize' it so that the integral of xx over R\R is 11. This gives ρXY(xy=y0)=ρXY(x,y0)RρXY(x,y0)dx. \rho_{X|Y}(x|y=y_0) = \frac{ \rho_{XY}(x, y_0)}{ \int_\R \rho_{XY}(x', y_0) dx' }.

How to play the probability game?

First thing first, find out what is Ω\Omega and P\P.

Be aware when someone say: “let me randomly choose ….”

Here is an interesting example: Bertrand Paradox

Another hard example: let MM be a random symmetric matrix of size N×NN \times N, where each entry is iid Gausian. Question: how does eigenvalues of this matrix distribute?

math121b/04-15.txt · Last modified: 2020/04/15 09:44 by pzhou