A random variable is a mapping, $X: \Omega \mapsto \mathbb{R}$, that assigns a real number $X(\omega)$ to each outcome $\omega$.
The cumulative distribution function, or CDF, is the function $F_X(x) = P(X \leq x)$.
Properites of a CDF:
X is discrete if it takes countably many values $\{x_1, x_2, ...\}$. We define the probablity function or probability mass function for X by $f_X(x) = P(X=x)$.
A random variable X is continuous if there exists a function $f_X$ such that
The function $f_X$ is called the probablity density function (PDF).
$F_X(x) = \int_{-\infty}^x f_X(t)dt$
$f_X(x) = F_X^{'}(x)$ at all points x at which $F_X$ is differentiable.
Let F be the CDF for a random variable X. Then:
Let X be a random variable with CDF F. The inverse CDF or quantile function is defined by $F^{-1}(q) = inf\{x: F(x) > q\}$ for $q \in [0,1]$.
$F^{-1}(1/4)$ is the first quartile. $F^{-1}(1/2)$ is the median.
X ~ F means "X has distribution F".
X has a standard normal distribution if $\mu=0$ and $\sigma=1$. The standard normal is traditionally denoted by Z, the PDF by $\phi(Z)$ and the CDF by $\Phi(Z)$.
If $X \sim N(\mu, \sigma^2)$, then
$P(a < X < b) = P(\cfrac{a-\mu}{\sigma} < Z < \cfrac{b-\mu}{\sigma}) = \Phi(\cfrac{b-\mu}{\sigma}) - \Phi(\cfrac{a-\mu}{\sigma})$
(There is no closed form expression for $\Phi$.)
Join mass function of a pair of discrete random variables X and Y, f(x,y) = P(X=x and Y=y) = P(X=x, Y=y)
In the continuous case, a function f(x,y) is a PDF of the random variables (X,Y) if
In the discrete or continuous case we define the join CDF as
$F_{X,Y}(x,y) = P(X \leq x, Y \leq y)$
If (X,Y) have a joint distribution with mass function $f_{X,Y}$, then the marginal mass function for X is defined by
$f_X(x) = P(X=x) = \sum_y P(X=x,Y=y) = \sum_y f(x,y)$
Similarly the marginal mass function for Y is defined by
$f_Y(y) = P(Y=y) = \sum_x P(X=x,Y=y) = \sum_x f(x,y)$
And for continuous random variables, the marginal densities are
$f_X(x) = \int f(x,y)dy$
$f_Y(y) = \int f(x,y)dx$
Two random variables X and Y are independent if, for every A and B,
$P(X \in A,Y \in B) = P(X \in A)P(Y \in B)$
and we write $X \coprod Y$
Let X and Y have join PDF $f_{X,Y}$. then $X \coprod Y$ iff $f_{X,Y}(x,y) = f_X(x)f_Y(y)$ for all values of x and y.
Suppose the range of X and Y is a (possibly infinite) rectangle. If f(x,y)=g(x)h(y) for some functions g and h (not necessarily probablity density functions) then X and Y are independent.
The conditional probability mass function is
$f_{X|Y}(x|y) = P(X=x|Y=y) = \cfrac{P(X=x,Y=y)}{P(Y=y)} = \cfrac{f_{X,Y}(x,y)}{f_Y(y)}$
For continuous random variables, the conditional probability density function is
$f_{X|Y}(x|y) = \cfrac{f_{X,Y}(x,y)}{f_Y(y)}$
assuming $f_Y(y) > 0$. Then
$P(X \in A|Y=y) = \int_A f_{X|Y}(x|y)dx$
In [ ]: