You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Theorem 1.8 For a partition $B = {B_1, B_2, \cdots, B_m}$ and any event $A$ in the sample space, let $C_i = A \cap B_i$ For $i ≠ j$, the events $C_i$ and $C_j$ are mutually exclusive and $A = C_1 \cup C_2 \cup \cdots$
Theorem 1.9 For any event $A$ and partition ${B_1, B_2, \cdots, B_m}$
$$P[A] = \sum_{i=1}^m P[A \cap B_i$$
Theorem 1.10 Law of total probability
For a partition ${ B_1, B_2, \cdots, B_m }$ with $P[B_i] > 0$ for all $i$,
$$ P[A] = \sum_{i=1}^m P[A|B_i] P[B_i] $$
Theorem 1.11 Bayes' theorem
$$ P[B|A] = \frac{P[A|B] P[B]}{P[A]} $$
Definition 1.1 Outcome An outcome of an experiment is a possible result of the experiment.
Definition 1.2 Sample space The sample space of an experiment is the finest-grain, mutually exclusive, collectively exhaustive set of all possible outcomes of the experiment.
Definition 1.3 Event An event is a subset of the sample space.
Definition 1.4 Axioms of Probability A probability measure $P[.]$ is a function that maps events in the sample spacce to real numbers such that
Axiom 1 For any event $A$, $P[A] \geq 0$
Axiom 2$P[S] = 1$
Axiom 3 For any countable collection $A_1, A_2, \cdots$ of mutually exclusive events,
Theorem 2.1 An experiment consists of two subexperiments. If one subexperiment has $k$ outcomes and the other has $n$ outcomes, then the experiment has $kn$ outcomes.
Theorem 2.2 The number of k-permutations of $n$ distinguishable objects is
Theorem 2.4 Given $m$ distinguishable objects, there are $m^n$ ways to choose ith replacement an ordered sample of n objects.
Theorem 2.5 For $n$ repitions of a subexperiment with sample space $S_sub = {s_1, s_2, \cdots, s_m-1}$, the sample space $S$ of the sequential experiment has $m^n$ outcomes.
Theorem 2.6 The number of observation sequences for $n$ subexperiments with sample space $S = {0,1}$ with $0$ appearing $n_0$ times and $1$ appearing $n_1 = n - n_0$ times is $n \choose n_1$.
Theorem 2.7 For n reptitions of a subexperiment with sample space $S = {s_0, s_1, \cdots, s_m-1}$, the number of length $n = n_0 + n_1 + \cdots + n_{m-1}$ observation sequences with $s_i$ appearing $n_i$ times is
Theorem 2.9 A subexperiment has sample space $S = {s_0, s_1, \cdots, s_m-1}$ with $P[s_i] = p_i$ for $n = n_0 + n_1 + \cdots + n_{m-1}$ independent trials, the probability of $n_i$ occurrences of $s_i$, $i = 0, 1, \cdots, m-1$ is
$\text{(c) For all } x' > x, F_X(x') > F_X(x) $
$\text{(d) } F_X(x) = F_X(x_i) \text{for all x such that } x_i ≤ x ≤ x_{i+1} $
Theorem 3.3 For all $b > a$, $F_X(b) - F_X(a) = P[a < X ≤ b] $
Theorem 3.4 The Bernoulli $(p)$ random variable $X$ has expected value $E[X] = p$
Theorem 3.5 The geometirc $(p)$ random variable $X$ has expect value $E[X] = 1/p$
Theorem 3.6
(a) For the binomial $(n, p)$ random variable $X$ of Definition 3.6
$$ E[X] = np \space $$
(b) For the Pascal $(k, p)$ random variable $X$ of Definition 3.7
$$ E[X] = k/p $$
(c) For the discrete uniform $(k, l)$ random variable $X$ of Definition 3.8
$$ E[X] = \frac{k + l}{2} $$
Theorem 3.8 Perfom $n$ Bernoulli trials. In each trial, let the probability of success be ${\alpha} / n$, where ${\alpha} > 0$ is a constant and $n >\alpha.$ Let the random variable $K_n$ be the number of successes in the $n$ trials. As $n \rightarrow \infty, P_{K_n}(k)$ converges to the PMF of a Poisson $(\alpha)$ random variable.
Theorem 3.9 For a discrete random variable $X$, the PMF of $Y = g(X)$ is
$$ P_Y(y) = \sum_{x: g(x) = y} P_X(x) $$
Theorem 3.10 Given a random variable $X$ with PMF $P_X(x),$ and the derived random variable $Y = g(x),$ the expected value of $Y$ is
$$ E[Y] = \mu_Y = \sum_{x \in S_x} g(x) P_X(x) $$
Theorem 3.11 For any random variable $X$
$$ E[X - \mu_{X}] = 0 $$
Theorem 3.12 For any random variable $X$
$$ E[aX + b] = aE[X] + b $$
Theorem 3.13 In the absence of observations, the minimum mean square error estimate random variable $X$ is
(a) If X is Bernoiulli $(p)$, then $Var[X] = p(1-p)$
(b) If X is geometric $(p)$, then $Var[X] = ({1-p})/{p^2}$
(c) If X is binomial $(n, p)$, then $Var[X] = np(1 - p)$
(d) If X is Pascal $(k, p)$, then $Var[X] = k(1 - p)/p^2$
(e) If X is Poisson $(\alpha)$ then $Var[X] = \alpha$
(f) If X is discrete uniform (k, l), then $Var[X] = (l - k)(l - k + 2)/12$
Definition 3.1 Random Variable
A random variable consists of an experiment with a probability measure $P[.]$ defined on a sample space S and a function that assigns a real number to each outcome in the sample spacce of the experiment.
Definition 3.2 Discrete Random Variable$X$ is a discrete random variable if the range of $X$ is a countable set.
$$ S_X = { x_1, x_2, \dots } $$
Definition 3.3 Probability Mass Function PMF The probability mass function (PMF) of a discrete random variable $X$ is a function that assigns a probability to each value in the range of $X$
$$ P_X(x) = P[X = x] $$
Definition 3.4 Bernoulii (p) Random Variable$X$ is a Bernoulli $(p)$ random variable if the PMF of X has the form
where the parameter $\alpha$ is in the range $\alpha > 0$
Definition 3.10 Cumulative Distribution Function (CDF) The cumulative distribution function (CDF) of a discrete random variable $X$ is a function that assigns a probability to each value in the range of $X$.
$$ F_X(x) = P[X \leq x] $$
Definition 3.11 Mode A mode of random variable $X$ is a number $x_{mod}$ satisfying $P_X(x_{mod}) ≥ P_X(x)$ for all $x$
Definition 3.12 Median A median $x_{med}$ of random variable $X$ is a number that satisfies
Definition 3.13 Expected Value The expected value of $X$ is
$$ E[X] = \mu_{X} = \sum_{x \in S_X} x P_X(x) $$
Definition 3.14 Derived Random Variable Each sample value y of a derived random variable $Y$ is a mathematical function $g(x)$ of a sample value $x$ of another random variable $X$. We adopt the notation $Y = g(X)$ to describe the relationship of the two random variables.
Definition 3.15 Variance The variance of random variable $X$ is
$$ Var[X] = \sigma^2_X = E[(X - \mu{X})^2] $$
Definition 3.16 Standard Deviation The standard deviation of random variable $X$ is
$$ \sigma_X = \sqrt{Var[X]} $$
Definition 3.17 Moments For random variable $X$
(a) The nth moment is $E[X^n]$
(b) The nth central moment is $E[(X - \mu_X)^n]$
4. Continuous Random Variables
Theorem 4.1 For any random variable $X$,
(a) $F_X(-\infty) = 0$
(b) $F_X(\infty) = 1$
(c) $P[x_1 < X ≤ x_2] = F_X(x_2) - F_X(x_1)$
Theorem 4.2 For a continuous random variable $X$, with PMF $f_X(x)$,
(a) $f_X(x) ≥ 0 \text{ for all x, } $
(b) $f_X(x) = \int_{-\infty}^{x} f_X(u) \space du, $
Theorem 4.6 If $X$ is a uniform $(a, b)$ random variable,
The CDF of $X$ is
$$F_X(x) = \begin{cases}
0 & x < a \\
{(x - a)}/{(b - a)} & a ≤ x ≤ b \\
1 & x > b \\
\end{cases} $$
The expected value of $X$ is $E[X] = {(a + b)}/{2} $
The variabce of $X$ is $Var[X] = {(b - a)^2}/{12} $
Theorem 4.7 Let $X$ be a uniform $(a, b)$ random variable, where $a$ and $b$ are both integers. Let $K = \lceil X \rceil$. Then $K$ is a discrete uniform $(a + 1, b)$ random variable.
Theorem 4.8 If $X$ is an exponential $(\lambda)$ random variable,
The expected value of $X$ is $E[X] = {1}/{\lambda} $
The variance of $X$ is $Var[X] = {1}/{\lambda^2} $
Theorem 4.9 If $X$ is an exponential $(\lambda)$ random variable, then $K = \lceil X \rceil$ is a geometric $(p)$ random variable with $p = 1 - e^{-\lambda}$
Theorem 4.10 If $X$ is an Erlang $(n, \lambda)$ random variable, then
$\text{(a) } E[X] = {n\over{(\lambda)}} $
$\text{(b) } Var[X] = {n\over{(\lambda)^2}} $
Theorem 4.11 Let $K_\alpha$ denote a Poisson $\alpha$ random variable. For any $x > 0$, the CDF of an Erlang $(n, \lambda)$ random variable $X$ satisfies,
Theorem 4.17$\int_{-\infty}^{x} \delta(v) dv = u(x) $
Theorem 4.18 For a random variable $X$, we have the folloing equivalent statements:
$\text{(a) } P[X = x_0] = q$
$\text{(b) } P[x_0] = q$
$\text{(c) } F_X(x_{0}^+) - F_X(x_{0}^-) = q$
$\text{(d) } f_x(x_0) = q \delta(0)$
Definition 4.1 Cumulative Distribution Function (CDF) The cumulative distribution function (CDF) of random variable $X$ is $F_X(x) = P[X ≤ x]$
Definition 4.2 Continuous Random Variable$X$ is a continuous random variable if the CDF F_X(x) is a continuous function.
Definition 4.3 Probability Density Function (PDF) The probability density function (PDF) of a continuous random variable $X$ is
$$f_X(x) = \frac{dF_X(x)}{dx}$$
Definition 4.4 Expected Value The expected value of a random variable $X$ is
$$E[X] = \int_{-\infty}^{\infty} x f_X(x) dx$$
Definition 4.5 Uniform Random Variable$X$ is a uniform $(a, b)$ random variable if the PDF of $X$ is $f_X(x)$, and where the parameter $\lambda > 0$
$$f_X(x) = \begin{cases}
{1}/{(b - a)} & \space a ≤ x ≤ b \\
0 & \space otherwise \\
\end{cases} $$
Definition 4.6 Exponential Random Variable$X$ is an exponential (\lambda) random variable if the PDF of $X$ is $f_X(x)$, and where the parameter $\lambda > 0$
Definition 4.7 Erlang Random Variable$X$ is an Erlang $(n, \lambda)$ random variable if the PDF of $X$ is $f_X(x)$ where the parameter $\lambda > 0$, and the parameter $n ≥ 1$ is an integer.
Definition 4.8 Gaussian Random Variable$X$ is a Gaussian $(\mu, \sigma)$ random variable if the PDF of $X$ is $f_X(x)$ where the parameter $\mu$ can be any real number and the parameter $\sigma > 0$