Probability & Statistics for Engineers & Scientists
===
# 積分公式
## 分部積分
$$
∫udv=uv−∫vdu
$$
# Chapter 1
## Definition 1.1
>平均
Suppose that the observations in a sample are $x_1, x_2, . . . , x_n$. The **sample mean**, denoted by $\overline x$, is
$$
\overline x =\sum\limits_{i=1}^n\frac{x_i}{n}= \frac{x_1+ x_2+ . . . + x_n}{n}.
$$
## Definition 1.2
> 中位數
Given that the observations in a sample are $x_1, x_2, . . . , x_n$, arranged in **increasing order** of magnitude, the sample median is
$$
\tilde{x}=
\left\{
\begin{array}{l}
x_{(n+1)/2}, & \text{if n is odd}\\
\frac{1}{2}(x_{n/2}+x_{n/2+1}), & \text{if n is even}
\end{array}
\right.
$$
## Definition 1.3
> 變異數
The **sample variance**, denoted by $s^2$, is given by
$$
s^2=\sum\limits_{i=1}^n\frac{(x_i-\overline{x})^2}{n-1}.
$$
> 標準差
The **sample standard deviation**, denoted by $s$, is the positive square root of $s^2$, that is,
$$
s=\sqrt{s^2}.
$$
# Chapter 2
## Definition 2.1
The set of all possible outcomes of a statistical experiment is called the **sample space** and is represented by the symbol $S$.
## Definition 2.2
An **event** is a subset of a sample space.
## Definition 2.3
The **complement** of an event $A$ with respect to $S$ is the subset of all elements of $S$ that are not in $A$. We denote the complement of $A$ by the symbol $A'$.
## Definition 2.4
The **intersection** of two events $A$ and $B$, denoted by the symbol $A \cap B$, is the
event containing all elements that are common to $A$ and $B$.
## Definition 2.5
Two events $A$ and $B$ are **mutually exclusive**, or **disjoint**, if $A \cap B = \phi$, that
is, if $A$ and $B$ have no elements in common.
## Definition 2.6
The **union** of the two events $A$ and $B$, denoted by the symbol $A \cup B$, is the event
containing all the elements that belong to $A$ or $B$ or both.
## Definition 2.7
A **permutation** is an arrangement of all or part of a set of objects.
## Definition 2.8
For any non-negative integer $n$, $n!$, called "$n$ factorial," is defined as
$$
n!=n(n-1)...(2)(1),
$$
with special case $0! = 1$.
## Theorem 2.1
The number of permutations of $n$ objects is $n!$.
## Theorem 2.2
The number of permutations of $n$ distinct objects taken $r$ at a time is
$$
C_r^n=\frac{n!}{(n-r)!}
$$
## Theorem 2.3
The number of permutations of n objects arranged in a circle is $(n-1)!$.
## Theorem 2.4
The number of distinct permutations of n things of which $n_1$ are of one kind, $n_2$ of a second kind, . . . , $n_k$ of a $k$th kind is
$$
\frac{n!}{n_1!n_2!...n_k!}
$$
## Theorem 2.5
The number of ways of partitioning a set of $n$ objects into $r$ cells with $n_1$ elements in the first cell, $n_2$ elements in the second, and so forth, is
$$
\begin{pmatrix}
n\\
n_1, n_2, ...n_r
\end{pmatrix}
=\frac{n!}{n_1!n_2!...n_r!}
$$
where $n_1 + n_2 + · · · + n_r = n$.
## Theorem 2.6
The number of combinations of $n$ distinct objects taken $r$ at a time is
$$
\begin{pmatrix}
n\\
r
\end{pmatrix}
=\frac{n!}{r!(n-r)!}
$$
## Definition 2.9
The **probability** of an event $A$ is the sum of the weights of all sample points in $A$. Therefore,
$$
0 \le P(A) \le 1, ~~~P(\phi) = 0,~~~ \text {and}~~~ P(S) = 1.
$$
Furthermore, if $A_1 , A_2 , A_3 , . . .$ is a sequence of mutually exclusive events, then
$$
P(A_1 \cup A_2 \cup A_3 \cup· · ·) = P(A_1) + P(A_2) + P(A_3) + · · · .
$$
## Theorem 2.7
If $A$ and $B$ are two events, then
$$
P(A \cup B) = P(A) + P(B) − P(A \cap B).
$$
## Theorem 2.8
For three events A, B, and C,
$$
P(A \cup B \cup C) = P(A) + P(B) + P(C) − P(A \cap B) − P(A \cap C) − P(B \cap C) + P(A \cap B \cap C).
$$
## Theorem 2.9
If $A$ and $A'$ are complementary events, then
$$
P(A) + P(A') = 1.
$$
## Definition 2.10
The conditional probability of $B$, given $A$, denoted by $P(B|A)$, is defined by
$$
P(B|A) = \frac{P(A \cap B)}{P(A)} ,~~~ \text {provided}~~~ P(A) > 0.
$$
## Definition 2.11
Two events $A$ and $B$ are independent if and only if
$$
P(B|A) = P(B) ~~~\text {or}~~~ P(A|B) = P(A),
$$
assuming the existences of the conditional probabilities. Otherwise, $A$ and $B$ are **dependent**.
## Theorem 2.10
If in an experiment the events $A$ and $B$ can both occur, then
$$
P(A \cap B) = P(A)P(B|A),~~~ \text{provided}~~~ P(A) > 0.
$$
## Theorem 2.11
Two events $A$ and $B$ are independent if and only if
$$
P(A \cap B) = P(A)P(B).
$$
Therefore, to obtain the probability that two independent events will both occur, we simply find the product of their individual probabilities.
## Theorem 2.12
If, in an experiment, the events $A_1,A_2, . . . , A_k$ can occur, then
$$
P(A_1 \cap A_2 \cap· · · \cap A_k )
= P(A_1)P(A_2|A_1)P(A_3|A_1 \cap A_2 ) · · · P(A_k |A_1 \cap A_2 \cap· · · \cap A_{k−1} ).
$$
If the events $A_1,A_2, . . . , A_k$ are independent, then
$$
P(A_1 \cap A_2 \cap· · · \cap A_k )= P(A_1)P(A_2). . .P(A_k)
$$
## Definition 2.12
A collection of events ${\oldstyle{A}} = \{A_1, . . . , A_n \}$ are mutually independent if for any subset of ${\oldstyle{A}}$, $A_{i_1} , . . . , A_{i_k}$ , for $k \le n$, we have
$$
P(A_{i_1} \cap · · · \cap A_{i_k} )= P(A_{i_1})· · ·P(A_{i_k}).
$$
## Theorem 2.13
If the events $B_1,B_2, . . . , B_k$ constitute a partition of the sample space $S$ such that $P(B_i) \ne 0$ for $i = 1, 2, . . . , k$, then for any event $A$ of $S$,
$$
P(A) = \sum_{i=1}^kP(B_i \cap A) =\sum_{i=1}^kP(B_i)P(A|B_i).
$$
## Theorem 2.14
**(Bayes’ Rule)** If the events $B_1,B_2, . . . , B_k$ constitute a partition of the sample space $S$ such that $P(B_i) \ne 0$ for $i = 1, 2, . . . , k$, then for any event $A$ in $S$ such that $P(A) \ne 0$,
$$
P(B_r|A) = \frac{P(B_r \cap A)}{\sum_{i=1}^kP(B_i \cap A)}= \frac{P(B_r)P(A|B_r)}{\sum_{i=1}^kP(B_i)P(A|B_i)}~~~
\text{for}~~~ r = 1, 2, . . . , k.
$$
# Chapter 3
## Definition 3.1
A **random variable** is a function that associates a real number with each element in the sample space.
## Definition 3.2
If a sample space contains a finite number of possibilities or an unending sequence with as many elements as there are whole numbers, it is called a **discrete sample space**.
## Definition 3.3
If a sample space contains an infinite number of possibilities equal to the number of points on a line segment, it is called a **continuous sample space**.
## Definition 3.4
The set of ordered pairs $(x, f(x))$ is a **probability function**, **probability mass function**, or **probability distribution** of the discrete random variable $X$ if, for each possible outcome $x$,
1. $f(x) \ge 0,$
2. ${\sum\limits_{x}} f(x) = 1,$
3. $P(X = x) = f(x).$
## Definition 3.5
The **cumulative distribution function** $F(x)$ of a discrete random variable $X$ with probability distribution $f(x)$ is
$$
F(x) = P(X \le x) = {\sum\limits_{t\le x}} f(t), \text{ for $−∞<x<∞$}.
$$
## Definition 3.6
>機率密度函數
The function $f(x)$ is a **probability density function** (pdf) for the continuous random variable $X$, defined over the set of real numbers, if
1. $f(x) \ge 0$, for all $x \in R.$
2. $\int_{-∞}^∞f(x) dx = 1.$
3. $P(a < X < b) =\int_{a}^bf(x) dx.$
## Definition 3.7
>累積分佈函數
The **cumulative distribution function** $F(x)$ of a continuous random variable $X$ with density function $f(x)$ is
$$
F(x)=P(X \le x) =\int_{-∞}^xf(t) dt, \text{ for $−∞<x<∞$}.
$$
## Definition 3.8
>聯合機率分佈、機率質量函數
The function $f(x, y)$ is a **joint probability distribution** or **probability mass function** of the discrete random variables $X$ and $Y$ if
1. $f(x, y) \ge 0$ for all $(x, y),$
1. ${\sum\limits_{x}}{\sum\limits_{y}}f(x,y)=1,$
1. $P(X = x, Y = y) = f(x, y).$
For any region $A$ in the $xy$ plane, $P[(X, Y ) \in A] ={\sum}{\sum\limits_{A}}f(x, y)$.
## Definition 3.9
>聯合密度函數
The function $f(x, y)$ is a **joint density function** of the continuous random variables $X$ and $Y$ if
1. $f(x, y) \ge 0$ for all $(x, y),$
1. $\int_{-∞}^∞\int_{-∞}^∞f(x,y)dxdy=1,$
1. $P[(X, Y ) \in A] =\int\int_{A}f(x, y)dxdy$, for any region $A$ in the $xy$ plane.
## Definition 3.10
>邊際分佈
The **marginal distributions** of X alone and of Y alone are
$g(x)={\sum\limits_{y}}f(x,y)$ and $h(y)={\sum\limits_{x}}f(x,y)$, for the discrete case
$g(x)=\int_{-∞}^∞f(x,y)dy$ and $h(y)=\int_{-∞}^∞f(x,y)dx$, for the continuous case
## Definition 3.11
>條件分佈
Let $X$ and $Y$ be two random variables, discrete or continuous.
The **conditional distribution** of the random variable $Y$ given that $X = x$ is
$$
f(y|x) = \frac{f(x, y)} {g(x)} , \text {provided $g(x) > 0$.}
$$
the **conditional distribution** of $X$ given that $Y = y$ is
$$
f(x|y) = \frac{f(x, y)} {h(y)} , \text {provided $h(y) > 0$.}
$$
## Definition 3.12
>獨立事件
Let $X$ and $Y$ be two random variables, discrete or continuous, with joint probability distribution $f(x, y)$ and marginal distributions $g(x)$ and $h(y)$, respectively.
The random variables $X$ and $Y$ are said to be **statistically independent** if and
only if
$$
f(x, y) = g(x)h(y)
$$
for all $(x, y)$ within their range.
## Definition 3.13
Let $X_1,X_2, ... , X_n$ be $n$ random variables, discrete or continuous, with joint probability distribution $f(x_1, x_2, ... , x_n)$ and marginal distribution $f_1 (x_1 ), f_2 (x_2 ), ... , f_n (x_n )$, respectively. The random variables $X_1,X_2, ... , X_n$ are said to be mutually **statistically independent** if and only if
$$
f(x_1, x_2, ... , x_n) = f_1 (x_1 )f_2 (x_2 ) ... f_n (x_n )
$$
for all $(x_1, x_2, ... , x_n)$within their range.
# Chapter 4
## Definition 4.1
>平均、期望值
Let $X$ be a random variable with probability distribution $f(x)$. The **mean**, or **expected value**, of $X$ is
$$
μ = E(X) =
\left\{
\begin{array}{l}
\sum\limits_{x}xf(x), & \text{if X is discrete}\\
\int_{-∞}^∞xf(x)\mathrm{d}x, & \text{if X is continuous}
\end{array}
\right.
$$
## Theorem 4.1
Let $X$ be a random variable with probability distribution $f(x)$. The expected value of the random variable $g(X)$ is
$$
μ_{g(X)} = E[g(X)] =
\left\{
\begin{array}{l}
\sum\limits_{x} g(x)f(x), & \text{if X is discrete}\\
\int_{-∞}^∞g(x)f(x)\mathrm{d}x, & \text{if X is continuous}
\end{array}
\right.
$$
## Definition 4.2
Let $X$ and $Y$ be random variables with joint probability distribution $f(x, y)$. The mean, or expected value, of the random variable $g(X, Y )$ is
$$
μ_{g(X,Y )} = E[g(X, Y )] =
\left\{
\begin{array}{l}
\sum\limits_{x}\sum\limits_{y}g(x,y)f(x,y), & \text{if X and Y are discrete}\\
\int_{-∞}^∞\int_{-∞}^∞g(x,y)f(x,y)\mathrm{d}x\mathrm{d}y, & \text{if X and Y are continuous}
\end{array}
\right.
$$
## Definition 4.3
Let $X$ be a random variable with probability distribution $f(x)$ and mean $μ$. The
variance of $X$ is
$$
σ^2=E[(X-μ)^2]=\left\{
\begin{array}{l}
{\sum\limits_{x}}(x-μ)^2f(x), & \text{if X is discrete}\\
\int_{-∞}^∞(x-μ)^2f(x)\mathrm{d}x, & \text{if X is continuous}
\end{array}
\right.
$$
The positive square root of the variance, $σ$, is called the **standard deviation** of
$X$.
## Theorem 4.2
The variance of a random variable $X$ is
$$
σ^2=E(X^2)−μ^2.
$$
## Theorem 4.3
Let $X$ be a random variable with probability distribution $f(x)$. The variance of the random variable $g(X)$ is
$$
σ^2_{g(X)}=E\{[g(X)-μ_{g(X)}]^2\}=\left\{
\begin{array}{l}
{\sum\limits_{x}}[g(x)-μ_{g(X)}]^2f(x), & \text{if X is discrete}\\
\int_{-∞}^∞[g(x)-μ_{g(X)}]^2f(x)\mathrm{d}x, & \text{if X is continuous}
\end{array}
\right.
$$
## Definition 4.4
Let $X$ and $Y$ be random variables with joint probability distribution $f(x, y)$. The
covariance of $X$ and $Y$ is
$$
σ_{XY}=E[(X−μ_X)(Y−μ_Y)] =
\left\{
\begin{array}{l}
{\sum\limits_{x}}{\sum\limits_{y}}
(x − μ_X )(y − μ_y )f(x, y), & \text{if X and Y are discrete}\\
\int_{-∞}^∞\int_{-∞}^∞(x − μ_X )(y − μ_y )f(x, y)\mathrm{d}x\mathrm{d}y, & \text{if X and Y are continuous}\\
\end{array}
\right.
$$
## Theorem 4.4
The covariance of two random variables $X$ and $Y$ with means $μ_X$ and $μ_Y$ , respectively, is given by
$$
σ_{XY}=E(XY)−μ_Xμ_Y.
$$
## Definition 4.5
Let $X$ and $Y$ be random variables with covariance $σ_{XY}$ and standard deviations $σ_X$ and $σ_Y$, respectively. The correlation coefficient of $X$ and $Y$ is
$$
\rho_{XY} = \frac{σ_{XY}}{σ_{X}σ_{Y}}.
$$
## Theorem 4.5
If $a$ and $b$ are constants, then
$$
E(aX + b) = aE(X) + b.
$$
## Theorem 4.6
The expected value of the sum or di!erence of two or more functions of a random variable $X$ is the sum or difference of the expected values of the functions. That
is,
$$
E[g(X) ± h(X)] = E[g(X)] ± E[h(X)].
$$
## Theorem 4.7
The expected value of the sum or difference of two or more functions of the random variables $X$ and $Y$ is the sum or difference of the expected values of the functions.
That is,
$$
E[g(X, Y ) ± h(X, Y )] = E[g(X, Y )] ± E[h(X, Y )].
$$
## Theorem 4.8
Let $X$ and $Y$ be two independent random variables. Then
$$
E(XY ) = E(X)E(Y ).
$$
## Theorem 4.9
If $X$ and $Y$ are random variables with joint probability distribution $f(x, y)$ and $a$, $b$, and $c$ are constants, then
$$
σ^2_{aX+bY+c} = a^2σ^2_X + b^2σ^2_Y + 2abσ_{X Y} .
$$
## Theorem 4.10
**(Chebyshev’s Theorem)** The probability that any random variable $X$ will assume a value within $k$ standard deviations of the mean is at least $1 − 1/k^2$. That
is,
$$
P(μ − kσ < X < μ + kσ) \ge 1 −\frac{1}{k^2}.
$$
# Chapter 5
## Binomial Distribution
> 二項式分佈
A Bernoulli trial can result in a success with probability $p$ and a failure with probability $q = 1−p$. Then the probability distribution of the binomial random variable $X$, the number of successes in n independent trials, is
$$
b(x;n,p)=\pmatrix{n\\x}p^xq^{n-x}, ~~~ x=0,1,2,...,n.
$$
## Multinomial Distribution
> 多項分佈
If a given trial can result in the $k$ outcomes $E_1,E_2, . . . , E_k$ with probabilities $p_1, p_2, . . . , p_k$, then the probability distribution of the random variables $X_1,X_2 , . . . , X_k$ , representing the number of occurrences for $E_1,E_2, . . . , E_k$ in $n$ independent trials, is
$$
f(x_1,x_2,...,x_k;p_1,p_2,...,p_k,n)=\pmatrix{n\\x_1,x_2,...,x_k}p_1^{x_1}p_2^{x_2}...p_k^{x_k},
$$
with
$$
\sum_{i=1}^{k}x_i = n ~~~\text{and}~~~ \sum_{i=1}^{k}p_i = 1.
$$
## Hypergeometric Distribution
> 超幾何分佈
(x:抽到瑕疵的數量,N:總數量,n:抽查數量,k:總瑕疵數量)
The probability distribution of the hypergeometric random variable $X$, the number of successes in a random sample of size $n$ selected from $N$ items of which $k$ are labeled **success** and $N − k$ labeled **failure**, is
$$
h(x; N, n, k) =\frac{\pmatrix{k\\x}\pmatrix{N-k\\n-x}}{\pmatrix{N\\n}}, ~~~ \max\{{0, n − (N − k)}\} \le x \le \min\{{n, k}\}.
$$
## Multivariate Hypergeometric Distribution
> 多變量超幾何分佈
If $N$ items can be partitioned into the $k$ cells $A_1,A_2, ... , A_k$ with $a_1, a_2, ... , a_k$ elements, respectively, then the probability distribution of the random variables $X_1,X_2, ... , X_k$, representing the number of elements selected from $A_1,A_2, ... , A_k$ in a random sample of size $n$, is
$$
f(x_1,x_2,...,x_k;a_1,a_2,...,a_k,N,n)=\frac{\pmatrix{a_1\\x_1}\pmatrix{a_2\\x_2}...\pmatrix{a_k\\x_k}}{\pmatrix{N\\n}},
$$
with $\sum_{i=1}^k x_i=n$ and $\sum_{i=1}^k a_i=N.$
## Negative Binomial Distribution
> 負二項式分佈
If repeated independent trials can result in a success with probability $p$ and a failure with probability $q = 1 − p$, then the probability distribution of the random variable $X$, the number of the trial on which the $k$th success occurs, is
$$
b^*(x; k, p) =\pmatrix{x-1\\k-1}p^k q^{x−k}, ~~~ x= k, k + 1, k + 2, . . . .
$$
## Geometric Distribution
> 幾何分布
If repeated independent trials can result in a success with probability $p$ and a failure with probability $q = 1 − p$, then the probability distribution of the random variable $X$, the number of the trial on which the first success occurs, is
$$
g(x; p) = pq^{x−1}, ~~~ x= 1, 2, 3,...
$$
## Poisson Distribution
The probability distribution of the Poisson random variable $X$, representing the number of outcomes occurring in a given time interval or specified region denoted by $t$, is
$$
p(x;𝜆t) = \frac{e^{−𝜆t}(𝜆t)^x}{x!}, ~~~ x= 0, 1, 2, . . . ,
$$
where $𝜆$ is the average number of outcomes per unit time, distance, area, or volume and $e = 2.71828...$.
## Theorem 5.1
The mean and variance of the binomial distribution $b(x; n, p)$ are
$$
μ = np ~~~ \text{and} ~~~ \sigma^2 = npq.
$$
## Theorem 5.2
The mean and variance of the hypergeometric distribution $h(x; N, n, k)$ are
$$
μ = \frac{nk}{N} ~~~ \text{and} ~~~ \sigma^2 = \frac{N-n}{N-1} \cdot n \cdot \frac{k}{N}(1-\frac{k}{N}).
$$
## Theorem 5.3
The mean and variance of a random variable following the geometric distribution are
$$
μ = \frac{1}{p} ~~~ \text{and} ~~~ \sigma^2 = \frac{1-p}{p^2}.
$$
## Theorem 5.4
Both the mean and the variance of the Poisson distribution $p(x;𝜆t)$ are $𝜆t$.
## Theorem 5.5
Let $X$ be a binomial random variable with probability distribution $b(x; n, p)$. When $n\to \infty$, $p \to 0$, and $np \xrightarrow{n\to \infty} μ$ remains constant,
$$
b(x; n, p) \xrightarrow{n\to \infty} p(x;μ).
$$
# Chapter 6
## Uniform Distribution
The density function of the continuous uniform random variable $X$ on the interval $[~A, B~]$ is
$$
f(x; A,B) = \left\{
\begin{array}{l}
\array{
\frac{1}{B-A}, & A\le x \le B\\
0, & \text{elsewhere.}
}
\end{array}
\right.
$$
## Normal Distribution
The density of the normal random variable $X$, with mean $μ$ and variance $\sigma^2$, is
$$
n(x; μ, \sigma) =\frac{1}{\sqrt{2\pi}\sigma}e^{-\frac{1}{2\sigma^2}(x-μ)^2}, ~~~ -\infty \lt x \lt \infty,
$$
where $\pi = 3.14159 . . .$ and $e = 2.71828 . . .$ .
## Normal Approximation to the Binomial Distribution
Let $X$ be a binomial random variable with parameters $n$ and $p$. For large $n$, $X$ has approximately a normal distribution with $μ = np$ and $\sigma^2 = npq = np(1−p)$ and
$$
\eqalign{
P(X \le x) &= \sum_{k=0}^x b(k;n,p) \\
&\approx \text{area under normal curve to the left of $x + 0.5$} \\
&= P\pmatrix{Z \le \frac{x+0.5-np}{\sqrt{npq}}},
}
$$
and the approximation will be good if $np$ and $n(1−p)$ are greater than or equal to 5.
## Gamma Distribution
The continuous random variable $X$ has a gamma distribution, with parameters $\alpha$ and $\beta$, if its density function is given by
$$
f(x;\alpha,\beta) = \left\{
\begin{array}{l}
\array{
\frac{1}{\beta^{\alpha}\Gamma(\alpha)}x^{\alpha-1}e^{-x/\beta}, & x \gt 0\\
0, & \text{elsewhere.}
}
\end{array}
\right.
$$
where $\alpha > 0$ and $\beta > 0$.
## Exponential Distribution
The continuous random variable $X$ has an **exponential distribution**, with parameter $\beta$, if its density function is given by
$$
f(x;\beta) = \left\{
\begin{array}{l}
\array{
\frac{1}{\beta}e^{-x/\beta}, & x \gt 0\\
0, & \text{elsewhere.}
}
\end{array}
\right.
$$
where $\beta \gt 0.$
## Chi-Squared Distribution
The continuous random variable $X$ has a **chi-squared distribution**, with $v$ **degrees of freedom**, if its density function is given by
$$
f(x;v) = \left\{
\begin{array}{l}
\array{
\frac{1}{2^{v/2}\Gamma(v/2)}x^{v/2-1}e^{-x/2}, & x \gt 0\\
0, & \text{elsewhere.}
}
\end{array}
\right.
$$
where $v$ is a positive integer.
## Beta Distribution
The continuous random variable $X$ has a **beta distribution** with parameters $\alpha > 0$ and $\beta > 0$ if its density function is given by
$$
f(x) = \left\{
\begin{array}{l}
\array{
\frac{1}{B(\alpha,\beta)}x^{\alpha-1}(1-x)^{\beta-1}, & 0 \lt x \lt 1\\
0, & \text{elsewhere.}
}
\end{array}
\right.
$$
## Lognormal Distribution
The continuous random variable $X$ has a **lognormal distribution** if the random variable $Y = \ln(X)$ has a normal distribution with mean $μ$ and standard deviation $\sigma$. The resulting density function of $X$ is
$$
f(x;μ,\sigma) = \left\{
\begin{array}{l}
\array{
\frac{1}{\sqrt{2\pi}\sigma x}e^{-\frac{1}{2\sigma^2}[\ln(x)-μ]^2}, & x \ge 0\\
0, & x<0
}
\end{array}
\right.
$$
## Weibull Distribution
The continuous random variable $X$ has a **Weibull distribution**, with parameters $\alpha$ and $\beta$, if its density function is given by
$$
f(x;\alpha,\beta) = \left\{
\begin{array}{l}
\array{
\alpha\beta x^{\beta-1}e^{-\alpha x^{\beta}}, & x > 0\\
0, & \text{elsewhere.}
}
\end{array}
\right.
$$
where $\alpha > 0$ and $\beta > 0$.
## cdf for Weibull Distribution
The **cumulative distribution function for the Weibull distribution** is given by
$$
F(x)=1-e^{-\alpha x^{\beta}}, ~~~ \text{for} ~ x \ge 0,
$$
where $\alpha > 0$ and $\beta > 0$.
## Failure Rate for Weibull Distribution
The **failure rate** at time t for the Weibull distribution is given by
$$
Z(t)=\alpha\beta t^{\beta-1}, ~~~ t>0.
$$
## Theorem 6.1
The mean and variance of the uniform distribution are
$$
μ=\frac{A+B}{2} ~~~ \text{and} ~~~ \sigma^2 = \frac{(B-A)^2}{12}.
$$
## Theorem 6.2
The mean and variance of $n(x; μ,\sigma)$ are $μ$ and $\sigma^2$, respectively. Hence, the standard deviation is $\sigma$.
## Definition 6.1
The distribution of a normal random variable with mean 0 and variance 1 is called a **standard normal distribution**.
## Theorem 6.3
If $X$ is a binomial random variable with mean $μ = np$ and variance $\sigma^2 = npq$, then the limiting form of the distribution of
$$
Z=\frac{X-np}{\sqrt{npq}},
$$
as $n\to \infty$, is the standard normal distribution $n(z; 0, 1)$.
## Definition 6.2
The gamma function is defined by
$$
\Gamma(\alpha) = \int_0^\infty x^{\alpha-1}e^{-x} \mathrm{d} x , ~~~ \text{for} ~~~ \alpha >0,
$$
## Theorem 6.4
The mean and variance of the gamma distribution are
$$
μ = \alpha\beta ~~~ \text{and} ~~~ \sigma^2 = \alpha\beta^2 .
$$
## Theorem 6.5
The mean and variance of the chi-squared distribution are
$$
μ = v ~~~ \text{and} ~~~ \sigma^2 = 2v.
$$
## Definition 6.3
A beta function is defined by
$$
B(\alpha,\beta) = \int_0^1 x^{\alpha-1}(1-x)^{\beta-1} \mathrm{d} x = \frac{\Gamma(\alpha)\Gamma(\beta)}{\Gamma(\alpha+\beta)}, ~~~ \text{for} ~~~ \alpha,\beta >0,
$$
where $\Gamma(\alpha)$ is the gamma function.
## Theorem 6.6
The mean and variance of a beta distribution with parameters $\alpha$ and $\beta$ are
$$
μ = \frac{\alpha}{\alpha+\beta} ~~~ \text{and} ~~~ \sigma^2 = \frac{\alpha\beta}{(\alpha+\beta)^2(\alpha+\beta+1)},
$$
respectively.
## Theorem 6.7
The mean and variance of the lognormal distribution are
$$
μ = e^{μ+\sigma^2 /2} ~~~ \text{and} ~~~ \sigma^2 = e^{2μ+\sigma^2} (e^{\sigma^2} − 1).
$$
## Theorem 6.8
The mean and variance of the Weibull distribution are
$$
μ = \alpha^{-1/\beta} \Gamma\left(1 + \frac{1}{\beta}\right) \quad \text{and} \quad \sigma^2 = \alpha^{-2/\beta} \left\{ \Gamma\left(1 + \frac{2}{\beta}\right) - \left[ \Gamma\left(1 + \frac{1}{\beta}\right) \right]^2 \right\}.
$$
# Chapter 8
## Definition 8.1
A **population** consists of the totality of the observations with which we are concerned.
## Definition 8.2
A **sample** is a subset of a population.
## Definition 8.3
Let $X_1,X_2, . . . , X_n$ be n independent random variables, each having the same probability distribution $f(x)$. Define $X_1,X_2, . . . , X_n$ to be a **random sample** of size $n$ from the population $f(x)$ and write its joint probability distribution as
$$
f(x_1, x_2, . . . , x_n) = f(x_1)f(x_2) · · · f(x_n).
$$
## Definition 8.4
Any function of the random variables constituting a random sample is called a **statistic**.
## Theorem 8.1
If $S^2$ is the variance of a random sample of size $n$, we may write
$$
S^2 = \frac{1}{n(n-1)}= \left[ n\sum_{i=1}^nX_i^2 - \left(\sum_{i=1}^nX_i\right)^2 \right].
$$
## Definition 8.5
The probability distribution of a statistic is called a **sampling distribution**.
## Theorem 8.2
**Central Limit Theorem:** If $\bar X$ is the mean of a random sample of size $n$ taken from a population with mean $μ$ and finite variance $\sigma^2$, then the limiting form of the distribution of
$$
Z = \frac{\bar X - μ}{\sigma / \sqrt{n}},
$$
as $n\to \infty$, is the standard normal distribution $n(z; 0, 1)$.
## Theorem 8.3
If independent samples of size $n_1$ and $n_2$ are drawn at random from two populations, discrete or continuous, with means $μ_1$ and $μ_2$ and variances $\sigma^2_1$ and $\sigma_2^2$, respectively, then the sampling distribution of the di!erences of means, $\bar X_1$ − $\bar X_2$, is approximately normally distributed with mean and variance given by
$$
μ_{\bar X_1 − \bar X_2} = μ_1 - μ_2 ~~~\text{and}~~~ \sigma^2_{\bar X_1 − \bar X_2} = \frac{\sigma^2_1}{n_1} + \frac{\sigma^2_2}{n_2}.
$$
Hence,
$$
Z = \frac{(\bar X_1 − \bar X_2)-(μ_1 - μ_2)}{\sqrt{(\sigma^2_1/n_1)+(\sigma^2_2/n_2)}}
$$
is approximately a standard normal variable.
## Theorem 8.4
If $S^2$ is the variance of a random sample of size n taken from a normal population having the variance $\sigma^2$, then the statistic
$$
X^2 = \frac{(n-1)S^2}{\sigma^2} = \sum_{i=1}^n\frac{(X_i-\bar X)^2}{\sigma^2}
$$
has a chi-squared distribution with $v = n − 1$ degrees of freedom.
## Theorem 8.5
Let $Z$ be a standard normal random variable and $V$ a chi-squared random variable with $v$ degrees of freedom. If $Z$ and $V$ are independent, then the distribution of the random variable $T$, where
$$
T = \frac{Z}{\sqrt{V/v}},
$$
is given by the density function
$$
h(t) = \frac{\Gamma [(v+1)/2]}{\Gamma (v/2)\sqrt{\pi v}}\left(1+\frac{t^2}{v}\right)^{-(v+1)/2}, ~~~ -\infty < t < \infty.
$$
This is known as the **$t$-distribution** with $v$ degrees of freedom.
## Theorem 8.6
Let $U$ and $V$ be two independent random variables having chi-squared distributions with $v_1$ and $v_2$ degrees of freedom, respectively. Then the distribution of the random variable $F = \frac{U/v_1}{V/v_2}$ is given by the density function
$$
h(t) = \left\{
\begin{array}{l}
\array{
\frac{\Gamma [(v_1+v_2)/2](v_1/v_2)^{v_1/2}}{\Gamma (v_1/2)\Gamma (v_2/2)}\frac{f^{(v_1/2)-1}}{(1+v_1f/{v_2})^{(v_1+v_2)/2}}, & f>0\\
0, & f \le 0
}
\end{array}
\right.
$$
This is known as the **$F$-distribution** with $v_1$ and $v_2$ degrees of freedom (d.f.).
## Theorem 8.7
Writing $f_{\alpha} (v_1, v_2)$ for $f_{\alpha}$ with $v_1$ and $v_2$ degrees of freedom, we obtain
$$
f_{1-\alpha}(v_1,v_2) = \frac{1}{f_{\alpha} (v_2, v_1)}.
$$
## Theorem 8.8
If $S^2_1$ and $S^2_2$ are the variances of independent random samples of size $n_1$ and $n_2$ taken from normal populations with variances $\sigma^2_1$ and $\sigma^2_2$, respectively, then
$$
F = \frac{S^2_1/\sigma^2_1}{S^2_2/\sigma^2_2} = \frac{\sigma^2_2S^2_1}{\sigma^2_1S^2_2}
$$
has an $F$-distribution with $v_1 = n_1 − 1$ and $v_2 = n_2 − 1$ degrees of freedom.
## Definition 8.6
A **quantile** of a sample, $q(f)$, is a value for which a specified fraction $f$ of the data values is less than or equal to $q(f)$.
## Definition 8.7
The **normal quantile-quantile plot** is a plot of $y_{(i)}$ (ordered observations) against $q_{0,1}(f_i)$, where $f_i = \frac{i-\frac{3}{8}}{n+\frac{1}{4}}$.