当前位置:网站首页>Probability theory and mathematical statistics 3 discrete random variables and probability distributions (Part 2)
Probability theory and mathematical statistics 3 discrete random variables and probability distributions (Part 2)
2022-07-25 09:59:00 【Lum0s!】
probability theory _3.4_3.5_3.6
3.4 The Binomial Probability Distribution( The binomial distribution )
There are many experiments that conform either exactly or approximately to the following list of requirements( Fully or approximately meet the following requirements ):
- The experiment consists of a sequence of n smaller experiments called trials( test ), where n is fixed in advance of the experiment.
- Each trial can result in one of the same two possible outcomes (dichotomous trials, Dichotomy test ), which we generically denote by success (S) and failure (F).
- The trials are independent, so that the outcome on any particular trial does not influence the outcome on any other trial( The results of any particular test will not affect the results of any other test ).
- The probability of success P(S) is constant from trial to trial; we denote this probability by p.
DEFINITION:
An experiment for which Conditions 1–4 are satisfied is called a binomial experiment
We will use the following rule of thumb( Rule of thumb ) in deciding whether a "without replacement( Take it out and don't put it back )"experiment can be treated as a binomial experiment.
RULE:
Consider sampling without replacement from a dichotomous population of size N. If the sample size (number of trials) n is at most 5% of the population size, the experiment can be analyzed as though it were exactly a binomial experiment.
The Binomial Random Variable and Distribution
DEFINITION:
The binomial random variable X associated with a binomial experiment consisting of n trials is defined as
X = t h e n u m b e r o f S ′ s a m o n g t h e n t r i a l s X=the \hspace{1mm} number \hspace{1mm} of \hspace{1mm} S's \hspace{1mm} among \hspace{1mm} the \hspace{1mm} n \hspace{1mm}trials X=thenumberofS′samongthentrials
NOTATION:
Because the pmf of a binomial rv X depends on the two parameters n and p, we denote the pmf by b(x; n, p).
THEOREM:
b ( x ; n , p ) = { ( n x ) p x ( 1 − p ) n − x , x = 0 , 1 , 2 , . . . , n 0 , o t h e r w i s e b(x;n,p) =\begin{cases} \dbinom{n}{x}p^x(1-p)^{n-x}, x= 0,1,2,...,n \\0, otherwise \end{cases} b(x;n,p)=⎩⎨⎧(xn)px(1−p)n−x,x=0,1,2,...,n0,otherwise
Using Binomial Tables*
NOTATION:
For X~Bin(n,p), the cdf will be denoted by
B ( x ; n , p ) = P ( X ≤ x ) = ∑ y = 0 x b ( y ; n , p ) x = 0 , 1 , . . . , n B(x;n,p)=P(X \leq x) = \sum_{y=0}^x b(y;n,p) \hspace{1cm} x=0,1,...,n B(x;n,p)=P(X≤x)=y=0∑xb(y;n,p)x=0,1,...,n
The Mean and Variance of X
PROPOSITION:
If X~Bin(n,p), then E(X)=np, V(X)=np(1-p)=npq, and σ X \sigma_X σX= n p q \sqrt{npq} npq (where q = 1 - p).
3.5 Hypergeometric and Negative Binomial Distributions( Hypergeometric distribution and negative binomial distribution )
The Hypergeometric Distribution
The assumptions leading to the hypergeometric distribution are as follows:
- The population or set to be sampled consists of N individuals, objects, or elements (a finite population).
- Each individual can be characterized as a success (S) or a failure (F), and there are M successes in the population.
- A sample of n individuals is selected without replacement in such a way that each subset of size n is equally likely to be chosen.
PROPOSITION:
If X is the number of S’s in a completely random sample of size n drawn from a population consisting of M S’s and (N-M)F’s, then the probability distribution of X, called the hypergeometric distribution, is given by
P ( X = x ) = h ( x ; n , M , N ) = ( M x ) ( N − M n − x ) ( N n ) P(X=x)=h(x;n,M,N)=\frac{\dbinom{M}{x}\dbinom{N-M}{n-x}}{\dbinom{N}{n}} P(X=x)=h(x;n,M,N)=(nN)(xM)(n−xN−M)
for x, an integer, satisfying max(0,n-N+M) ≤ \leq ≤ x ≤ \leq ≤ min(n,M).
PROPOSITION:
The mean and variance of the hypergeometric rv X having pmf h(x; n, M, N) are
E ( X ) = n ⋅ M N V ( X ) = ( N − n N − 1 ) ⋅ n ⋅ M N ⋅ ( 1 − M N ) E(X)= n \cdot \frac{M}{N} \hspace{1cm} V(X)=(\frac{N-n}{N-1}) \cdot n \cdot \frac{M}{N}\cdot (1-\frac{M}{N}) E(X)=n⋅NMV(X)=(N−1N−n)⋅n⋅NM⋅(1−NM)
The means of the binomial and hypergeometric rv’s are equal, whereas the variances of the two rv’s differ by the factor ( N − n ) ( N − 1 ) \frac{(N-n)}{(N-1)} (N−1)(N−n), often called the finite population correction factor( Finite population correction factor ). This factor is less than 1, so the hypergeometric variable has smaller variance than does the binomial rv. The correction factor can be written ( 1 − n N ) ( 1 − 1 N ) \frac{(1-\frac{n}{N})}{(1-\frac{1}{N})} (1−N1)(1−Nn), which is approximately 1 when n is small relative to N.
The Negative Binomial Distribution
The negative binomial rv and distribution are based on an experiment satisfying the following conditions:
- The experiment consists of a sequence of independent trials.
- Each trial can result in either a success (S) or a failure (F).
- The probability of success is constant from trial to trial, so for i = 1,2,3,…
- The experiment continues (trials are performed) until a total of r successes have been observed, where r is a specified positive integer.
The random variable of interest is X = the number of failures that precede the rth success; X is called a negative binomial random variable because, in contrast to the binomial rv, the number of successes is fixed and the number of trials is random.
PROPOSITION:
The pmf of the negative binomial rv X with parameters r=number of S’s and p=P(S) is
n b ( x ; r , p ) = ( x + r − 1 r − 1 ) p r ( 1 − p ) x x = 0 , 1 , 2 , . . . nb(x;r,p)=\dbinom{x+r-1}{r-1}p^r(1-p)^x \hspace{1mm} x=0,1,2,... nb(x;r,p)=(r−1x+r−1)pr(1−p)xx=0,1,2,...
In some sources, the negative binomial rv is taken to be the number of trials X+r rather than the number of failures.
In the special case r=1, the pmf is
n b ( x ; 1 , p ) = ( 1 − p ) x p x = 0 , 1 , 2 , . . . nb(x;1,p)=(1-p)^xp \hspace{1cm}x=0,1,2,... nb(x;1,p)=(1−p)xpx=0,1,2,...
Both X=number of F’s and Y=number of trials (=1+X) are referred to in the literature as geometric random variables( Geometric random variable ), and the pmf above is called the geometric distribution( Geometric distribution ).
PROPOSITION:
If X is a negative binomial rv with pmf nb(x; r, p), then
E ( X ) = r ( 1 − p ) p V ( X ) = r ( 1 − p ) p 2 E(X)=\frac{r(1-p)}{p} \hspace{1cm} V(X)=\frac{r(1-p)}{p^2} E(X)=pr(1−p)V(X)=p2r(1−p)
3.6 The poisson Probability Distribution( Poisson distribution )
DEFINITION:
A discrete random variable X is said to have a Poisson distribution with parameter μ ( μ > 0 ) \mu(\mu > 0) μ(μ>0) if the pmf of X is
p ( x ; μ ) = e − μ ⋅ μ x x ! x = 0 , 1 , 2 , 3 , . . . p(x;\mu)=\frac{e^{-\mu} \cdot \mu^{x}}{x!} \hspace{1cm} x=0,1,2,3,... p(x;μ)=x!e−μ⋅μxx=0,1,2,3,...
μ \mu μ is in fact the expected value of X. The letter e in the pmf represents the base of the natural logarithm system; its numerical value is approximately 2.71828. In contrast to the binomial and hypergeometric distributions, the Poisson distribution spreads probability over all non-negative integers, an infinite number of possibilities.
e μ = 1 + μ + μ 2 2 ! + μ 3 3 ! + . . . = ∑ x = 0 ∞ μ x x ! e^{\mu}=1+\mu+\frac{\mu^2}{2!}+\frac{\mu^3}{3!}+...=\sum_{x=0}^{\infin} \frac{\mu^x}{x!} eμ=1+μ+2!μ2+3!μ3+...=x=0∑∞x!μx
If the two extreme terms are multiplied by and then this quantity is moved inside the summation on the far right, the result is
1 = ∑ x = 0 ∞ e − μ ⋅ μ x x ! 1=\sum_{x=0}^{\infin} \frac{e^{-\mu} \cdot \mu^x}{x!} 1=x=0∑∞x!e−μ⋅μx
The Poisson Distribution as a Limit
PROPOSITION:
Suppose that in the binomial pmf b(x; n, p), we let n → ∞ n \to \infin n→∞ and p → 0 p \to 0 p→0 in such a way that np approaches a value μ > 0 \mu > 0 μ>0 . Then b(x; n, p) → \to → p(x; μ \mu μ).
According to this proposition, in any binomial experiment in which n is large and p is small, b ( x ; n , p ) ≈ p ( x ; μ ) b(x;n,p) \approx p(x;\mu) b(x;n,p)≈p(x;μ) , where μ = n p \mu = np μ=np. As a rule of thumb, this approximation can safely be applied if n>50 and np<5.
The Mean and Variance of X
Since b ( x ; n , p ) → p ( x ; μ ) b(x;n,p) \to p(x;\mu) b(x;n,p)→p(x;μ) as n → ∞ n \to \infin n→∞ , p → 0 p \to 0 p→0 , n p → μ np \to \mu np→μ, the mean and variance of a
binomial variable should approach those of a Poisson variable. These limits are
n p → μ np \to \mu np→μ and n p ( 1 − p ) → μ np(1-p) \to \mu np(1−p)→μ.
PROPOSITION:
If X has a Poisson distribution with parameter μ \mu μ, then E ( X ) = V ( X ) = μ E(X) = V(X) = \mu E(X)=V(X)=μ.
The Poisson Process
Assumption:
- There exists a parameter α > 0 \alpha >0 α>0 such that for any short time interval of length Δ \Delta Δt, the probability that exactly one event occurs is α ⋅ Δ t + o ( Δ t ) ∗ \alpha \cdot \Delta t+o(\Delta t)^* α⋅Δt+o(Δt)∗.
- The probability of more than one event occurring during Δ t \Delta t Δt is o ( Δ t ) o(\Delta t) o(Δt) [which, along with Assumption 1, implies that the probability of no events during Δ t \Delta t Δt is 1 − α ⋅ Δ t − o ( Δ t ) 1-\alpha \cdot \Delta t - o(\Delta t) 1−α⋅Δt−o(Δt).
- The number of events occurring during the time interval Δ t \Delta t Δt is independent of the number that occur prior to this time interval.
Informally, Assumption 1 says that for a short interval of time, the probability of a single event occurring is approximately proportional to the length of the time interval, where a is the constant of proportionality. Now let P k ( t ) P_k(t) Pk(t) denote the probability that k events will be observed during any particular time interval of length t.
PROPOSITION:
P k ( t ) = e − α t ⋅ ( α t ) k / k ! P_k(t)=e{-\alpha t} \cdot (\alpha t)^k / k! Pk(t)=e−αt⋅(αt)k/k! ,so that the number of events during a time interval of length t is a Poisson rv with parameter μ = α t \mu=\alpha t μ=αt. The expected number of events during any such time interval is then α t \alpha t αt, so the expected number during a unit interval of time is α \alpha α.
The occurrence of events over time as described is called a Poisson process; the parameter α \alpha α specifies the rate for the process.
边栏推荐
- NLM5系列无线振弦传感采集仪的工作模式及休眠模式下状态
- SOC芯片内部结构
- 手持振弦VH501TC采集仪传感器的连接与数据读取
- 概率机器人学习笔记第二章
- ¥ 1-2 example 2.2 put the union of two sets into the linear table
- Creation of adjacency matrix of undirected connected graph output breadth depth traversal
- 低功耗和UPF介绍
- [Android studio] batch data import to Android local database
- TM1637带秒点四位LED显示器模块ARDUINO驱动程序
- SystemVerilog syntax
猜你喜欢

Store to-do items locally (improve on to-do items)

ARMV8体系结构简介

Mlx90640 infrared thermal imager temperature measurement module development notes (4)
![[RNN] analyze the RNN from rnn- (simple|lstm) to sequence generation, and then to seq2seq framework (encoder decoder, or seq2seq)](/img/6e/da80133e05b18c87d7167c023b6c93.gif)
[RNN] analyze the RNN from rnn- (simple|lstm) to sequence generation, and then to seq2seq framework (encoder decoder, or seq2seq)

关于MLOps中的数据工程,你一定要知道的.......

Get to know opencv4.x for the first time --- add Gaussian noise to the image

Mlx90640 infrared thermal imaging sensor temperature measurement module development notes (III)

Introduction to armv8 architecture

MLX90640 红外热成像传感器测温模块开发笔记(三)

单目深度估计自监督模型Featdepth解读(上)——论文理解和核心源码分析
随机推荐
ARMV8 datasheet学习
ADC简介
ISP图像信号处理
CDA LEVELⅠ2021新版模拟题二(附答案)
无线振弦采集仪参数配置工具的设置
Mlx90640 infrared thermal imager temperature measurement module development notes (V)
SOC芯片内部结构
数字IC设计SOC入门进阶
概率机器人学习笔记第二章
Introduction to arm GIC
MLX90640 红外热成像仪测温模块开发笔记(四)
Introduction to Verdi Foundation
概率论与数理统计 4 Continuous Random Variables and Probability Distributions(连续随机变量与概率分布)(上篇)
Swift creates weather app
ARMV8体系结构简介
【机器翻译】SCONES——用多标签任务做机器翻译
多通道振弦、温度、模拟传感信号采集仪数据查看和参数修改
testbench简介
¥ 1-2 example 2.2 put the union of two sets into the linear table
VCS常用命令