17-Concentration_Inequalities_and_the_Laws_of_Large_Numbers

Attention

在阅读下面的内容前,您可能需要先修读 微积分 - 极限 部分内容以便更好理解下面的讲解

我们有时说一个东西的概率为 p ,但在 empirical experience(实践实验,与理论推理相对应) 中,我们要进行多少次才有足够的把握让实验概率 p^ 与 p 足够接近呢?下面的论述给出的答案:

不难看出,这与微积分中证明极限存在时取 n 足够大的情况是一样的。

对于上边最后的结论,我们下面将一步一步进行证明。

I Markov’s Inequality(马尔可夫不等式)

Theorem

(Markov’s Inequality). For a nonnegative random variable X (i.e., X(ω) ≥ 0 for all ω ∈ Ω) with finite mean, P[Xc]E[X]c , for any positive constant c.

proof is shown below:

Info

Indicator function

IA={1,if xA0,if xAorI{ϵ}={1,if ϵ is true0,if ϵ is false

II Chebyshev’s Inequality(切比雪夫不等式)

Theorem

(Chebyshev’s Inequality) For a random variable X with finite expectation E[X] = µ, P[|Xµ|c]Var(X)c2 , for any positive constant c.

The proof of Chebyshev's Inequality is easy since we just need to take

|Xμ|c|Xμ|2c2

using Markov’s Inequality,so we get that

P[|Xμ|c]=P[|Xμ|2c2]E[(Xμ)2]c2=Var(X)c2

take c=kσ where σ=Var(X) , we get that

P[|Xμ|kσ]1k2

which is of great importance.

III Estimating the Bias of a Coin

Now, let's solve the problem come up with at the begin.

IV Law of Large Numbers(大数定律)

Theorem

(Law of Large Numbers) . Let X1,X2,..., be a sequence of i.i.d. (independent and identically distributed) random variables with common finite expectation E[Xi ] = µ for all i. Then, their partial sums Sn = X1 +X2 +···+Xn satisfy

P[|1nSnμ|ϵ0]asn

for every ε > 0, however small.

That means if n is big enough, Snnμ .