Alec's sample mean bound

From Maths
Revision as of 21:02, 11 January 2018 by Alec (Talk | contribs) (Started to separate out into multiple pages)

(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to: navigation, search
Stub grade: A*
This page is a stub
This page is a stub, so it contains little or minimal information and is on a to-do list for being expanded.The message provided is:
Needs some work, like what is a random variable for which expectation and variance are defined? Can we have complex or vector ones for example?
[ilmath]\newcommand{\P}[2][]{\mathbb{P}#1{\left[{#2}\right]} } \newcommand{\Pcond}[3][]{\mathbb{P}#1{\left[{#2}\!\ \middle\vert\!\ {#3}\right]} } \newcommand{\Plcond}[3][]{\Pcond[#1]{#2}{#3} } \newcommand{\Prcond}[3][]{\Pcond[#1]{#2}{#3} }[/ilmath]
[ilmath]\newcommand{\E}[1]{ {\mathbb{E}{\left[{#1}\right]} } } [/ilmath][ilmath]\newcommand{\Mdm}[1]{\text{Mdm}{\left({#1}\right) } } [/ilmath][ilmath]\newcommand{\Var}[1]{\text{Var}{\left({#1}\right) } } [/ilmath][ilmath]\newcommand{\ncr}[2]{ \vphantom{C}^{#1}\!C_{#2} } [/ilmath]

Notice

It appears that this is actually 3 inequalities in one, which we shall name as follows:

  1. Alec's remaining probability bound - that for [ilmath]X[/ilmath] a real and non-negative random variable, [math]\forall \alpha\in\mathbb{R}_{>0}\left[\P{X\ge \alpha}\le\frac{\E{X} }{\alpha}\right][/math]
  2. Alec's deviation probability bound - that for a real random variable (possibly negative) [ilmath]X[/ilmath] that [math]\forall\beta\in\mathbb{R}_{>0}\left[\mathbb{P}\Big[\big\vert X-\E{X}\big\vert\ge\beta\Big]\le\frac{\Var{X} }{\beta^2} \right][/math]
  3. Alec's sample mean bound (this page)

Inequality

Let [ilmath]X_1,\ldots,X_n[/ilmath] be a collection of [ilmath]n[/ilmath] random variables which are pairwise independent, such that:

  • [ilmath]\exists\mu\forall i\in\{1,\ldots,n\}\big[\mathbb{E}[X_i]\eq\mu\big][/ilmath] - all of the [ilmath]X_i[/ilmath] have the same expectation and
    • Alternatively: [ilmath]\forall i,j\in\{1,\ldots,n\}\big[\mathbb{E}[X_i]\eq\mathbb{E}[X_j]\big][/ilmath], but note we need [ilmath]\mu[/ilmath] in the expression
  • [ilmath]\exists\sigma\forall i\in\{1,\ldots,n\}\big[\text{Var}(X_i)\eq\sigma^2\big][/ilmath] - all the [ilmath]X_i[/ilmath] have the same variance
    • Alternatively: [ilmath]\forall i,j\in\{1,\ldots,n\}\big[\text{Var}(X_i)\eq\text{Var}(X_j)\big][/ilmath], but note again we need [ilmath]\sigma[/ilmath] in the expression

Then

  • For all [ilmath]\epsilon>0[/ilmath] we have:
    • [math]\mathbb{P}\left[\left\vert\frac{\sum^n_{i\eq 1}X_i}{n}-\mu\right\vert<\epsilon\right]\ge 1-\frac{\sigma^2}{\epsilon^2n}[/math]
      • Note that the notation here differs from that in my 2011 research journal slightly, but [ilmath]\sigma[/ilmath] and [ilmath]\mu[/ilmath] were present.
    • [math]\mathbb{P}\left[\left\vert\frac{\sum^n_{i\eq 1}X_i}{n}-\mu\right\vert<\epsilon\right]\ge 1-\frac{\sigma^2}{\epsilon^2n}[/math]

History of form

When I "discovered" this inequality I was looking to say something like "the chance of the sample mean being within so much of the mean is at least ..."

I didn't know how to handle [ilmath]\vert X-\E{X}\vert[/ilmath] (what we'd now call [ilmath]\Mdm{X} [/ilmath]) which is why I applied it to variance, and of course [ilmath]\big(X-\E{X}\big)^2\ge 0[/ilmath] - the only condition required for the first inequality.