# Alec's sample mean bound

This page is a stub, so it contains little or minimal information and is on a to-do list for being expanded.The message provided is:
Needs some work, like what is a random variable for which expectation and variance are defined? Can we have complex or vector ones for example?
[ilmath]\newcommand{\P}[]{\mathbb{P}#1{\left[{#2}\right]} } \newcommand{\Pcond}[]{\mathbb{P}#1{\left[{#2}\!\ \middle\vert\!\ {#3}\right]} } \newcommand{\Plcond}[]{\Pcond[#1]{#2}{#3} } \newcommand{\Prcond}[]{\Pcond[#1]{#2}{#3} }[/ilmath]
[ilmath]\newcommand{\E}{ {\mathbb{E}{\left[{#1}\right]} } } [/ilmath][ilmath]\newcommand{\Mdm}{\text{Mdm}{\left({#1}\right) } } [/ilmath][ilmath]\newcommand{\Var}{\text{Var}{\left({#1}\right) } } [/ilmath][ilmath]\newcommand{\ncr}{ \vphantom{C}^{#1}\!C_{#2} } [/ilmath]

## Notice

It appears that this is actually 3 inequalities in one, which we shall name as follows:

1. Alec's remaining probability bound - that for [ilmath]X[/ilmath] a real and non-negative random variable, $\forall \alpha\in\mathbb{R}_{>0}\left[\P{X\ge \alpha}\le\frac{\E{X} }{\alpha}\right]$
2. Alec's deviation probability bound - that for a real random variable (possibly negative) [ilmath]X[/ilmath] that $\forall\beta\in\mathbb{R}_{>0}\left[\mathbb{P}\Big[\big\vert X-\E{X}\big\vert\ge\beta\Big]\le\frac{\Var{X} }{\beta^2} \right]$

## Inequality

Let [ilmath]X_1,\ldots,X_n[/ilmath] be a collection of [ilmath]n[/ilmath] random variables which are pairwise independent, such that:

• [ilmath]\exists\mu\forall i\in\{1,\ldots,n\}\big[\mathbb{E}[X_i]\eq\mu\big][/ilmath] - all of the [ilmath]X_i[/ilmath] have the same expectation and
• Alternatively: [ilmath]\forall i,j\in\{1,\ldots,n\}\big[\mathbb{E}[X_i]\eq\mathbb{E}[X_j]\big][/ilmath], but note we need [ilmath]\mu[/ilmath] in the expression
• [ilmath]\exists\sigma\forall i\in\{1,\ldots,n\}\big[\text{Var}(X_i)\eq\sigma^2\big][/ilmath] - all the [ilmath]X_i[/ilmath] have the same variance
• Alternatively: [ilmath]\forall i,j\in\{1,\ldots,n\}\big[\text{Var}(X_i)\eq\text{Var}(X_j)\big][/ilmath], but note again we need [ilmath]\sigma[/ilmath] in the expression

Then

• For all [ilmath]\epsilon>0[/ilmath] we have:
• $\mathbb{P}\left[\left\vert\frac{\sum^n_{i\eq 1}X_i}{n}-\mu\right\vert<\epsilon\right]\ge 1-\frac{\sigma^2}{\epsilon^2n}$
• Note that the notation here differs from that in my 2011 research journal slightly, but [ilmath]\sigma[/ilmath] and [ilmath]\mu[/ilmath] were present.
• $\mathbb{P}\left[\left\vert\frac{\sum^n_{i\eq 1}X_i}{n}-\mu\right\vert<\epsilon\right]\ge 1-\frac{\sigma^2}{\epsilon^2n}$

## History of form

When I "discovered" this inequality I was looking to say something like "the chance of the sample mean being within so much of the mean is at least ..."

I didn't know how to handle [ilmath]\vert X-\E{X}\vert[/ilmath] (what we'd now call [ilmath]\Mdm{X} [/ilmath]) which is why I applied it to variance, and of course [ilmath]\big(X-\E{X}\big)^2\ge 0[/ilmath] - the only condition required for the first inequality.