Multidimensional Chebyshev's inequality
In probability theory, the multidimensional Chebyshev's inequality is a generalization of Chebyshev's inequality, which puts a bound on the probability of the event that a random variable differs from its expected value by more than a specified amount.
Let X be an N-dimensional random vector with expected value
and covariance matrix
If
is a positive-definite matrix, for any real number
:
Proof
Since
is positive-definite, so is
.
Define the random variable
Since
is positive, Markov's inequality holds:
Finally,
This article is issued from Wikipedia - version of the Saturday, February 15, 2014. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.
![V=\mathbb{E} \left[ \left(X - \mu \right) \left( X - \mu \right)^T \right]. \,](../I/m/56f872eea22164b78ba0889da601fc93.png)


![\begin{array}{lll}\mathrm{Pr}\left( \sqrt{\left( X-\mu\right)^T \, V^{-1} \, \left( X-\mu\right) } > t\right) &= \mathrm{Pr}\left( \sqrt{y} > t\right)\\
&=\mathrm{Pr}\left( y > t^2 \right) \\
&\le \frac{\mathbb{E}[y]}{t^2} .\end{array}](../I/m/0c83278af5c190f0f9688911222448a6.png)