Chebyshev inequality estimates the probability for exceeding the
deviation of a random variable from its mathematical expectation in terms
of the variance of the random variable. In modern probability theory, the
Chebyshev inequality is the most frequently used tool for proving
different convergence processes; for example, it plays a fundamental role
in proofs of various forms of laws of large numbers. The mathematical
expression of the bound on the probability in the Chebyshev inequality is
very simple and can be modified easily for different kinds of sequence of
random variables (e.g., for the case of sums of independent random
variables). This fact lies behind these frequent applications. In this
setting, the Chebyshev inequality has pure theoretical
“applications” in probability theory and its role is to
provide “a guarantee” of convergence but not to give a bound
on concrete probability content.
In the present article we consider the Chebyshev inequality as a
probability bound that is essential for the translation from its
conventional theoretical applications to the practical setting if
easy-to-compute multivariate generalizations are derived.
Such an inequality for the random vectors having multivariate Normal
distribution is proved. The new inequality gives a lower bound in terms of
variances on the probability that the random vector in question falls into
an Euclidean ball with center at mean vector. The need and importance of
consideration of this kind of multivariate Chebyshev inequality stemmed
from several problems in engineering and informational sciences (Hassibi
and Boyd [9], Jeng [10], Jeng and Woods [11], Molina, Katseggelos, Mateos, Hermoso, and
Segall [13]). Jeng [10] derived an inequality that gives an upper
bound for the probability in question. The simultaneous application of the
established multivariate Chebyshev inequality and Jeng's inequality
is useful in practical problems by providing lower and upper bounds on the
probability content.
The inequality is attractive by its being easy to compute and its
similarity to the original Chebyshev inequality, in contrast to well-known
complicated multivariate Chebyshev inequalities. The present article also
gives some insights into the very origin of the Chebyshev inequality,
which makes the article self-contained.