Aalto Dictionary of ML – Markov's inequality

Consider a real-valued non-negative random variable (RV) $x$ for which the expectation $\mathbb{E} { x}$ exists. Markov’s inequality provides an upper bound on the probability $\mathbb{P}\left(x\geq a\right)$ that $x$ exceeds a given positive threshold $a>0$. In particular, \(\mathbb{P}\left(x \geq a\right) \leq \frac{\mathbb{E} \{ x\}}{a} \qquad \mbox{ holds for any } a > 0. \label{eq:markovsinequality_dict}\)
This inequality can be verified by noting that $\mathbb{P}\left(x \geq a\right)$ is the expectation $\mathbb{E} {g(x)}$ with the function \(g: \mathbb{R} \rightarrow \mathbb{R}: x' \mapsto \mathbb{I}_{\{x \geq a\}}(x').\)
As illustrated in the Figure below, for any positive $a>0$, \(g(x') \leq x'/a \mbox{ for all } x' \in \mathbb{R}.\) This obvious inequality implies Markov’s inequality via the monotonicity of the Lebesgue integral (Folland 1999, 50).

Illustration of Markov's inequality


See also: expectation, probability, concentration inequality.

Folland, Gerald B. 1999. Real Analysis: Modern Techniques and Their Applications. 2nd ed. New York, NY, USA: Wiley.


📚 This explanation is part of the Aalto Dictionary of Machine Learning — an open-access multi-lingual glossary developed at Aalto University to support accessible and precise communication in ML.

Written on November 9, 2025