Chebyshev's inequality: Difference between revisions
CSV import Tags: mobile edit mobile web edit |
CSV import |
||
| (One intermediate revision by the same user not shown) | |||
| Line 37: | Line 37: | ||
[[Category:Mathematical Theorems]] | [[Category:Mathematical Theorems]] | ||
{{stub}} | {{stub}} | ||
{{No image}} | |||
__NOINDEX__ | |||
Latest revision as of 06:43, 17 March 2025
Chebyshev's Inequality is a fundamental theorem in probability theory and statistics, named after the Russian mathematician Pafnuty Chebyshev. It provides a universal bound on the probability that the outcome of a random variable deviates from its mean.
Etymology[edit]
The theorem is named after Pafnuty Chebyshev, a prominent 19th-century Russian mathematician who made significant contributions to number theory, probability theory, and mechanics.
Statement of the Theorem[edit]
In the field of probability theory, Chebyshev's Inequality states that for a wide class of probability distributions, no more than a certain amount of values can be more than a certain distance from the mean. Specifically, for any random variable with a finite expected value μ and finite non-zero variance σ², the inequality is defined as:
- Pr(|X - μ| ≥ kσ) ≤ 1/k²
where:
- Pr denotes the probability
- X is a random variable
- μ is the expected value (mean) of X
- σ is the standard deviation of X
- k is a positive real number
Applications[edit]
Chebyshev's Inequality has wide applications in various fields including statistics, economics, computer science, finance, and physics. It is often used to prove the Weak Law of Large Numbers. It is also used in the proof of the Central Limit Theorem.
Related Terms[edit]
- Probability Theory
- Random Variable
- Expected Value
- Standard Deviation
- Variance
- Weak Law of Large Numbers
- Central Limit Theorem


