Borel-Cantelli Lemma: Difference between revisions
Line 42: | Line 42: | ||
<math> \{x: \cup_{k}\cap_n \cup_{m\geq n} |\overline X_m| \geq \frac{1}{k}\} </math>. By continuity of measure, we have <math> P(\{x: \cup_{k}\cap_n \cup_{m\geq n} |\overline X_m| \geq \frac{1}{k}\}) = \lim_{k \to \infty} P \{x: \cap_n \cup_{m\geq n} |\overline X_m| \geq \frac{1}{k}\} </math>. | <math> \{x: \cup_{k}\cap_n \cup_{m\geq n} |\overline X_m| \geq \frac{1}{k}\} </math>. By continuity of measure, we have <math> P(\{x: \cup_{k}\cap_n \cup_{m\geq n} |\overline X_m| \geq \frac{1}{k}\}) = \lim_{k \to \infty} P \{x: \cap_n \cup_{m\geq n} |\overline X_m| \geq \frac{1}{k}\} </math>. | ||
It suffices to show that for any fixed <math display="inline">k</math>, <math display="inline">P(\cap_{n}\cup_{k\geq n} |\overline X_n|> \frac{1}{k}) = 0</math>. This closely resembles the conclusion of the Borel-Cantelli lemma, so we might be tempted to show that <math display="inline">\sum_{n=1}^\infty P(|\overline X_n| > \frac{1}{k}) < \infty</math>. Unfortunately, a simple application of Chebyshev's inequality reveals that this fall short | It suffices to show that for any fixed <math display="inline">k</math>, <math display="inline">P(\cap_{n}\cup_{k\geq n} |\overline X_n|> \frac{1}{k}) = 0</math>. This closely resembles the conclusion of the Borel-Cantelli lemma, so we might be tempted to show that <math display="inline">\sum_{n=1}^\infty P(|\overline X_n| > \frac{1}{k}) < \infty</math>. Unfortunately, a simple application of Chebyshev's inequality reveals that this fall short, as | ||
<math display="block">\sum_{n=1}^\infty P(|\overline X_n|> \frac{1}{k}) \leq \sum_{n=1}^\infty \frac{k^2\sigma^2}{n} = \infty.</math> | <math display="block">\sum_{n=1}^\infty P(|\overline X_n|> \frac{1}{k}) \leq \sum_{n=1}^\infty \frac{k^2\sigma^2}{n} = \infty.</math> |
Revision as of 20:59, 18 December 2020
Many interesting theorem in probability utilizes measure theory, as it often provides some form of one-line argument that allows the proof to go through. Among one of them is the law of large numbers, which informally states that the average of iid samples approaches the mean of the distribution as the sample size grows large.
We would like to be more precise about what the words "approaches" mean. Clearly we cannot replace the word "approaches" with "converges to", as one could be unlucky and consistently draw below or above the mean so that the sample average may never even be close to the true mean. Despite of this, one thing that we do know is that consistently drawing below or above the mean, while not impossible, becomes increasingly unlikely as the sample grows large. This naturally leads to the idea of convergence in probability.
Convergence in probability
A sequence of random variables converges in probability to if for any ,
Notice that this is simply convergence in measure in probability space. Equipped with this definition, we can now state the weak law of large numbers: the sample average converges in probability to the mean. Bernoulli first proved this theorem for Bernoulli random variables in 1713, when tools like Chebyshev's inequality wasn't discovered. Later on, mathmaticians proved more general cases that does not the require the assumption of finite variance or iid random variables.
Proof. Without loss of generality assume that the random variables are centered, and let denote the sample average. By Chebyshev's inequality, as approaches infinity. Note that for Chebyshev's inequality to hold we need finite variance.
It is natural to wonder the possibility of obtaining a stronger notion of convergence, perhaps under stronger condition than finite variance. To do so we need to formally define the stronger type of convergence that we need, which in study of probability we often refer to as "almost sure convergence".
Almost sure convergence
A sequence of random variables converges almost surely to if
The strong law of large number carries this stronger notion of converges and states: the sample average converges almost surely to the mean. Before the proof, we first introduce a lemma.
Borel Catelli Lemma
Let be a probability space, and any sequence of events in . If , then .
On the other hand, if , then .
Proof.
Fix , there exists sufficiently large such that . By union bound we obtain . So . By continuity of measure , we have .
The other statement follows easily from the similar technique.
Strong law of large numbers
Instead of directly stating the theorem, we explore a little bit on our own as to what assumptions we might need in order to achieve this "almost sure convergence" to the mean. Here, we are interested in the set where converges to , and we can rewrite this set as . To relate it to Borel-Cantelli lemma, we instead study its complement . By continuity of measure, we have .
It suffices to show that for any fixed , . This closely resembles the conclusion of the Borel-Cantelli lemma, so we might be tempted to show that . Unfortunately, a simple application of Chebyshev's inequality reveals that this fall short, as