Stochastic convergence: Difference between revisions

From Citizendium
Jump to navigation Jump to search
imported>Ragnar Schroder
(→‎Related topics: added link to yet unwritten time series article)
imported>Pat Palmer
mNo edit summary
 
(18 intermediate revisions by 4 users not shown)
Line 1: Line 1:
{{subpages}}
{{ dambigbox| Stochastic convergence | Convergence }}
'''Stochastic convergence''' is a mathematical concept intended to formalize the idea that a sequence of essentially random or unpredictable events sometimes is expected to settle into a pattern.
'''Stochastic convergence''' is a mathematical concept intended to formalize the idea that a sequence of essentially random or unpredictable events sometimes is expected to settle into a pattern.


The pattern may for instance be
The pattern may for instance be
*[[convergence]] in the classical sense to a fixed value,  perhaps itself coming from a random event
*[[Convergence]] in the classical sense to a fixed value,  perhaps itself coming from a random event
*an increasing similarity of outcomes to what a purely deterministic function would produce
*An increasing similarity of outcomes to what a purely deterministic function would produce
*an increasing preference towards a certain outcome
*An increasing preference towards a certain outcome
*an increasing "aversion" against straying far away from a certain outcome
*An increasing "aversion" against straying far away from a certain outcome


Some less obvious, more theoretical patterns could be  
Some less obvious, more theoretical patterns could be  
*that the probability distribution describing the next outcome may grow increasingly similar to a certain distribution
*That the probability distribution describing the next outcome may grow increasingly similar to a certain distribution
*that the series formed by calculating the [[expected value]] of the outcome's distance from a particular value may converge to 0
*That the series formed by calculating the [[expected value]] of the outcome's distance from a particular value may converge to 0
*that the variance of the [[random variable]] describing the next event grows smaller and smaller.
*That the variance of the [[random variable]] describing the next event grows smaller and smaller.
 
 




==Various possible modes of stochastic convergence==
==Various possible modes of stochastic convergence==


Four different varieties of stochastic convergence are noted:
Four different varieties of stochastic convergence are noted:
Line 27: Line 27:




----
==Almost sure convergence==
===Almost sure convergence===
This is the type of stochastic convergence that is most similar to ordinary [[convergence]] known from elementary [[real analysis]].
 
This is the type of stochastic convergence that is most similar to [[pointwise convergence]] known from elementary [[real analysis]].
 
{|align="right" cellpadding="10" style="background-color:lightblue; width:40%; border: 1px solid #aaa; margin:2px; font-size: 90%;"
|'''Examples of almost sure convergence.'''


===Examples===
<b>Basic example 1</b>
====Basic example 1====
Consider an animal of some short-lived species. We note the exact amount of food that this animal consumes day by day. This sequence of numbers will be unpredictable in advance, but we may be ''quite certain'' that one day the number will become zero, and will stay zero forever after.
Consider an animal of some short-lived species. We note the exact amount of food that this animal consumes day by day. This sequence of numbers will be unpredictable in advance, but we may be ''quite certain'' that one day the number will become zero, and will stay zero forever after.


====Basic example 2====
<b>Basic example 2</b>
Consider a man who starts tomorrow to toss seven coins once every morning.  Each afternoon, he donates a random amount of money to a certain charity.  The first time the result is all tails, however, he will stop permanently.
Consider a man who starts tomorrow to toss seven coins once every morning.  Each afternoon, he donates a random amount of money to a certain charity.  The first time the result is all tails, however, he will stop permanently.


Line 44: Line 47:
However,  when we consider ''any finite number'' of days,  there is a nonzero probability the terminating condition will not occur.
However,  when we consider ''any finite number'' of days,  there is a nonzero probability the terminating condition will not occur.


====Intermediate example====
<b>Intermediate example</b>
A business owner has two sources of income:  His business,  and interest from a large bank deposit with fixed interest and no withdrawal or deposits.
A business owner has two sources of income:  His business,  and interest from a large bank deposit with fixed interest and no withdrawal or deposits.


Line 55: Line 58:
Then the total monthly income <math>U_i</math> has almost sure convergence to the function f(i).  
Then the total monthly income <math>U_i</math> has almost sure convergence to the function f(i).  


===Formal definition===
|}
===Informal description of almost sure convergence===
 
We are confronted with an infinite sequence of random experiments: Experiment 1, experiment 2, experiment 3 ... , where the outcome of each experiment will generate a real number. 
The random experiments will thus generate a  sequence of real numbers,  typically denoted ''x''<sub>1</sub>, ''x''<sub>2</sub>, ''x''<sub>3</sub>... .
 
If we have formulas available that describe the probabilities involved in each experiment,  then we may say something about the [[probability]] that this [[sequence]] will [[convergence|converge]] to a given value.


Let <math>\scriptstyle X_0, X_1, ... </math> be an infinite sequence of [[random variable|random variables]] defined over a subset of R.  
If this probability is 1, then the phenomenon of "almost sure convergence" is present.
 
Note that in [[Almost sure convergence|advanced treatments]] the outcomes are not restricted to real numbers.
 
====Formal definition====
 
Let '''X'''<sub>0</sub>, '''X'''<sub>1</sub>, '''X'''<sub>2</sub>... be an infinite sequence of [[random variable|random variables]] defined over a subset of R.  


Then the actual outcomes will be an ordinary sequence of real numbers.  
Then the actual outcomes will be an ordinary sequence of real numbers.  
Line 73: Line 88:
Commonly used notation: <math>X_i \stackrel{a.s.}{\rightarrow} a </math>, <math>X_i \stackrel{a.s.}{\rightarrow} X </math>.
Commonly used notation: <math>X_i \stackrel{a.s.}{\rightarrow} a </math>, <math>X_i \stackrel{a.s.}{\rightarrow} X </math>.


----


==Convergence in probability==
The basic idea is that the probability of an "freaky" outcome becomes smaller and smaller, while the idea of "non-freaky" may become stricter and stricter.


===Examples===


====Basic example====
===Convergence in probability===
Assume someone has an infinite supply of envelopes and money,  and likes to play the following game: He shows you a number of sealed envelopes. Inside exactly one is a $5 bill, any and all of the others all contain $1. You get to pick exactly one envelope, and keep the content.
 
The basic idea is that the probability of an "unusual" outcome becomes smaller and smaller as the sequence progresses.
 
 
{|align="right" cellpadding="10" style="background-color:lightblue; width:40%; border: 1px solid #aaa; margin:2px; font-size: 90%;"
|'''Illustration of convergence in probability.'''
{{Image|Convergenceinprobabilitysample1.gif|right|224px|Sequence of stochastic variables exhibiting typical behavior}}


Tomorrow he plays the game using 1 envelope, the second day he uses two envelopes, the 3rd day there'll be 3 envelopes, etc.
Inspecting this graph, you'll notice the behavior typical of convergence in probability.  


Then the amount of dollars you earn from the game on a given day will be a random variable that converges in probability to $1, because the probability of guessing the "right" envelope steadily shrinks.
|}


In the long run, then, an outcome different from $1 will become an increasingly "freaky" event.


===Formal definition===
 
 
 
====Formal definition====
Let <math>\scriptstyle X_0, X_1, ... </math> be an infinite sequence of [[random variable|random variables]] defined over a subset of R.  
Let <math>\scriptstyle X_0, X_1, ... </math> be an infinite sequence of [[random variable|random variables]] defined over a subset of R.  


Line 97: Line 116:


Commonly used notation: <math>X_i \stackrel{P}{\rightarrow} a</math>.
Commonly used notation: <math>X_i \stackrel{P}{\rightarrow} a</math>.
----


==Convergence in distribution==
 
 
===Convergence in distribution===
With this mode of convergence,  we increasingly expect to see our next outcome in a sequence of random experiments becoming better and better modeled by a given [[probability distribution]].
With this mode of convergence,  we increasingly expect to see our next outcome in a sequence of random experiments becoming better and better modeled by a given [[probability distribution]].


===Examples===
{|align="right" cellpadding="10" style="background-color:lightblue; width:40%; border: 1px solid #aaa; margin:2px; font-size: 90%;"
====Basic example====
|'''Examples of convergence in distribution.'''
 
<b>Basic example:</b>
The outcome from tossing a non-biased dice follows the [[uniform discrete distribution]].
The outcome from tossing a non-biased dice follows the [[uniform discrete distribution]].


Line 112: Line 134:
As the factory is improved,  the dices will be less and less loaded,  and the outcomes from tossing a newly produced dice will follow the desired distribution more and more closely.
As the factory is improved,  the dices will be less and less loaded,  and the outcomes from tossing a newly produced dice will follow the desired distribution more and more closely.


====Intermediate example====
<b>Intermediate example:</b>


Let <math>\scriptstyle X_n</math> be the result of flipping n unbiased coins,  and noting the fraction of heads.
Let <math>\scriptstyle X_n</math> be the result of flipping n unbiased coins,  and noting the fraction of heads.
Line 124: Line 146:
(This convergence follows from the famous [[central limit theorem]]).
(This convergence follows from the famous [[central limit theorem]]).


===Formal definition===
|}
 
 
====Formal definition====


Given a random variable X with a [[cumulative distribution function]] F(x), let <math>X_i</math> be a sequence of random variables,  each with cumulative distribution function <math>F_i (x)</math>, respectively.
Given a random variable X with a [[cumulative distribution function]] F(x), let <math>X_i</math> be a sequence of random variables,  each with cumulative distribution function <math>F_i (x)</math>, respectively.
Line 130: Line 155:
If <math>\scriptstyle \lim_{i \to \infty} F_i (x) = F(x)</math> for all x where F(x) is continuous,  then the sequence <math>X_i</math> of stochastic variables converges in distribution to the distribution of <math>X</math>.
If <math>\scriptstyle \lim_{i \to \infty} F_i (x) = F(x)</math> for all x where F(x) is continuous,  then the sequence <math>X_i</math> of stochastic variables converges in distribution to the distribution of <math>X</math>.


Commonly used notation: <math>X_i \stackrel{L}{\rightarrow} X</math>.  One can also use the distribution directly,  so if f.i. X is normally distributed with mean 0 and variance 1,  one could write <math>X_i \stackrel{L}{\rightarrow} N(0,1)</math>.  
Commonly used notation: <math>X_i \stackrel{L}{\rightarrow} X</math>.  One can also use the distribution directly,  so if f.i. X is normally distributed with mean 0 and variance 1,  one could write <math>X_i \stackrel{L}{\rightarrow} N(0,1)</math>.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 


----


==Convergence in rth order mean==
===Convergence in r-th order mean===
This is a rather "technical" mode of convergence. We essentially compute a sequence of real numbers,  one number for each random variable,  and check if this sequence is [[convergence|convergent]] in the ordinary sense.  
This is a rather "technical" mode of convergence. We essentially compute a sequence of real numbers,  one number for each random variable,  and check if this sequence is [[convergence|convergent]] in the ordinary sense.  


===Example===
{|align="right" cellpadding="10" style="background-color:lightblue; width:40%; border: 1px solid #aaa; margin:2px; font-size: 90%;"
|'''Examples of convergence in r-th order mean.'''
 
<b>Basic example:</b>
A newly built factory produces cans of beer.  The owners want each can to contain ''exactly'' a certain amount.
A newly built factory produces cans of beer.  The owners want each can to contain ''exactly'' a certain amount.


Line 145: Line 197:


This example illustrates convergence in first order mean.
This example illustrates convergence in first order mean.
|}


===Formal definition===
====Formal definition====
If <math>\scriptstyle \lim_{n \to \infty} E(|X_n - a|^r ) =0</math>  for some real number a,  then {<math>X_n</math>} converges in rth order mean to a.
If <math>\scriptstyle \lim_{n \to \infty} E(|X_n - a|^r )=0</math>  for some real number a,  then {<math>X_n</math>} converges in r-th order mean to a.


Commonly used notation: <math>X_n \stackrel{L_r}{\rightarrow} a</math>.
Commonly used notation: <math>X_n \stackrel{L_r}{\rightarrow} a</math>.
----


==Relations between the different modes of convergence==
==Relations between the different modes of convergence==
*If a sequence of random variables has almost sure convergence,  then it also has convergence in probability.
*If a sequence of random variables has almost sure convergence,  then it also has convergence in probability.
*If a sequence of random variables has convergence in probability,  then it also has convergence in distribution.
*If a sequence of random variables has convergence in probability,  then it also has convergence in distribution.
*If a sequence of random variables has convergence in (r+1)th order mean, then it also has convergence in rth order mean (r>0).
*If a sequence of random variables has convergence in (r+1)-th order mean, then it also has convergence in r-th order mean (r>0).
*If a sequence of random variables has convergence in rth order mean, then it also has convergence in probability.
*If a sequence of random variables has convergence in rth order mean, then it also has convergence in probability.


Line 164: Line 216:
*[[Convergence in distribution]]
*[[Convergence in distribution]]
*[[Convergence in probability]]
*[[Convergence in probability]]
*[[Convergence in rth order mean]]
*[[Convergence in r-th order mean]]




Line 172: Line 224:
*[[Random variable]]
*[[Random variable]]
*[[Stochastic process]]
*[[Stochastic process]]
*[[time series]]
*[[Time series]]
*[[Stochastic differential equations]]
*[[Stochastic differential equation]]
*[[Stochastic modeling]]
*[[Stochastic modeling]]


== References==
== References==
 
#P. Billingsley, ''Probability and Measure'' (2 ed.), ser. Wiley Series in Probability and Mathematical Statistics, Wiley, 1986.
#D. Williams, ''Probability with Martingales'', Cambridge : Cambridge University Press, 1991.
#E. Wong and B. Hajek, ''Stochastic Processes in Engineering Systems'', New York: Springer-Verlag,    1985.


== External links ==
== External links ==
 
#[http://www.probability.net Probability tutorial at Probability.net]
 
[[Category:CZ Live]]
[[Category:Mathematics Workgroup]]
[[Category:Physics Workgroup]]
[[Category:Chemistry Workgroup]]

Latest revision as of 19:57, 29 September 2020

This article is developing and not approved.
Main Article
Discussion
Related Articles  [?]
Bibliography  [?]
External Links  [?]
Citable Version  [?]
 
This editable Main Article is under development and subject to a disclaimer.
This article is about Stochastic convergence. For other uses of the term Convergence , please see Convergence (disambiguation).

Stochastic convergence is a mathematical concept intended to formalize the idea that a sequence of essentially random or unpredictable events sometimes is expected to settle into a pattern.

The pattern may for instance be

  • Convergence in the classical sense to a fixed value, perhaps itself coming from a random event
  • An increasing similarity of outcomes to what a purely deterministic function would produce
  • An increasing preference towards a certain outcome
  • An increasing "aversion" against straying far away from a certain outcome

Some less obvious, more theoretical patterns could be

  • That the probability distribution describing the next outcome may grow increasingly similar to a certain distribution
  • That the series formed by calculating the expected value of the outcome's distance from a particular value may converge to 0
  • That the variance of the random variable describing the next event grows smaller and smaller.


Various possible modes of stochastic convergence

Four different varieties of stochastic convergence are noted:

  • Almost sure convergence
  • Convergence in probability
  • Convergence in distribution
  • Convergence in rth order mean


Almost sure convergence

This is the type of stochastic convergence that is most similar to pointwise convergence known from elementary real analysis.

Examples of almost sure convergence.

Basic example 1 Consider an animal of some short-lived species. We note the exact amount of food that this animal consumes day by day. This sequence of numbers will be unpredictable in advance, but we may be quite certain that one day the number will become zero, and will stay zero forever after.

Basic example 2 Consider a man who starts tomorrow to toss seven coins once every morning. Each afternoon, he donates a random amount of money to a certain charity. The first time the result is all tails, however, he will stop permanently.

Let be the day by day amounts the charity receives from him.

We may be almost sure that one day this amount will be zero, and stay zero forever after that.

However, when we consider any finite number of days, there is a nonzero probability the terminating condition will not occur.

Intermediate example A business owner has two sources of income: His business, and interest from a large bank deposit with fixed interest and no withdrawal or deposits.

The business income varies unpredictably from month to month, while income from interest is predictable and given by a simple function f.

The income for month i can thus be modeled by a random variable , where is the income from the business.

Now assume converges almost surely to 0 (history bears out that all businesses sooner or later fold up).

Then the total monthly income has almost sure convergence to the function f(i).

Informal description of almost sure convergence

We are confronted with an infinite sequence of random experiments: Experiment 1, experiment 2, experiment 3 ... , where the outcome of each experiment will generate a real number. The random experiments will thus generate a sequence of real numbers, typically denoted x1, x2, x3... .

If we have formulas available that describe the probabilities involved in each experiment, then we may say something about the probability that this sequence will converge to a given value.

If this probability is 1, then the phenomenon of "almost sure convergence" is present.

Note that in advanced treatments the outcomes are not restricted to real numbers.

Formal definition

Let X0, X1, X2... be an infinite sequence of random variables defined over a subset of R.

Then the actual outcomes will be an ordinary sequence of real numbers.

If the probability that this sequence will converge to a given real number a equals 1, then we say the original sequence of stochastic variables has almost sure convergence to a.

In more compact notation:

If for some a, then the sequence has almost sure convergence to .

Note that we may replace the real number a above by a real-valued function of i, and obtain almost sure convergence to a function rather than a fixed number.

The number a may also be the outcome of a random variable X. In that case the compact but somewhat confusing notation is often used.

Commonly used notation: , .



Convergence in probability

The basic idea is that the probability of an "unusual" outcome becomes smaller and smaller as the sequence progresses.


Illustration of convergence in probability.
PD Image
Sequence of stochastic variables exhibiting typical behavior

Inspecting this graph, you'll notice the behavior typical of convergence in probability.



Formal definition

Let be an infinite sequence of random variables defined over a subset of R.

If there exists a real number a such that for all , then the sequence has convergence in probability to a.


Commonly used notation: .


Convergence in distribution

With this mode of convergence, we increasingly expect to see our next outcome in a sequence of random experiments becoming better and better modeled by a given probability distribution.

Examples of convergence in distribution.

Basic example: The outcome from tossing a non-biased dice follows the uniform discrete distribution.

Assume a new dice factory has just been built.

The first few dices come out quite biased, due to imperfections in the production process. The outcome from tossing any of them will follow a distribution markedly different from the desired uniform discrete distribution.

As the factory is improved, the dices will be less and less loaded, and the outcomes from tossing a newly produced dice will follow the desired distribution more and more closely.

Intermediate example:

Let be the result of flipping n unbiased coins, and noting the fraction of heads.

will then follow the uniform discrete probability distribution with expected value and variance , but as n grows larger, will follow a distribution that gradually takes on more and more similarity to the gaussian distribution .

Forming the sequence , we find the random variables becoming distributed more and more like the standard normal distribution as n increases.

We then say the sequence converges in distribution to the standard normal distribution.

(This convergence follows from the famous central limit theorem).


Formal definition

Given a random variable X with a cumulative distribution function F(x), let be a sequence of random variables, each with cumulative distribution function , respectively.

If for all x where F(x) is continuous, then the sequence of stochastic variables converges in distribution to the distribution of .

Commonly used notation: . One can also use the distribution directly, so if f.i. X is normally distributed with mean 0 and variance 1, one could write .














Convergence in r-th order mean

This is a rather "technical" mode of convergence. We essentially compute a sequence of real numbers, one number for each random variable, and check if this sequence is convergent in the ordinary sense.

Examples of convergence in r-th order mean.

Basic example: A newly built factory produces cans of beer. The owners want each can to contain exactly a certain amount.

Knowing the details of the current production process, engineers may compute the expected error in a newly produced can.

They are continuously improving the production process, so as time goes by, the expected error in a newly produced can tends to zero.

This example illustrates convergence in first order mean.


Formal definition

If for some real number a, then {} converges in r-th order mean to a.

Commonly used notation: .

Relations between the different modes of convergence

  • If a sequence of random variables has almost sure convergence, then it also has convergence in probability.
  • If a sequence of random variables has convergence in probability, then it also has convergence in distribution.
  • If a sequence of random variables has convergence in (r+1)-th order mean, then it also has convergence in r-th order mean (r>0).
  • If a sequence of random variables has convergence in rth order mean, then it also has convergence in probability.


See also


Related topics

References

  1. P. Billingsley, Probability and Measure (2 ed.), ser. Wiley Series in Probability and Mathematical Statistics, Wiley, 1986.
  2. D. Williams, Probability with Martingales, Cambridge : Cambridge University Press, 1991.
  3. E. Wong and B. Hajek, Stochastic Processes in Engineering Systems, New York: Springer-Verlag, 1985.

External links

  1. Probability tutorial at Probability.net