{"id":8576,"date":"2015-11-30T23:37:55","date_gmt":"2015-11-30T22:37:55","guid":{"rendered":"http:\/\/djalil.chafai.net\/blog\/?p=8576"},"modified":"2015-12-02T08:56:20","modified_gmt":"2015-12-02T07:56:20","slug":"back-to-basics-polya-urns","status":"publish","type":"post","link":"https:\/\/djalil.chafai.net\/blog\/2015\/11\/30\/back-to-basics-polya-urns\/","title":{"rendered":"Back to basics - P\u00f3lya urns"},"content":{"rendered":"<figure id=\"attachment_8580\" aria-describedby=\"caption-attachment-8580\" style=\"width: 227px\" class=\"wp-caption aligncenter\"><a href=\"https:\/\/fr.wikipedia.org\/wiki\/George_P%C3%B3lya\"><img loading=\"lazy\" src=\"http:\/\/djalil.chafai.net\/blog\/wp-content\/uploads\/2015\/12\/Georges-Polya-227x300.jpg\" alt=\"Georges P\u00f3lya\" width=\"227\" height=\"300\" class=\"size-medium wp-image-8580\" srcset=\"https:\/\/djalil.chafai.net\/blog\/wp-content\/uploads\/2015\/12\/Georges-Polya-227x300.jpg 227w, https:\/\/djalil.chafai.net\/blog\/wp-content\/uploads\/2015\/12\/Georges-Polya-113x150.jpg 113w, https:\/\/djalil.chafai.net\/blog\/wp-content\/uploads\/2015\/12\/Georges-Polya.jpg 242w\" sizes=\"(max-width: 227px) 100vw, 227px\" \/><\/a><figcaption id=\"caption-attachment-8580\" class=\"wp-caption-text\">Georges P\u00f3lya (1887  - 1985)<\/figcaption><\/figure>\n<p style=\"text-align: justify;\">This post is devoted to <b>P\u00f3lya urns<\/b>, probably the simplest stochastic model of <b>reinforcement<\/b>. At time \\( {n=0} \\), we prepare an urn containing \\( {a&gt;0} \\) <b>azure<\/b> balls and \\( {b&gt;0} \\) <b>black<\/b> balls. To produce the configuration of the urn at time \\( {n=1} \\), we draw at random a ball in the urn, and then we replace this ball in the urn together with a new ball of the same color. We repeat this mechanism independently to produce the configuration of the urn at any time \\( {n\\in\\mathbb{N}} \\).<\/p>\n<p style=\"text-align: justify;\">Let \\( {M_n\\in[0,1]} \\) be the proportion of azure balls in the urn at time \\( {n\\in\\mathbb{N}} \\). At time \\( {n\\in\\mathbb{N}} \\), the urn contains \\( {a+b+n} \\) balls, and \\( {(a+b+n)M_n} \\) azure balls. Conditional on \\( {M_n} \\), an azure ball is added to the urn at time \\( {n+1} \\) with probability \\( {M_n} \\). Hence<\/p>\n<p style=\"text-align: center;\">\\[ M_0=\\frac{a}{a+b} \\quad\\text{and}\\quad M_{n+1}=\\frac{(a+b+n)M_n+\\mathbf{1}_{\\{U_{n+1}\\leq M_n\\}}}{a+b+n+1} \\]<\/p>\n<p style=\"text-align: justify;\">where \\( {{(U_n)}_{n\\geq1}} \\) are iid uniform random variables on \\( {[0,1]} \\), and the event \\( {\\{U_{n+1}\\leq M_n\\}} \\) corresponds to add an azure ball. The stochastic recursive sequence \\( {{(M_n)}_{n\\geq0}} \\) is a non-homogeneous <b>Markov chain<\/b> with state space \\( {[0,1]} \\), and also a <b>martingale<\/b>.<\/p>\n<p style=\"text-align: justify;\">If we encode the result of the \\( {n} \\) drawing by a random variable \\( {X_n} \\) with values in \\( {\\{\\alpha,\\beta\\}} \\) where \\( {\\alpha} \\) and \\( {\\beta} \\) indicate respectively that the ball is azure or black. Let \\( {Y_n} \\) and \\( {Z_n} \\) be respectively the number of azure and black balls added to the urn after the first \\( {n} \\) drawing. At time \\( {n} \\), the urn contains<\/p>\n<p style=\"text-align: center;\">\\[ a+Y_n+b+Z_n=a+b+n \\]<\/p>\n<p style=\"text-align: justify;\">balls, more precisely the number of azure balls is<\/p>\n<p style=\"text-align: center;\">\\[ a+Y_n=a+\\sum_{k=1}^n\\mathbf{1}_{\\{X_k=\\alpha\\}}=(a+b+n)M_n \\]<\/p>\n<p style=\"text-align: justify;\">while the number of black balls is<\/p>\n<p style=\"text-align: center;\">\\[ b+Z_n=b+\\sum_{k=1}^n\\mathbf{1}_{\\{X_k=\\beta\\}}. \\]<\/p>\n<p style=\"text-align: justify;\">We have \\( {Y_n+Z_n=n} \\). The sequence \\( {{(X_n,Y_n,Z_n)}_{n\\geq1}} \\) satisfies the stochastic recursion<\/p>\n<p style=\"text-align: center;\">\\[ (X_{n+1},Y_{n+1},Z_{n+1}) = \\left\\{ \\begin{array}{rl} (\\alpha,Y_n+1,Z_n) & \\mbox{if } U_{n+1}\\leq\\frac{a+Y_n}{a+b+n},\\\\ (\\beta,Y_n,Z_n+1) & \\mbox{if } U_{n+1}&gt;\\frac{a+Y_n}{a+b+n}, \\end{array} \\right. \\]<\/p>\n<p style=\"text-align: justify;\">with by convention \\( {Y_0=0} \\) and \\( {Z_0=0} \\) (the value of \\( {X_0} \\) is useless).<\/p>\n<p style=\"text-align: justify;\"><b>Martingale.<\/b> The sequence \\( {{(M_n)}_{n\\geq0}} \\) is a <b>martingale<\/b> taking its values in \\( {[0,1]} \\) for the filtration \\( {{(\\mathcal{F}_n)}_{n\\geq0}} \\) defined by \\( {\\mathcal{F}_n=\\sigma(U_1,\\ldots,U_n)} \\). Indeed \\( {M_n\\in L^1(\\mathcal{F}_n)} \\) and<\/p>\n<p style=\"text-align: center;\">\\[ \\mathbb{E}(M_{n+1}\\,|\\,\\mathcal{F}_n) =\\mathbb{E}(M_{n+1}\\,|\\,M_n) =\\frac{(a+b+n)M_n+M_n}{a+b+n+1} =M_n. \\]<\/p>\n<p style=\"text-align: justify;\">In particular we have the following <b>conservation law<\/b>: for all \\( {n\\in\\mathbb{N}} \\),<\/p>\n<p style=\"text-align: center;\">\\[ \\mathbb{E}(M_n)=\\mathbb{E}(M_0)=\\frac{a}{a+b}. \\]<\/p>\n<p style=\"text-align: justify;\">It follows from martingale limit theorems (the martingale is uniformly bounded and thus uniformly integrable!) that there exists a random variable \\( {M_\\infty} \\) on \\( {[0,1]} \\) such that<\/p>\n<p style=\"text-align: center;\">\\[ \\lim_{n\\rightarrow\\infty}M_n = M_\\infty \\]<\/p>\n<p style=\"text-align: justify;\">almost surely and in \\( {L^p} \\) for any \\( {p\\geq1} \\). In particular \\( {\\mathbb{E}(M_\\infty)=\\mathbb{E}(M_0)=\\frac{a}{a+b}} \\).<\/p>\n<p style=\"text-align: justify;\"><b>Limiting distribution.<\/b> The random variable \\( {M_\\infty} \\) follows the <b>Beta distribution<\/b> on \\( {[0,1]} \\) with parameter \\( {(a,b)} \\) and density<\/p>\n<p style=\"text-align: center;\">\\[ u\\in[0,1]\\mapsto\\frac{u^{a-1}(1-u)^{b-1}}{\\mathrm{Beta}(a,b)} \\quad\\text{where}\\quad \\mathrm{Beta}(a,b):=\\int_0^1\\!p^{a-1}(1-p)^{b-1}\\,dp. \\]<\/p>\n<p style=\"text-align: justify;\">In particular if \\( {a=b=1} \\) then \\( {M_\\infty} \\) follows the uniform distribution on \\( {[0,1]} \\).<\/p>\n<p style=\"text-align: justify;\">To see it let us recall first of all that<\/p>\n<p style=\"text-align: center;\">\\[ \\mathrm{Beta}(a,b) =\\frac{\\Gamma(a)\\Gamma(b)}{\\Gamma(a+b)} \\quad\\text{where}\\quad \\Gamma(x):=\\int_0^\\infty\\!t^{x-1}e^{-x}\\,dx. \\]<\/p>\n<p style=\"text-align: justify;\">For any \\( {c} \\) and \\( {k} \\) we set<\/p>\n<p style=\"text-align: center;\">\\[ c^{(k)}=c(c+1)\\cdots(c+k-1)=\\frac{(c+k-1)!}{(c-1)!}=\\frac{\\Gamma(c+k)}{\\Gamma(c)}. \\]<\/p>\n<p style=\"text-align: justify;\">Now for any \\( {x_1,\\ldots,x_n} \\) in \\( {\\{0,1\\}} \\), we have, denoting \\( {k=x_1+\\cdots+x_n} \\),<\/p>\n<p style=\"text-align: center;\">\\[ \\mathbb{P}(\\mathbf{1}_{\\{X_1=\\alpha\\}}=x_1,\\ldots,\\mathbf{1}_{\\{X_n=\\alpha\\}}=x_n) =\\frac{a^{(k)}b^{(n-k)}}{(a+b)^{(n)}}. \\]<\/p>\n<p style=\"text-align: justify;\">This probability is invariant by permutation of \\( {x_1,\\ldots,x_n} \\): this means that the distribution of the random vector \\( {(\\mathbf{1}_{\\{X_1=\\alpha\\}},\\ldots,\\mathbf{1}_{\\{X_n=\\alpha\\}})} \\) is <b>exchangeable<\/b>. Hence the random variable<\/p>\n<p style=\"text-align: center;\">\\[ Y_n=\\sum_{k=1}^n\\mathbf{1}_{\\{X_k=\\alpha\\}} \\]<\/p>\n<p style=\"text-align: justify;\">satisfies, for any \\( {k\\in\\{0,1,\\ldots,n\\}} \\),<\/p>\n<p style=\"text-align: center;\">\\[ \\begin{array}{rcl} \\mathbb{P}(Y_n=k) &=&\\binom{n}{k}\\frac{a^{(k)}b^{(n-k)}}{(a+b)^{(n)}}\\\\ &=&\\binom{n}{k} \\frac{\\Gamma(a+k)\\Gamma(b+n-k)\\Gamma(a+b)}{\\Gamma(a)\\Gamma(b)\\Gamma(a+b+n)}\\\\ &=&\\binom{n}{k} \\frac{\\mathrm{Beta}(a+k,b+n-k)}{\\mathrm{Beta}(a,b)}\\\\ &=&\\int_0^1\\! \\binom{n}{k}p^{k}(1-p)^{n-k}\\frac{p^{a-1}(1-p)^{b-1}}{\\mathrm{Beta}(a,b)}\\,dp. \\end{array} \\]<\/p>\n<p style=\"text-align: justify;\">We say that \\( {Y_n} \\) follows the <b>Beta-binomial distribution<\/b>, which is a a mixture of binomial laws of size \\( {n} \\) with a parameter \\( {p} \\) which follows the Beta distribution of parameter \\( {(a,b)} \\). We have \\( {M_n=(a+Y_n)\/(a+b+n)} \\) with \\( {Y_0=0} \\). When \\( {a=b=1} \\), then formula for the law of \\( {Y_n} \\) indicates that \\( {Y_n} \\) is uniform on \\( {\\{0,1,\\ldots,n\\}} \\), and thus \\( {M_n} \\) is uniform on \\( {\\{1\/(n+2),\\ldots,(n+1)\/(n+2)\\}} \\), which implies that \\( {M_\\infty} \\) is uniform on \\( {[0,1]} \\). In the general case, for any \\( {t\\in[0,1]} \\),<\/p>\n<p style=\"text-align: center;\">\\[ \\begin{array}{rcl} \\mathbb{P}(M_\\infty\\leq t) &=&\\lim_{n\\rightarrow\\infty}\\mathbb{P}(Y_n\\leq(a+b+n)t-a)\\\\ &=&\\cdots =\\frac{1}{\\mathrm{Beta}(a,b)}\\int_0^t\\!u^{a-1}(1-u)^{b-1}\\,du. \\end{array} \\]<\/p>\n<p style=\"text-align: justify;\">The computation is based on the so-called Beta-binomial correspondence, which states that if \\( {U_1,\\ldots,U_n} \\) are iid and uniform on \\( {[0,1]} \\) and if \\( {U_{(1)}\\leq\\cdots\\leq U_{(n)}} \\) is their reordering, then for any \\( {k=1,\\ldots,n} \\) the random variable \\( {U_{(k)}} \\) follows the Beta distribution on \\( {[0,1]} \\) with parameter \\( {(k,n-k+1)} \\) and<\/p>\n<p style=\"text-align: center;\">\\[ \\mathbb{P}(S_n\\geq k) =\\mathbb{P}(\\mathbf{1}_{\\{U_1\\leq p\\}}+\\cdots+\\mathbf{1}_{\\{U_n\\leq p\\}}\\geq k) =\\mathbb{P}(U_{(k)}\\leq p). \\]<\/p>\n<p style=\"text-align: justify;\"><b>From discrete to continuous.<\/b> The results remain valid when \\( {a} \\) and \\( {b} \\) are positive reals. This motivates the usage of Beta and Gamma functions instead of factorials. In this case we may use an <b>interval partition<\/b> instead of an urn.<\/p>\n<p style=\"text-align: justify;\"><b>General urns and sampling schemes.<\/b> We generalized the reinforcement mechanism as follows: we fix an integer \\( {r\\geq-1} \\), and, at each drawing, we replace in the urn \\( {1+r} \\) balls of the drawn color.<\/p>\n<ul>\n<li>if \\( {r=1} \\) then we recover the P\u00f3lya urn studied above for which \\( {Y_n} \\) follows the <b>Beta-binomial distribution<\/b>;<\/li>\n<li>if \\( {r=0} \\) then we recover a sampling with replacement and \\( {Y_n} \\) follows then the <b>binomial distribution<\/b>;<\/li>\n<li>if \\( {r=-1} \\) then we recover a sampling without replacement and \\( {Y_n} \\) follows the <b>hypergeometric distribution<\/b>;<\/li>\n<li>more generally, for any \\( {r\\geq-1} \\) and \\( {k\\in\\{0,1,\\ldots,n\\}} \\), we have\n<p style=\"text-align: center;\">\\[ \\mathbb{P}(Y_n=k)=\\frac{a^{(r,k)}b^{(r,k)}}{(a+b)^{(r,k)}} \\quad\\text{where}\\quad c^{(r,k)}:=c(c+r)\\cdots(c+(k-1)r) \\]<\/p>\n<p> and in this case \\( {M_\\infty} \\) follows the Beta distribution of parameter \\( {(a\/r,b\/r)} \\).<\/li>\n<\/ul>\n<p style=\"text-align: justify;\">It is also possible to consider an arbitrary number of colors, which produces a model which includes as a particular case the sampling model of the multivariate hypergeometric distribution. It is also possible to use a replacement matrix as follows: fix a matrix<\/p>\n<p style=\"text-align: center;\">\\[ A={(A_{i,j})}_{1\\leq i,j\\leq k}\\in\\mathcal{M}_{k}(\\mathbb{N}) \\]<\/p>\n<p style=\"text-align: justify;\">and consider the generalized P\u00f3lya urn with \\( {k} \\) colors defined as follows: we draw a ball at random, we denote by \\( {i} \\) its color, and then we replace in the urn \\( {A_{i,j}} \\) balls of color \\( {j} \\) for any \\( {j=1,\\ldots,k} \\). The standard P\u00f3lya urn corresponds to \\( {k=2} \\) and \\( {A=2I_2} \\).<\/p>\n<p style=\"text-align: justify;\"><b>Rubin's theorem on generalized P\u00f3lya urns.<\/b> We consider two non-decreasing functions<\/p>\n<p style=\"text-align: center;\">\\[ S_\\alpha,S_\\beta:\\mathbb{N}\\rightarrow\\mathbb{R}_+ \\]<\/p>\n<p style=\"text-align: justify;\">such that \\( {S_\\alpha(0)&gt;0} \\) and \\( {S_\\beta(0)&gt;0} \\). We construct a random recursive sequence \\( {{(X_n)}_{n\\geq1}} \\) taking its values in \\( {\\{\\alpha,\\beta\\}} \\) as follows: for any \\( {n\\in\\mathbb{N}} \\), conditional on \\( {X_1,\\ldots,X_n} \\), denoting \\( {Y_n=\\sum_{k=1}^n\\mathbf{1}_{\\{X_k=\\alpha\\}}} \\) and \\( {Z_n=n-Y_n} \\), we set<\/p>\n<p style=\"text-align: center;\">\\[ (X_{n+1},Y_{n+1},Z_{n+1})= \\left\\{ \\begin{array}{rl} (\\alpha,Y_n+1,Z_n) & \\text{if } U_{n+1}\\leq \\frac{S_\\alpha(Y_n)}{S_\\alpha(Y_n)+S_\\beta(Z_n)}, \\\\ (\\beta,Y_n,Z_n+1) & \\text{if } U_{n+1}&gt;\\frac{S_\\alpha(Y_n)}{S_\\alpha(Y_n)+S_\\beta(Z_n)}, \\end{array} \\right. \\]<\/p>\n<p style=\"text-align: justify;\">where \\( {{(U_n)}_{n\\geq1}} \\) are iid uniform random variables on \\( {[0,1]} \\). Such a sequence \\( {{(X_n)}_{n\\geq1}} \\) encodes the drawings of a generalized P\u00f3lya urn. If \\( {(S_\\alpha(0),S_\\beta(0))=(A_0,B_0)} \\) is deterministic and \\( {S_\\alpha(n)=S_\\beta(n)=r(n)} \\) for any \\( {n&gt;0} \\), then we say that we have<\/p>\n<ul>\n<li><b>no reinforcement<\/b> when \\( {r} \\) is linear (sampling with replacement);<\/li>\n<li><b>linear reinforcement<\/b> when \\( {r} \\) is affine (standard P\u00f3lya urn);<\/li>\n<li><b>geometric reinforcement<\/b> when \\( {r} \\) is a power function.<\/li>\n<\/ul>\n<p style=\"text-align: justify;\">To study the general case, we introduce the probability \\( {p_\\alpha} \\) (respectively \\( {p_\\beta} \\)) that the sequence \\( {{(X_n)}_{n\\geq1}} \\) contains a finite number of \\( {\\beta} \\) (respectively of \\( {\\alpha} \\)) in other words only \\( {\\alpha} \\) (respectively \\( {\\beta} \\)) for large enough \\( {n} \\). These probabilities are given by<\/p>\n<p style=\"text-align: center;\">\\[ p_\\alpha =\\mathbb{P}\\left(\\cup_n\\cap_{k\\geq n}\\{X_k=\\alpha\\}\\right) \\quad\\text{and}\\quad p_\\beta =\\mathbb{P}\\left(\\cup_n\\cap_{k\\geq n}\\{X_k=\\beta\\}\\right). \\]<\/p>\n<p style=\"text-align: justify;\">We have<\/p>\n<p style=\"text-align: center;\">\\[ p_\\alpha+p_\\beta\\leq1. \\]<\/p>\n<p style=\"text-align: justify;\">The following numbers in \\( {[0,\\infty]} \\) will play an crucial role:<\/p>\n<p style=\"text-align: center;\">\\[ \\varphi_\\alpha=\\sum_{n=0}^\\infty\\frac{1}{S_\\alpha(n)} \\quad\\text{and}\\quad \\varphi_\\beta=\\sum_{n=0}^\\infty\\frac{1}{S_\\beta(n)}. \\]<\/p>\n<p style=\"text-align: justify;\">We are now ready to state <b>Rubin<\/b>'s theorem: \\( {p_\\alpha,p_\\beta} \\) are linked to \\( {\\varphi_\\alpha,\\varphi_\\beta} \\) via<\/p>\n<p style=\"text-align: center;\">\\[ \\begin{array}{c||c|c} & \\varphi_\\beta&lt;\\infty & \\varphi_\\beta=\\infty \\\\\\hline\\hline \\varphi_\\alpha&lt;\\infty & p_\\alpha&gt;0, p_\\beta&gt;0, p_\\alpha+p_\\beta=1 & p_\\alpha=1 (\\Rightarrow p_\\beta=0)\\\\\\hline \\varphi_\\alpha=\\infty & p_\\beta=1 (\\Rightarrow p_\\alpha=0) & p_\\alpha=0, p_\\beta=0\\\\ \\end{array} \\]<\/p>\n<p style=\"text-align: justify;\">In particular the triviality of \\( {p_\\alpha} \\) and \\( {p_\\beta} \\) depends only on the asymptotic behavior of the reinforcement functions \\( {S_\\alpha} \\) and \\( {S_\\beta} \\), and the critical cases are related to the Riemann condition. In the case of absence of reinforcement (\\( {S_\\alpha} \\) and \\( {S_\\beta} \\) are linear) and in the case of linear reinforcement (\\( {S_\\alpha} \\) and \\( {S_\\beta} \\) are affine) we have \\( {\\varphi_\\alpha=\\varphi_\\beta=\\infty} \\) since the harmonic series diverges. As soon as the reinforcement is over-linear, we get \\( {\\varphi_\\alpha&lt;\\infty} \\) and \\( {\\varphi_\\beta&lt;\\infty} \\), and this is the case in particular for the quadratic reinforcement and more generally for the geometric reinforcement.<\/p>\n<p style=\"text-align: justify;\"><b>Proof of Rubin's theorem.<\/b> Let \\( {{(E^\\alpha_n)}_{n\\geq0}} \\) and \\( {{(E^\\beta_n)}_{n\\geq0}} \\) be two independent sequences of independent random variables with, for any \\( {n\\geq0} \\),<\/p>\n<p style=\"text-align: center;\">\\[ \\mathbb{E}(E^\\alpha_n)=\\frac{1}{S_\\alpha(n)} \\quad\\text{and}\\quad \\mathbb{E}(E^\\beta_n)=\\frac{1}{S_\\beta(n)} \\]<\/p>\n<p style=\"text-align: justify;\">Now let us define the random sets<\/p>\n<p style=\"text-align: center;\">\\[ \\mathcal{A}=\\left\\{\\sum_{k=0}^n E^\\alpha_k:n\\geq0\\right\\} \\quad\\text{and}\\quad \\mathcal{B}=\\left\\{\\sum_{k=0}^n E^\\beta_k:n\\geq0\\right\\} \\quad\\text{and}\\quad \\mathcal{G}=\\mathcal{A}\\cup\\mathcal{B}. \\]<\/p>\n<p style=\"text-align: justify;\">Let \\( {\\xi_0&lt;\\xi_1&lt;\\cdots} \\) be the reordering of the elements of \\( {\\mathcal{G}} \\). We consider the random sequence \\( {{(X_n')}_{n\\geq1}} \\) taking values in \\( {\\{\\alpha,\\beta\\}} \\) defined by \\( {X_n'=\\alpha} \\) if \\( {\\xi_{n-1}\\in\\mathcal{A}} \\) and \\( {X_n'=\\beta} \\) if \\( {\\xi_{n-1}\\in\\mathcal{B}} \\). The sequences \\( {{(X_n')}_{n\\geq1}} \\) and \\( {{(X_n)}_{n\\geq1}} \\) have same distribution, and this follows from the properties of exponential distributions (including the lack of memory). Namely, let us examine the equality in distribution of \\( {X_1} \\) and \\( {X'_1} \\). If \\( {U} \\) and \\( {V} \\) are two independent exponential random variables with respective means \\( {1\/u} \\) and \\( {1\/v} \\) then \\( {\\mathbb{P}(U&lt;V)=u\/(u+v)} \\) and \\( {\\mathbb{P}(V&lt;U)=v\/(u+v)} \\), which gives<\/p>\n<p style=\"text-align: center;\">\\[ \\mathbb{P}(X_1'=\\alpha) =\\mathbb{P}(\\xi_0\\in\\mathcal{A}) =\\mathbb{P}(E^\\alpha_0\\leq E^\\beta_0) =\\frac{S_\\alpha(0)}{S_\\alpha(0)+S_\\beta(0)} =\\mathbb{P}(X_1=\\alpha). \\]<\/p>\n<p style=\"text-align: justify;\">The same idea gives (tedious!) the equality in distribution of \\( {{(X_n)}_{n\\geq1}} \\) and \\( {{(X_n')}_{n\\geq1}} \\). Next, by the zero-one law for the sum of independent exponential random variables,<\/p>\n<p style=\"text-align: center;\">\\[ p:=\\mathbb{P}\\left(\\sum_{n=0}^\\infty E^\\alpha_n&lt;\\infty\\right)\\in\\{0,1\\}, \\]<\/p>\n<p style=\"text-align: justify;\">with \\( {p=1} \\) if and only if \\( {\\varphi_\\alpha&lt;\\infty} \\). The same property holds for the sequence \\( {{(E^\\beta_n)}_{n\\geq0}} \\) with \\( {\\varphi_\\beta} \\). On the other hand we have the formulas<\/p>\n<p style=\"text-align: center;\">\\[ p_\\beta =\\mathbb{P}\\left(\\sum_{n=0}^\\infty E^\\alpha_n&lt;\\sum_{n=0}^\\infty E^\\beta_n\\right) \\quad\\text{and}\\quad p_\\alpha =\\mathbb{P}\\left(\\sum_{n=0}^\\infty E^\\beta_n&lt;\\sum_{n=0}^\\infty E^\\alpha_n\\right). \\]<\/p>\n<p style=\"text-align: justify;\"><b>Notes and further reading.<\/b> The content of this post is extracted and translated from a <a href=\"\/docs\/RMA\/\">book<\/a> written with Florent Malrieu in French. P\u00f3lya urns are named after the Hungarian mathematician George P\u00f3lya. We refer to the books by <a href=\"http:\/\/www.ams.org\/mathscinet-getitem?mr=MR2435823\">Hosam Mahmoud<\/a>, by <a href= \"http:\/\/www.ams.org\/mathscinet-getitem?mr=MR0488211\">Norman Johnson and Samuel Kotz<\/a>, and to the survey articles by <a href= \"http:\/\/www.ams.org\/mathscinet-getitem?mr=MR1456736\">Samuel Kotz and Narayanaswamy Balakrishnan<\/a>, and by <a href= \"http:\/\/www.ams.org\/mathscinet-getitem?mr=MR2282181\">Robin Pemantle<\/a>. Rubin's theorem can be found in the appendix of a paper by <a href= \"http:\/\/www.ams.org\/mathscinet-getitem?mr=MR1030727\">Burgess Davis<\/a> (suggested to us by Amine Asselah). P\u00f3lya urns are fundamental models of <b>reinforcement<\/b>. They play an important role in learning theory. They are also connected to a class of stochastic algorithms such as the stochastic approximation algorithm of Robbins-Monro, linked with the theory of dynamical systems and attractors. The concept of martingale is a central tool in the stochastic approach of game theory and strategies.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>This post is devoted to P&oacute;lya urns, probably the simplest stochastic model of reinforcement. At time \\( {n=0} \\), we prepare an urn containing \\(&#8230;<\/p>\n<div class=\"more-link-wrapper\"><a class=\"more-link\" href=\"https:\/\/djalil.chafai.net\/blog\/2015\/11\/30\/back-to-basics-polya-urns\/\">Continue reading<span class=\"screen-reader-text\">Back to basics - P\u00f3lya urns<\/span><\/a><\/div>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"iawp_total_views":1450},"categories":[1],"tags":[],"_links":{"self":[{"href":"https:\/\/djalil.chafai.net\/blog\/wp-json\/wp\/v2\/posts\/8576"}],"collection":[{"href":"https:\/\/djalil.chafai.net\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/djalil.chafai.net\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/djalil.chafai.net\/blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/djalil.chafai.net\/blog\/wp-json\/wp\/v2\/comments?post=8576"}],"version-history":[{"count":6,"href":"https:\/\/djalil.chafai.net\/blog\/wp-json\/wp\/v2\/posts\/8576\/revisions"}],"predecessor-version":[{"id":8583,"href":"https:\/\/djalil.chafai.net\/blog\/wp-json\/wp\/v2\/posts\/8576\/revisions\/8583"}],"wp:attachment":[{"href":"https:\/\/djalil.chafai.net\/blog\/wp-json\/wp\/v2\/media?parent=8576"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/djalil.chafai.net\/blog\/wp-json\/wp\/v2\/categories?post=8576"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/djalil.chafai.net\/blog\/wp-json\/wp\/v2\/tags?post=8576"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}