UMD Probability Qualifying Exams/Jan2015Probability
Problem 1
edit
A die is rolled infinitely many times. (a) Find the expected number of combinations (6,6) among the first 1000 rolls. Find the expected number of combinations (1,2) among the first 1000 rolls. (b) What is larger and by how much: the expected number of rolls till the first occurrence of (6,6) or the expected number of rolls till the first occurrence of (1,2)? (c) What is the probability that (6,6) occurs before (1,2)? |
Solution
edit(a) Let be the value of the nth roll. The number of (6,6)'s in the first 1000 rolls is given by
By linearity of expectation,
Similarly, the number of (1,2)'s in the first 1000 rolls is 999/36.
(b) For the general approach to solve such problems see [1]. The expected waiting time till the first (6,6) is 6+36=42, while the expected waiting time till the first (1,2) is 0+36=36.
The argument is as follows. Suppose we continue rolling dice until we hit the desired pattern. At each time n a new gambler begins a game with 1 coin. On a gambler's i-th bet, he bets the i-th character of the pattern will occur. If the gambler loses a bet, he forfeits all his coins. Otherwise he has 6 times as many as he started with. This is a martingale, so at the end of the game, by the optional stopping theorem the expected total sum of all gamblers' coins is equal to the number of gamblers who joined the game, i.e. the expected waiting time for the pattern.
For the pattern (6,6), the second to last gambler has 6 coins and the last gambler has 36 coins. The others have 0.
For the pattern (1,2), the last gambler has 36 coins. The others have 0.
- Note the optional stopping theorem applies because the expected value of the stopping time is finite and the martingale increments are bounded.
- A more direct way to do this problem is with absorbing Markov chains
(c) Since the patterns do not overlap, the odds that (6,6) occurs first are given by the expected waiting times, 36 : 42. That is the probability (6,6) occurs first is 36/(42+36)=6/13
For patterns which do overlap, the analysis is slightly more complicated, see Corollary 3.2 in the Shuo-Yen Li paper [1].
[1] A Martingale Approach to the Study of Occurrence of Sequence Patterns in Repeat Experiments, Shuo-Yen Li, Annals of Probability
Problem 2
edit
We'll say that a random variable is discrete if the measure it induces is discrete. A random variable is absolutely continuous if the measure it induces is absolutely continuous. Prove or disprove. (a) A sum of two discrete random variables is always a discrete random variable (b) A sum of a discrete and an absolutely continuous random variable is always an absolutely continuous random variable. (c) A sum of two absolutely continuous random variables is always an absolutely continuous random variable. (d) A sum of two independent discrete random variables is always a discrete random variable. (e) A sum of a discrete and an absolutely continuous random variable that are independent is always an absolutely continuous random variable. (f) A sum of two independent absolutely continuous random variables is always an absolutely continuous random variable. |
Solution
edit(a) True. A random variable is discrete iff its range is countable. The range of the sum of two random variables A, B is a subset of the sum-set range(A)+range(B). The sum-set is countable if range(A), and range(B) are countable.
(b). True. A random variable X is absolutely continuous iff for any Lebesgue-null set E. Let Y be a discrete random variable. since a countable sum of Lebesgue-null sets is Lebesgue-null.
(c). False. For any absolutely continuous random variable X, -X is also absolutely continuous, and X + (-X) =0 which is a discrete random variable.
(d). True, by (a)
(e). True, by (b)
(f). True, the distribution function is given by the convolution theorem.
Problem 3
edit
Suppose that are independent random variables uniformly distributed on the intervals respectively, where are positive. Let (a). Prove that tend to zero almost surely if . (b). Prove that do not have a limit if . |
Solution
edit(a). Let where is uniformly distributed in .
By the strong law off large numbers,
converges to -1 almost surely.
.
So converges to a.s. So converges to 0 a.s.
(b). In this case
So converges to a.s. So converges to almost surely. However the sign of changes infinitely often a.s. (by the second Borel-Cantelli lemma), so has no limit almost surely.
Problem 4
edit
Let , be a positive continuous martingale that converges to zero almost surely as . (a) Prove that for each . (b) Let be a standard one-dimensional Brownian motion. Prove that for each the random variable
has exponential distribution with parameter . |
Solution
edit(a) Let be the stopping time given by where u is the first time . By the optional stopping theorem,
Since converges to zero a.s., taking the limit as goes to infinity we have if .
Clearly, if , .
(b) Wald's martingale is a well-known martingale.
By (a),
Problem 5
edit
Suppose that is a bounded random variable on where X is a metric space, is its Borel -algebra, and P is a probability measure. Suppose that for all . Prove that there is a bounded continuous function on X such that . |
Solution
editSuppose for all x. By Lusin's theorem, for all , there exists a continuous such that and for all x.
Taking and the result follows.
Problem 6
edit
Suppose that n points are independently and uniformly distributed on a unit circle. Let be the smallest of the distances between the points. Prove that tends to zero in probability as . |
Solution
editLet . Partition the circle into non-intersecting arcs of equal length. Assume The probability that n points are selected at random and at least two fall into the same arc is given by
Taking C sufficiently small, we can ensure
If two points fall into the same arc, the distance between them is less than
and
which tends to 0. Since may be taken to be arbitrarily small, we get that tends to zero in probability.