Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

1. Let XiX_i, i=1,2i = 1,2 be independent random variables such that XiX_i has the exponential (λi)(\lambda_i) distribution. Suppose λ1λ2\lambda_1 \ne \lambda_2.

(a) Use the convolution formula to find the density of the sum S=X1+X2S = X_1 + X_2.

(b) Show by algebra that the expression you got for the density in Part (a) is non-negative for all positive λ1λ2\lambda_1 \ne \lambda_2.

(c) For i=1,2i=1, 2, let fif_i denote the exponential density of XiX_i. Show that the density you got in part (a) is equal to c1f1+c2f2c_1f_1 + c_2f_2 for two constants c1c_1 and c2c_2 such that c1+c2=1c_1+c_2 = 1. Are c1c_1 and c2c_2 both positive?

2. A coin lands heads with probability pp. Let XX be the number of tosses till the first head appears and let YY be the number of tails before the first head.

(a) Find the moment generating function of XX.

(b) Use the answer to (a) to find E(X)E(X). Note that by now you have found E(X)E(X) in several ways: by the tail sum formula, by conditioning on the first toss, by the pgf, and now by the mgf.

(c) Use the answer to (a) to find the moment generating function of YY.

3. Recognizing distributions saves a lot of work. See if you can do the following without explicit integration or differentation.

(a) WW is a normal (0,σ2)(0, \sigma^2) random variable. For n=0,1,2,3n = 0, 1, 2, 3, and 4, find E(Wn)E(W^n).

(b) XX and YY are i.i.d. with moment generating function M(t)=et+t2M(t) = e^{t + t^2}, <t<-\infty < t < \infty. What is the distribution of (XY)2(X-Y)^2?

4. Markov and Chebyshev Bounds on the Poisson-Binomial Upper Tail

For j1j \ge 1 let IjI_j be independent indicators such that P(Ij=1)=pjP(I_j = 1) = p_j. Let X=I1+I2++InX = I_1 + I_2 + \ldots + I_n. Then XX is the number of successes in nn independent trials that are not necessarily identically distributed.

We say that XX has the Poisson-binomial distribution with parameters p1,p2,,pnp_1, p_2, \ldots, p_n. As you saw in lab, the binomial is the special case when all the pjp_j’s are equal.

These distributions arise in statistical learning theory, the theory of randomized algorithms, and other areas.

Let E(X)=μE(X) = \mu. For c>0c > 0, you are going to find an upper bound on P(X(1+c)μ)P(X \ge (1+c)\mu). That’s the chance that XX exceeds its mean by some percent.

In the special case of the binomial, μ=np\mu = np and so P(X(1+c)μ)P(X \ge (1+c)\mu) can be rewritten as P(Xnpcp)P(\frac{X}{n} - p \ge cp). That’s the chance that the sample proportion exceeds pp by some percent.

(a) Find μ=E(X)\mu = E(X) and σ2=Var(X)\sigma^2 = Var(X) in terms of p1,p2,,pnp_1, p_2, \ldots, p_n.

(b) Find Markov’s bound on P(X(1+c)μ)P(X \ge (1+c)\mu).

(c) Find Chebyshev’s bound on P(X(1+c)μ)P(X \ge (1+c)\mu) in terms of μ\mu and σ\sigma.

(d) If all the pjp_j’s are equal to pp, what is the value of the bound in (c)?

5. Chernoff Bound on Poisson-Binomial Upper Tail

This exercise continues the previous one and uses the same notation.

(a) Show that the mgf of IjI_j is given by MIj(t)=1+pj(et1)M_{I_j}(t) = 1 + p_j(e^t - 1) for all tt.

(b) Use (a) to derive an expression for MX(t)M_X(t), the mgf of XX evaluated at tt.

(c) An useful exponential bound is that ex1+xe^x \ge 1 + x for all xx. You don’t have to show it but please look at the graphs. Use the fact to show that MX(t)exp(μ(et1))M_X(t) \le \exp\big(\mu(e^t -1)\big) for all tt. Notice that the right hand side is the mgf of a Poisson random variable that has the same mean as XX.

(d) Use Chernoff’s method and the bound in (c) to show that

P(X(1+c)μ)  (exp(c)(1+c)1+c)μP\big(X \ge (1+c)\mu\big) ~ \le ~ \Big( \frac{\exp(c)}{ (1+c)^{1+c}} \Big)^\mu

Remember that μ=np\mu = np when all the pjp_j’s are equal. If g(c)=exp(c)/(1+c)1+cg(c) = \exp(c)/(1+c)^{1+c} is small then the bound above will decrease exponentially as nn gets large. That is the focus of the next exercise.

6. Simplified Chernoff Bounds on Poisson-Binomial Upper Tail

The bound in the previous exercise is a bit complicated. Often, simpler versions are used because they are good enough even though they are weaker.

(a) It is not hard to show that log(1+c)2c2+c\log(1+c) \ge \frac{2c}{2+c} for c>0c > 0. You don’t have to show it but please look at the graphs. Use the fact to show that c(1+c)log(1+c)c22+cc - (1+c)\log(1+c) \le -\frac{c^2}{2+c} for c>0c > 0.

(b) Show that if XX has a Poisson-binomial distribution with mean μ\mu then

P(X(1+c)μ)  exp(c22+cμ),    c>0P\big( X \ge (1+c)\mu\big) ~ \le ~ \exp\big(-\frac{c^2}{2+c}\mu\big), ~~~~ c > 0

(c) A simpler but weaker version of the bound in (b) is also often used. Show that

P(X(1+c)μ)  exp(c23μ),    c(0,1)P\big( X \ge (1+c)\mu\big) ~ \le ~ \exp\big(-\frac{c^2}{3}\mu\big), ~~~~ c \in (0, 1)

7. Let NN be a non-negative integer valued random variable, and let X,X1,X2,X, X_1, X_2, \ldots be i.i.d. and independent of NN. As before, define the random sum SS by

S  =  0  if N=0=  X1+X2++Xn  if N=n>0\begin{align*} S ~~&=~~0~~ \mbox{if}~N=0 \\ &=~~ X_1 + X_2 + \cdots + X_n ~~ \mbox{if}~N = n > 0 \end{align*}

(a) Let MM be our usual notation for moment generating functions. By conditioning on NN, show that

MS(t)  =  MN(logMX(t))M_S(t) ~~=~~ M_N\big(\log M_X(t) \big)

You can assume that all the quantities above are well defined.

(b) Let NN have the geometric (p)(p) distribution on {1,2,3,}\{1, 2, 3, \ldots \}. Find the mgf of NN. This doesn’t use Part (a).

(c) Let X1,X2,X_1, X_2, \ldots be i.i.d. exponential (λ)(\lambda) variables and let NN be geometric as in Part (b). Use the results of Parts (a) and (b) to identify the distribution of SS.

8. Let U1,U2,U_1, U_2, \ldots be i.i.d. uniform on (0,1)(0, 1). For n1n \ge 1, let Sn=i=1nUiS_n = \sum_{i=1}^n U_i.

Let fSnf_{S_n} be the density of SnS_n. The formula for fSnf_{S_n} is piecewise polynomial on the possible values (0,n)(0, n). In this problem we will just focus on the density on the interval (0,1)(0, 1) and discover a nice consequence.

(a) For 0<x<10 < x < 1, find fS2(x)f_{S_2}(x).

(b) Use Part (a) and the convolution formula to find fS3(x)f_{S_3}(x) for 0<x<10 < x < 1.

(c) Guess a formula for fSn(x)f_{S_n}(x) for 0<x<10 < x < 1 and prove it by induction.

(d) Use Part (c) to find P(Sn<1)P(S_n < 1).

(e) Let N=min{n:Sn>1}N = \min\{n: S_n > 1\}. Use Part (d) to find E(N)E(N).