0% found this document useful (0 votes)
37 views

Statistics Homework Help

The document provides 7 statistics homework problems and their solutions. It begins by providing contact information for statistics homework help. The problems cover topics like probability distribution functions, transforms of random variables, expectations, variances, and convolving probability mass functions. The solutions provide detailed steps and explanations for finding PDFs, parameters, probabilities, expectations, variances, and the probability of a sum being a certain value by convolving PMFs.
Copyright
© © All Rights Reserved
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
37 views

Statistics Homework Help

The document provides 7 statistics homework problems and their solutions. It begins by providing contact information for statistics homework help. The problems cover topics like probability distribution functions, transforms of random variables, expectations, variances, and convolving probability mass functions. The solutions provide detailed steps and explanations for finding PDFs, parameters, probabilities, expectations, variances, and the probability of a sum being a certain value by convolving PMFs.
Copyright
© © All Rights Reserved
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
You are on page 1/ 14

For any Homework related queries, call us at :- +1 678 648 4277

You can mail us at :- [email protected] or


reach us at :-https://www.statisticshomeworksolver.com/

Statistics Homework Help


Problem
1. Suppose that

What is the PDF of X?

2. Find the transform of the random variable X with density function:

where p is a constant with 0 ≤ p ≤ 1.

3. Consider random variable Z with transform

(a) Find the numerical value for the parameter a.


(b) Find P(Z ≥ 0.5).
(c) Find E[Z] by using the probability distribution of Z.
(d) Find E[Z] by using the transform of Z and without explicity using the probability
distribution of Z.
(e) Find var(Z) by using the probability distribution of Z.
(f) Find var(Z) by using the transform of Z and without explicity using the probability
distribution of Z.
4. A coin is tossed repeatedly, heads appearing with probability q on each toss. Let
random variable T denote the number of tosses when a run of n consecutive heads
has appeared for the first time.
(a) Show that the PMF for T can be expressed as

(b) Determine the transform MT (s) associated with random variable T.


(c) Compute E[T], the expectation of random variable T.

5. This problem is based on an example covered in Monday’s lecture (lecture


11). Let X and N be two independent normal random variables. Say X ∼ N(0,
σ2 x ) and N ∼ N(0, σ2 n ). Let Y = X + N. In lecture we saw that Y is also
normal. The lecture slides also prove that the conditional PDF fY |X(y|x) is
normal.

Prove that for every value of y, the conditional density fX|Y (x|y) is normal.
6. Four fair 6-sided dice are rolled independently of each other. Let X1 be
the sum of the numbers on the first and second dice, and X2 be the sum of
the numbers on the third and fourth dice. Convolve the PMFs of the
random variables X1 and X2 to find the probability that the outcomes of
the four dice rolls sum to 8.
7. Consider two independent random variables X and Y . Let fX(x) = 1 − x/2 for
x ∈ [0, 2] and 0 otherwise. Let fY (y) = 2 − 2y for y ∈ [0, 1] and 0 otherwise.
Give the PDF of W = X + Y .

G1† . Let X1, X2, . . ., Xn be drawn i.i.d from the uniform distribution on [0, 1].
Let Y be the minimum of the Xi , and let Z be the maximum of the Xi . Let W =
Y + Z. Compute fW (w), and prove that for all ǫ > 0, limn→∞ P(|W − 1| > ǫ) =
0. Thus, for large n, with very high probability W is close to 1.
Solutions
1. X is the mixture of two exponential random variables with parameters 1 and 3,
which are selected with probability 1/3 and 2/3, respectively. Hence, the PDF of X is

2. X is a mixture of two exponential random variables, one with parameter λ


and one with parameter µ. We select the exponential with parameter λ with
probability p, so the transform

3. (a) The definition of the transform is

(b) We approach this problem by first finding the PDF of Z using partial fraction
expansion:
4. (a) Since it is impossible to get a run of n heads with fewer than n
tosses, it is clear that pT (k) = 0 for k < n. In addition, the probability of
getting n heads in n tosses is q n so pT (n) = q n . Lastly, for k ≥ n + 1, we
have T = k if there is no run of n heads in the first k − n − 1 tosses,
followed by a tail, followed by a run of n heads, so

(b) We use the PMF we obtained in the previous part to compute the
moment generating function. Thus,

We observe that the set of pairs {(i, k) | k ≥ n + 1, i ≥ k − n} is equal to the


set of pairs {(i, k) | i ≥ 1, n + 1 ≤ k ≤ i + n}, so by reversing the order of the
summations, we have
Note that for n = 1, this equation reduces to E[T] = 1/q, which is the mean of a
geometrically-distributed random variable, as expected.

5. We calculate fX|Y (x|y) using the definition of a conditional density. To


find the density of Y , recall that Y is normal, so the mean and variance
completely specify fY (y). Y = X + N, so E[Y ] = E[X] + E[N] = 0 + 0 = 0.
Because X and N are independent, var(Y ) = var(X) + var(N) = σ 2 x + σ 2 n .
So,

We can simplify the exponent as follows.


Thus, we obtain

Looking at this formula, we see that the conditional density is normal


with mean
6. Let Ri be the number rolled on the i th die. Since each number is equally likely
to rolled, the PMF of each Ri is uniformly distributed from 1 to 6. The PMF of X1
is obtained by convolving the PMFs of R1 and R2. Similarly, the PMF of X2 is
obtained by convolving the PMFs of R3 and R4. X1 and X2 take on values from 2
to 12 and are independent and identically distributed random variables. The
PMF of either one is given by

Note that the sum X1 + X2 takes on values from 4 to 24. The discrete
convolution formula tells us that for n from 4 to 24:
and thus we find the desired probability is

7. The PDF for X and Y are as follows,

Because X and Y are independent and W = X + Y , the pdf of W, fW (w), can


be written as the convolution of fX(x) and fY (y):

There are five ranges for w: 1. w ≤ 0


G1† . To compute fW (w), we will start by computing the joint PDF fY,Z(y, z).
Computing the joint density is quite simple. Define the joint CDF FY,Z(y, z) =
P(Y ≤ y, Z ≤ z). Now, FZ(z) = P(Z ≤ z) = z n , because the maximum is less than z if
and only if every one of the Xi is less than z, and all the Xi ’s are independent.
We can also compute P(y ≤ Y, Z ≤ Z) = (z − y) n because the minimum is greater
than y and the maximum is less that z if and only if every Xi falls between y
and z. Subtraction gives

Now, we find the joint PDF by differentiating, which gives fY,Z(y, z) = n(n−1)
(z−y) n−2 , 0 ≤ y ≤ z ≤ 1. Because Y and Z are not independent, convolving the
individual densities for Y and Z will not give us the density for W. Instead, we
must calculate the CDF FW (w) by integrating PY,Z(y, z) over the appropriate
region. We consider the cases w ≤ 1 and w > 1 separately.

When w ≤ 1, we need to compute

You might also like