icon nav

Probability (I)

Discrete Random Variables
  • Probability Theory
  • Mathematics

Discrete Random Variables

Random Variables

Definition: In probability theory, the sample space of an experiment is the set of all possible outcomes of that experiment. Here, we will refer to the sameple space as S, though you may also see it denoted as \Omega of U (for universal set).



For example, if our experiment is flipping a coin once, our sample space is S = \{H, T\}.



Definition: A random variable (RV), denoted X, is a function X: S \rightarrow \mathbb{R} The range of this function is denoted R_X = X(S).



Definition: X is a discrete random variable (DRV) if its range is countable.



Definition: Let X be a discrete random variable with R_X = \{x_1, x_2, x_3,...\}. The function P_X : R_X \rightarrow [0,1] defined by P_X(x_k) = P(X=x_k) = P \Big( \{ s \in S : X(s) = x_k \} \Big) \quad k=1,2,3,... is called the probability mass function (PMF) of X.



In general, the probability mass function P_X(x) will look like P_X(x) =  \begin{cases}<br>P(X=x) &  x \in R_X \\[10pt]<br>0 &  \text{otherwise}<br>\end{cases}

Example 3.3 We toss a fair coin twice and let our random variable X be the number of heads. Find R_X and the probability mass function of X.



Solution: First, we can form the sample space from all possible results of flipping a coin twice: S = \{HH, HT, TH, TT\} Flipping a coin twice, we will either get 0 heads, 1 head, or 2 heads, making our range R_X = \{0,1,2\} To find the probability mass function, we need to find the probability that each of these options will occur. P_X(0) = P(X=0) = P(TT) = \frac{1}{4} P_X(1) = P(X=1) = P(HT) + P(TH) = \frac{1}{2} P_X(2) = P(X=2) = P(HH) = \frac{1}{4} All together, our probability mass function will be

P_X(x) =  \begin{cases}<br>\frac{1}{4} &  x = 0, 2 \\[10pt]<br>\frac{1}{2}  & x = 1\\[10pt]<br>0 &  \text{otherwise}<br>\end{cases}

 

Example 3.4: Now say we have an unfair coin for which P(H) = p, 0<p<1. We toss the coin repeatedly until we get our first heads. Let Y be the number of coin tosses. Find the PMF of Y.



Solution: Here, our range is all natural numbers. R_Y = \{1,2,3,... \} To find the distribution of Y, we need to find P_Y(k) = P(Y=k), \quad k \in \mathbb{N} Given an arbitrary k, we have P_Y(k) = P(Y=k) = P(T\underset{\text{k-1 times}}{\ldots}TH) This represents the probability that we get k-1 tails followed by 1 head, or (1-p)^{k-1} p.



So our probability mass function is

P_Y(y) =  \begin{cases}<br>(1-p)^{y-1} p &  y \in \mathbb{N} \\[10pt]<br>0 & y \notin \mathbb{N}<br>\end{cases}

Properties of the Probability Mass Function:



(a) 0 \leq P_X(x) \leq 1 \forall x



(b) \displaystyle \sum_{x \in R_X} P_X(x) = 1



(c) \displaystyle \forall A \subset R_X, \, P(X \in A) = \sum_{x \in A} P_X(x)



Example 3.5 For the random variable Y from Example 3.4:



(a) Show that \displaystyle \sum_{x \in R_Y} P_Y(y) = 1



(b) Letting p= \frac{1}{2}, find P(2 \leq Y < 5)



Solution:

(a) \sum_{x \in R_Y} P_Y(y) = \sum_{k=1}^{\infty} (1-p)^{k-1} p = \left(\frac{1}{1-(1-p)}\right) p = 1

(b) P(2 \leq Y < 5) = \sum_{k=2}^4 (1-p)^{k-1} p = \sum_{k=2}^4 \left(\frac{1}{2}\right)^{k-1} \frac{1}{2} = \sum_{k=2}^4 \frac{1}{2^k} =  \frac{7}{16}

Definition 3.2: Let X and Y be two random variables. We say that X and Y are independent if P(X=x,Y=y) = P(X=x)P(Y=y) \quad \forall (x,y) \in \mathbb{R}^2 or, in general, P(X \in A,Y \in B) = P(X \in A)P(Y \in B) \quad \forall A,B

Example 3.6: Say we toss and coin twice and let X be the number of heads. Then, we toss the coin another two times and let Y be the number of heads from these two tosses. Find P(X <2, Y>1).



Solution: Since X and Y are the results of separate, independent coin tosses, we can say that the two random variables are independent. Hence: P(X <2, Y>1) = P(X<2)P(Y>1) = \Big((P_X(0) + P(_X(1) \Big) P_Y(2) = = \left(\frac{1}{4} + \frac{1}{2}  \right) \frac{1}{4} = \frac{3}{16}

Definition 3.3: Consider n discrete random variables X_1, X_2, ..., X_n. We can say that these variables are independent if P(X_1=  x_1, X_2= x_2, ... , X_n=x_n) = \prod_{k=1}^n P(X_k = x_k) \quad \forall (x_1, x_2, ..., x_n) \in \mathbb{R}^n

Definition: A probability distribution is a function which gives the probabilities of which different possible outcomes of an experiment occur.



Certain probability distributions are significant enough that they are given specific names. One example is the well-known normal distribution. Some important probability distributions for discrete random variables are defined the following section.

The Bernoulli Distribution

Definition: A random variable X is said to be a Bernoulli random variable with parameter p, written as X \sim \text{Bernoulli}(p), if its probability mass function if given by

P_X(x) =  \begin{cases}<br>p &  x=1 \\[5pt]<br>1-p &  x=0 \\[5pt]<br>0 &  \text{otherwise}<br>\end{cases}

where 0 < p <1.



Definition: The indicator random variable I_A for some event A is defined by

I_A =  \begin{cases}<br>1 &  \text{the event occurs} \\[5pt]<br>0 &  \text{otherwise}<br>\end{cases}

I_A is an example of a random variable which has a Bernoulli distribution with parameter p = P(A), so we can write I_A \sim \text{Bernoulli}\big(P(A)\big)

The Geometric Distribution

Suppose we have a coin where P(H) = p. We toss the coin until we get our first heads and define X as the number of coin tosses. This X is said to have a geometric distribution with parameter p, R_X = \mathbb{N}, and P_X(k) = (1-p)^{k-1}p.



Definition 3.5: A random variable X is said to be a geometric random variable with parameter p, written as X \sim \text{Geometric}(p), if its probability mass function is given by

P_X(k) =  \begin{cases}<br>(1-p)^{k-1} p &  k \in \mathbb{N} \\[10pt]<br>0 & k \notin \mathbb{N}<br>\end{cases}

where 0 < p <1.

The Binomial Distribution

Suppose we have a coin with P(H) = p. We toss the coin n times and define X to be the total number of heads. X, in this case, is said to have a binomial distribution with parameters n and p.



Definition: A random variable X is said to be a Binomial Random Variable with parameters n and p, written X \sim \text{Binomial}(n,p) if its probability mass function is given by

P_X(k) =  \begin{cases}<br>\displaystyle \binom{n}{k} p^k (1-p)^{n-k}&  k = 0, 1,2,...,n \\[15pt]<br>0 & \text{otherwise}<br>\end{cases}



where 0 < p <1 and \displaystyle \binom{n}{k} is the binomial coefficient \displaystyle \binom{n}{k} = \frac{n!}{k!(n-k)!}.



Lemma 3.1: While the Bernoulli distribution models the probabilities for a single trial, the Binomial distribution considers any number of independent trials. Hence, we can think of the Binomial distribution above as the sum of n Bernoulli random variables.



In other words, if X_1, X_2, ..., X_n are independent Bernoulli random variables, then the random variable X defined as X = X_1 + X_2 + ... + X_n has a Binomial(n,p) distribution.



Example: Say we flip a coin with P(H)=p three times. Let X be the number of heads we observe. Show that X \sim \text{Binomial}(3,p).



Solution: We can break X up into three random variables. Let X_1 represent the results of the first coin toss, where

X_1 =  \begin{cases}<br>1 &  \text{we get heads} \\[5pt]<br>0 &  \text{we get tails}<br>\end{cases}

We can see that X_1 \sim \text{Bernoulli}(p).



Now let X_2 and X_3 represent the second and third coin tosses the same way. These are all Bernoulli random variables and are independent from one another, each with the probability mass function

P_{X_i}(x) =  \begin{cases}<br>p &  x=1 \\[5pt]<br>1-p &  x=0 \\[5pt]<br>0 &  \text{otherwise}<br>\end{cases}

Now we can use X = X_1 + X_2 + X_3 to find the probability mass function of X:

P_X(0) = P(X_1=0, X_2=0, X_3=0)= P(X_1=0)  P(X_2=0)  P(X_3=0)= (1-p)^3

P_X(1) = P(X_1=1, X_2=0, X_3=0) + P(X_1=0, X_2=1, X_3=0) +P(X_1=0, X_2=0, X_3=1) = 3p(1-p)^2

P_X(2) = P(X_1=1, X_2=1, X_3=0) + P(X_1=0, X_2=1, X_3=1) +P(X_1=1, X_2=0, X_3=1) = 3 p^2 (1-p)

P_X(3) = P(x=0) =P(X_1=1)  P(X_2=1)  P(X_3=1) = p^3

So we have

P_{X}(x) =  \begin{cases}<br>(1-p)^3 &  x=0 \\[5pt]<br>3p(1-p)^2 &  x=1 \\[5pt]<br>3p^2(1-p) &  x=2 \\[5pt]<br>p^3 &  x=3 \\[5pt]<br>0 &  \text{otherwise}<br>\end{cases}

or, equivalently,

P_X(k) =  \begin{cases}<br>\displaystyle \binom{n}{k} p^k (1-p)^{n-k}&  k = 0, 1,2,3 \\[15pt]<br>0 & \text{otherwise}<br>\end{cases} Therefore X \sim \text{Binomial}(3,p).



Example 3.7: Say we have two independent random variables X \sim \text{Binomial}(n,p) and Y \sim \text{Binomial}(m,p). Define a new random variable Z = X + Y. Find the probability mass function of Z.



Solution: We can write X \sim \text{Binomial}(n,p) as the sum of n independent random variables X = X_1 + X_2 + ... + X_n where each X_i \sim \text{Bernoulli}(p).



We can write Y the same way as the sum of m independent Bernoulli random variables Y = Y_1 + Y_2 + ... Y_m So adding these together we have that Z = X + Y = X_1 + X_2 + ... + X_n + Y_1 + Y_2 + ... Y_m making Z the sum of m+n independent Bernoulli random variables, all with parameter p. Hence Z \sim \text{Binomial} (m+n, p), giving us the probability mass function P_Z(k) =  \begin{cases}<br>\displaystyle \binom{m+n}{k} p^k (1-p)^{m+n-k}&  k = 0, 1,2,...,m+n \\[15pt]<br>0 & \text{otherwise}<br>\end{cases}

The Negative Binomial (Pascal) Distribution

Let’s once again suppose we have a coin with P(H) = p. We toss the coin until we observe m heads and define X to be the total number of coin tosses. This X is said to have a negative binomial or Pascal distribution with parameters m and p.



Why is this called the negative binomial distribution? While the binomial distribution models the number of successes, the negative binomial distribution models the number of failures.



The parameter m represents the smallest number X can take, the case where every coin comes up heads. So our range is R_X = \{m, m+1, m+2, ... \} To find the probability mass function of X, define A = \{X=k\} = BC, where B is the event that we observe m=1 heads in the first k-1 trials and C is the event that we get a heads in the kth trial. So we can write.

P(A) = P(BC) = P(B)P(C) = \binom{k-1}{m-1} p^{m-1} (1-p)^{(k-1)-(m-1)} p = \binom{k-1}{m-1} p^m (1-p)^{k-m}

Definition: A random variable X is said to be a Pascal random variable, written X \sim \text{Pascal}(m,p) if its probability density function is given by P_X(k) =  \begin{cases}<br>\displaystyle \binom{k-1}{m-1} p^m (1-p)^{k-m}&  k = m, m+1, m+2, ... \\[15pt]<br>0 & \text{otherwise}<br>\end{cases} where 0 < p <1.

The Hypergeometric Distribution

Now suppose we have a bag that contains b blue marbles and r red marbles. We choose k \leq b+r marbles at random without replacement. Let X be the number of blue marbles in our sample. By definition, we must have X \leq \min(b,k) and our number of red marbles must be less than or equal to r. We can conclude from this that X \geq \max(0,k-r).



Therefore we have that R_X = \{\max(0,k-r), ..., \min(k,b) \} This is an example of a hypergeometric distribution.



textbfDefinition: A random variable X is said to be a hypergeometric random variable with parameters b, r, and k, written X \sim \text{Hypergeometric}(b,r,k) if its range is R_X = \{\max(0,k-r), ..., \min(k,b) \} and it has the probability mass function P_X(x) =  \begin{cases}<br>\displaystyle \frac{\binom{b}{x}\binom{r}{k-x}}{\binom{b+r}{k}}&  x \in R_X \\[15pt]<br>0 & \text{otherwise}<br>\end{cases}

Example: Say we have a bag 4 blue marbles and 2 red marbles. We choose 3 randomly without replacement. Let X be the number of blue marbles we choose. Show that X \sim \text{Hypergeometric}(4,2,3).



Solution: In this case, b=4, r=2, k=3, and our range is R_X = R_X = \{\max(0,3-2), ..., \min(3,4) \} = \{1,2,3\}

Now let’s look at probabilities. X=1 would mean that we have 1 "success" (blue marble) and 2 "failures" (red marbles). The probability that we’ll have this combination out of all possible combinations of 3 marbles will look like \frac{\text{(\# of ways to get 1 success)(\# of ways to get 2 failures)}}{\text{total number of ways to select 3 marbles}} or P_X(1) = \frac{\binom{3}{1}\binom{2}{2}}{\binom{6}{3}} = \frac{\binom{k}{1}\binom{r}{k-1}}{\binom{b+r}{k}} We can use the same argument for X=2,3, giving us the probability mass function

P_X(x) =  \begin{cases}<br>\displaystyle \frac{\binom{b}{x}\binom{r}{k-x}}{\binom{b+r}{k}}&  x = 1,2,3 \\[15pt]<br>0 & \text{otherwise}<br>\end{cases}

The Poisson Distribution

Say we run a website, www.probabilityiscool.org, which receives an average of 12.5 visitors an hour. We let our random variable X be the number of visitors we receive during some random hour. First we note that even though our mean is 12.5, the actual number of visitors has to be an integer. We should expect the probability to peak at the 12-13 range. The lowest possible number we can receive is zero and, speaking theoretically, the upper limit does not exist. Although the probability of such an event is infintessimally small, we could, in theory, find ourselves in a situation where every person on earth has 4 browsers open and is using all of them to access our website. So our probability as we approach infinitely large numbers will approach zero, but never actually touch it. This situation is an example of a Poisson distribution with paramter 2.5.



Definition: A random variable X is said to have a Poisson distribution with parameter \lambda, written X \sim \text{Poisson}(\lambda) if its range is R_X = \mathbb{N} \cup \{0\} = \{0,1,2,...\} and its probability mass function is given by

P_X(x) =  \begin{cases}<br>\dfrac{e^{-\lambda}\lambda^k}{k!}&  x \in R_X \\[15pt]<br>0 & \text{otherwise}<br>\end{cases}

Notice that



(a) P_X(k) \geq 0 \quad \forall k \in R_X(k)



(b) \displaystyle \sum_{k=0}^{\infty} P_X(k) = \sum_{k=0}^{\infty} \frac{e^{-\lambda \lambda^k}}{k!} = e^{-\lambda} \sum_{k=0}^{\infty} \frac{\lambda^k}{k!} = e^{-\lambda} e^{\lambda} = 1



Here are the graphs of some Poisson distributions for \lambda = 1,2,3 (blue, red, and brown respectively).

 

Example 3.8: The number of emails you get in a weekday can be modeled by a Poisson distribution with an average of 0.2 emails per minute.

(a) What is the probability that you get 0 emails over a span of five minutes.

(b) What is the probability that you get more than 3 emails over a span of 10 minutes?



Solution:

(a) The parameter \lambda will represent the expected number of emails during the given time period. In this case \lambda = 5(0.2) = 1 So P_X(0) = \frac{e^{-\lambda} \lambda^k}{k!} = \frac{e^{-1} (-1)^0}{0!} = e^{-1} \approx 0.3679 (b) Here we have \lambda = 10(0.2) = 2 and so our probability of X being greater than 3 will be P(X>3) = 1- P(X \leq 3) = 1 - \sum_{k=0}^3 \frac{e^{-2} \, 2^k}{k!} = 1 - \left( e^{-2} + \frac{e^{-2} \, 2^1}{1!} + \frac{e^{-2} \, 2^2}{2!} + \frac{e^{-2} \, 2^3}{3!} \right) \approx 0.1429

The Poisson Distribution as an Approximation for the Binomial

Theorem 3.1: Let X \sim \text{Binomial}\left(n, p=\frac{\lambda}{n}\right) where \lambda>0 is fixed. Then, for and k \in \mathbb{N} \cup \{0\}, we have

\lim_{n \rightarrow \infty} P_X(k) = \frac{e^{-\lambda} \, \lambda^k}{k!}

Proof: For our binomial distribution, the PMF looks like

P_X(k) =  \begin{cases}<br>\displaystyle \binom{n}{k} \left(\frac{\lambda}{n} \right)^k \left(1-\frac{\lambda}{n}\right)^{n-k}&  k = 1, 2, ..., n \\[15pt]<br>0 & \text{otherwise}<br>\end{cases}

Taking the limit as n \rightarrow \infty, we get

\lim_{n \rightarrow \infty} \displaystyle \left[ \binom{n}{k} \left(\frac{\lambda}{n} \right)^k \left(1-\frac{\lambda}{n}\right)^{n-k} \right] = \lim_{n\rightarrow \infty} \left[ \left(\frac{n!}{k!(n-k)!} \right)\left( \frac{\lambda^k}{n^k}\right) \left(1-\frac{\lambda}{n}\right)^{n-k}\right] = \frac{\lambda^k}{k!} \lim_{n \rightarrow \infty} \left[\left(\frac{n(n-1)...(n-k+1)}{n^k} \right) \left(1-\frac{\lambda}{n} \right)^n \left(1-\frac{\lambda}{k} \right)^{-k} \right] = \frac{\lambda^k}{k!} e^{-\lambda} = \frac{e^{-\lambda} \lambda^k}{k!} \blacksquare

3.1.6 Solved Problems

Problem 1: Let X be a discrete random variable with

P_X(k) =  \begin{cases}<br>0.1 &  x=0.2 \\[5pt]<br>0.2 &  x=0.4, 0.5, 1 \\[5pt]<br>0.1 &  x=0.8 \\[5pt]<br>0 & \text{otherwise}<br>\end{cases}

Find:

(a) R_X



(b) P(x \leq 0.5)



(c) P(0.25 < X < 0.75)



(d) P(X=0.2 | X < 0.6)



Solution:

(a) R_X = \{0.2,0.4,0.5,0.8,1\}



(b) P(X \leq 0.5) = 0.1+0.2+0.2 = 0.5



(c) P(0.25 < X <0.75) = 0.2+0.2 = 0.4



(d) P(X=0.2 | X <0.6) = \dfrac{P(x < 0.6 | X=2) P(X=0.2)}{P(X < 0.6)} = \dfrac{(1)  (0.1)}{0.1 + 0.2 + 0.2} = 0.2



Problem 2: You roll two 6-sided dice and observe two numbers, X and Y.

(a) Find R_X, R_Y, and the PMFs of X and Y.



(b) Find P(X=2,Y=6)



(c) Find P(X>3|Y=2)



(d) Let Z = X+Y. Find the range and PMF of Z.



(e) Find P(X=4 | Z=8)



Solution:

(a) R_X = R_Y = \{1,2,3,4,5,6\} P_X(k) = P_Y(k) =  \begin{cases}<br>\frac{1}{6} &  k = 1,2,3,4,5,6 \\[5pt]<br>0 & \text{otherwise}<br>\end{cases}

(b) X and Y are clearly independent so P(X=2, Y=6) = P(X=2)P(Y=6) = \frac{1}{6} \cdot \frac{1}{6} = \frac{1}{36}

(c) P(X>3|Y=2) = P(X>3) = \dfrac{1}{2}



(d) R_Z = \{1,2,3,4,5,6,7,8,9,10,11,12\}



To find the PMF, let’s look at all of our possible options:

So we have P_Z(2) = \frac{1}{36}, \quad P_Z(3) = \frac{2}{36}, \quad P_Z(4) = \frac{3}{36}, P_Z(5) = \frac{4}{36}, \quad P_Z(6) = \frac{5}{36}, \quad P_Z(7) = \frac{6}{36}, P_Z(8) = \frac{5}{36}, \quad P_Z(9) = \frac{4}{36}, \quad P_Z(10) = \frac{3}{36}, P_Z(11) = \frac{2}{36}, \quad P_Z(12) = \frac{1}{36}

Just to make sure: \sum_{k=2}^{12} P_Z(k) = \frac{1+2+3+4+5+6+5+4+3+2+1}{36} =1

Putting it all together we get

P_Z(k) =  \begin{cases}<br>\dfrac{1}{36} &  k=2,12 \\[10pt]<br>\dfrac{1}{18} &  k=3,11 \\[10pt]<br>\dfrac{1}{12} &  k=4,10 \\[10pt]<br>\dfrac{1}{9} &  k=5,9 \\[10pt]<br>\dfrac{5}{36} &  k=6,8 \\[10pt]<br>\dfrac{1}{6} &  k=7 \\[10pt]<br>0 & \text{otherwise}<br>\end{cases}

(e) We can use our PMF above to solve this one. We can either refer to our grid above or use Bayes’ Rule: P(X=4|Z=8) = \frac{P(Z=8|X=4) P(X=4)}{P(Z=8)} =\cfrac{\left(\cfrac{1}{6}\right)\left( \cfrac{1}{6}\right)}{\cfrac{5}{36}} = \frac{1}{5}

Problem 3: Say we roll a fair 6-sided die repeatedly until a number larger than 4 is observed. If N is the total number of rolls, find P(N=k) for k \in \mathbb{N}.



Solution: N is an example of a geometric random variable. The parameter will be the probability that a roll gives a number greater than 4, or p= \frac{1}{3}.



So N \sim \text{Geometric}(\left(\frac{1}{3} \right), which means

P_N(k) =  \begin{cases}<br>\dfrac{1}{3} \left( \dfrac{2}{3} \right)^{k-1} &  k \in \mathbb{N} \\[15pt]<br>0 &  k \notin \mathbb{N}<br>\end{cases}

Problem 4: You take an exam which contains 20 multiple choice questions with 4 options each. You know the answer to 10 questions, but have no idea about the other 10 so you choose your options randomly. Let X represent your total number of correct answers.



(a) Find the PMF of X



(b) What is P(X>15)?



Solution:

(a) Let X = Y+10 where Y is the number of correct answers to the 10 questions you’ve randomly guessed. Y is an example of a binomial random variable with parameters n=10 and p=\frac{1}{4}.



So P_Y(y) =  \begin{cases}<br>\displaystyle \binom{10}{y} \left( \frac{1}{4}\right)^{y}  \left( \frac{3}{4}\right)^{10-y} &  y =0,1,2,...,10 \\[15pt]<br>0 &  \text{otherwise}<br>\end{cases} Now to find the PMF of X, we just need to shift into our new range R_X=\{10,11,12,...,20 \} P_X(k) =  \begin{cases}<br>\displaystyle \binom{10}{k-10} \left( \frac{1}{4}\right)^{k-10}  \left( \frac{3}{4}\right)^{20-k} &  k =10,11,12,...,20 \\[15pt]<br>0 &  \text{otherwise}<br>\end{cases} (b) P(X>15) = \sum_{k=16}^{20} \binom{10}{k-10} \left( \frac{1}{4}\right)^{k-10}  \left( \frac{3}{4}\right)^{20-k} \approx 0.0197

Problem 5: Let X \sim \text{Pascal}(m,p) and Y \sim \text{Pascal}(l,p) be two independent random variables. Define a new random variable Z = X + Y. Find the PMF of Z.



Solution: This one is easy if we understand the definition of the Pascal distribution. Remember X \sim \text{Pascal}(m,p) represents the number of trials needed to achieve m successes given the probability of each success is p.



Hence, Z = X + Y represents the number of trials needed to achieve m+l successes given given the probability of each success is p, meaning Z \sim \text{Pascal}(m+l,p).



So we have P_Z(k) =  \begin{cases}<br>\displaystyle \binom{k-1}{m+l-1} p^{m+l} (1-p)^{k-(m+l)}&  k = m+l, m+l+1, m+l+2, ... \\[15pt]<br>0 & \text{otherwise}<br>\end{cases}

Problem 6: The number of customers arriving at a grocery store can be represented as a Poisson random variable. On average, 10 customers will arrive per hour. Let X be the number of customers arriving from 10:00 to 11:30. What is P(10 < X \leq 15)?



Solution: In this case, we have \lambda = 1.5 \cdot 10 = 15, so X \sim \text{Poisson}(15).



Thus P(10 < X \leq 15) = \sum_{k=11}^{15} \frac{e^{-15} \, 15^k}{k!} \approx 0.4496

Problem 7: Let X \sim \text{Poisson}(\alpha) and Y \sim \text{Poisson}(\beta) be two independent random variables. Define a new random variable Z = X+Y. Find the PMF of Z.



Solution: First note that the range for a Poisson random variable will always be the same R_x = R_Y = R_Z = \{0,1,2,...\} Now let’s find our PMF P_Z(k) = P(X+Y=k) = \sum_{i=0}^k P(X+Y=k | X=i) P(X=i) =\sum_{i=0}^k P(Y = k-i | X=i) P(X=i) =\sum_{i=0}^k P(Y = k-i) P(X=i) =\sum_{i=0}^k \left(\frac{e^{-\beta} \, \beta^{k-i}}{(k-i)!} \right) \left(\frac{e^{-\alpha} \, \alpha^{i}}{i!} \right) = e^{-(\alpha + \beta)} \sum_{i=0}^k \frac{\alpha^i \, \beta^{k-i}}{(k-i)! \, i!} = \frac{e^{-(\alpha + \beta)}}{k!} \sum_{i=0}^k  \binom{k}{i} \alpha^i \, \beta^{k-i} = \frac{e^{-(\alpha + \beta)}(\alpha + \beta)^k}{k!}

Therefore, Z \sim \text{Poisson}(\alpha + \beta) and has the PMF P_Z(k) =  \begin{cases}<br>\displaystyle \frac{e^{-(\alpha + \beta)}(\alpha + \beta)^k}{k!} &  k =1,2,3,... \\[15pt]<br>0 &  \text{otherwise}<br>\end{cases}

Problem 8: Let X be a discrete random variable with PMF

P_X(k) =  \begin{cases}<br>\frac{1}{4} & k =-2 \\[10pt]<br>\frac{1}{8} & k =-1  \\[10pt]<br>\frac{1}{8} & k =0 \\[10pt]<br>\frac{1}{4} & k =1 \\[10pt]<br>\frac{1}{4} & k =2 \\[10pt]<br>0 &  \text{otherwise}<br>\end{cases}

Define a new random variable Y = (X+1)^2. Find the range and PMF of Y.



Solution: We have that R_X = \{-2,-1,0,1,2\}, so R_Y = \{0,1,4,9 \}.



Now let’s find the probabilities for each value in the range. P_Y(0) = P\Big((X+1)^2 = 0 \Big) = P(X=-1) = \frac{1}{8} P_Y(1) = P\Big((X+1)^2 = 1 \Big) = P(X=-2) + P(X=0) = \frac{3}{8} P_Y(4) = P\Big((X+1)^2 = 4 \Big) = P(X=1) = \frac{1}{4} P_Y(9) = P\Big((X+1)^2 = 9 \Big) = P(X=2) = \frac{1}{4} Check: \sum_{y \in R_Y} P_Y(y) = \frac{1}{8}+ \frac{3}{8}+\frac{1}{4}+\frac{1}{4} = 1 So we have P_Y(y) =  \begin{cases}<br>\frac{1}{8} & y= 0 \\[10pt]<br>\frac{3}{8} & y=1 \\[10pt]<br>\frac{1}{4} & y=4\\[10pt]<br>\frac{1}{4} & y=4 \\[10pt]<br>0 &  \text{otherwise}<br>\end{cases}