[1] 0.0256
A random variable is a quantified observation from a process whose outcome is not known with uncertainty. \[\begin{align} X=x \end{align}\]
The capital letter “X” is the definition of the random variable.
The small caps letter “x” is the value it can take.
\[\begin{align} X=\{0,1,3\} \end{align}\]
\[\begin{align} X = \{0,1\} \end{align}\]
There are different interpretations of probability.
Words such as chance and likelihood can be used colloquially but it does not define anything. In fact in statistics, likelihood refers to something else.
Objective interpretation: Relative frequency of the value of a random variable in a large number of trials which obeys laws of probability.
Subjective interpretation: Quantification of uncertainty which obeys laws of probability.
Imagine a coin. You toss the coin and you will assign a probability to heads (X=1) and tails (X=0)
P(X=1)=0.5 & P(X=0) = 0.5
But why? It can not be just because there are 2 outcomes. Would you assign 0.5 as probability to sun rising tomorrow?
You either did a mental calculation of having a really large number of coin tosses and calculated the relative frequency (objective) or you used your mental image of the world and what you have experienced in your life about coin tosses (subjective).
Let 1 represent heads, 0 for tails.
Marginal: \(P(X=x)\) Probability associated with the value of a single random variable. What is the probability of the first coin toss being heads. P(X=1)
Joint: \(P(X_{1}=x_{1},X_{2}=x_{2})\) Probability of having two or more random variables happening together. What is the variable I will toss a coin two times and the first time it lands heads and the second time it lands tails. \(P(X_{1}=1,X_{2}=0)\)
Conditional: \(P(X_{1}=x_{1}|X_{2}=x_{2})\) Probability of one random variable happening given information on the right side of the sign \(\vert\) is referred to as given. What is the probability of the first coin toss to be heads if I know the second coin toss is tails. \(P(X_{1}=1|X_{2}=0)\)
The definition of conditional probability is \[P(X=x \vert Y=y) = \frac{P(X=x,Y=y)}{P(Y=y)} \]
Random Variable definitions
X=1/0 Middle Aged/Not Middle Aged
Y=1/0 Buys product/Does not Buy Product
Contingency table (table of counts)
| X=0 | X=1 | |
|---|---|---|
| Y=0 | 22 | 33 |
| Y=1 | 44 | 55 |
Technically, the numerator and denominator should really be divided by the sum of the of the contingency table (22+44+33+55=154).
This would be done to turn the numerator and the denominator into probabilities before the ratio is evaluated but the sum is common to both so they would cancel out. \(\frac{\frac{22}{154}}{\frac{22+33}{154}}=0.40\)
Conditional probability of a random variables given information on the other one by ignoring the common division of numerator and denominator by 154. \[P(X=0 \vert Y=0)=\frac{(22)}{(22+33)}=0.40 \]
All of this is required to understand the laws of probability. These laws are used to define properties of probability.
Convexity: \(0<P(X=x)<1\)
Addition: \(P(X_{1}=x_{1} \cup X_{2}=x_{2} )=\)
Multiplication: \(P(X_{1}=x_{1} , X_{2}=x_{2} )=\) \[ \begin{split} & P(X_{1}=x_{1}\vert X_{2}=x_2) \times P(X_{2}=x_2) \\ & P(X_{2}=x_{2}\vert X_{1}=x_1) \times P(X_{1}=x_1) \end{split} \]
\(P(\Omega)=1\) where \(\Omega\) is sample space of X.
\[\begin{align} P(Y=y|X=x) = P(Y=y) \\ P(X=x|Y=y) = P(X=x) \end{align}\]
\[P(S_{t}=s_{t} \vert S_{t-1}=s_{t-1}) \ne P(S_{t}=s_{t}) \]
i=If we assume that yesterday provides some information regarding today’s stock price we can think about a possible probabilistic structure.
However formulating the whole distribution might be a challenge for this part of the course so we will focus on the expected value of a probability distribution.
Compare this with the structure below for dependence to become apparent where \(s_{t-1}\) is the observed previous day’s closing price. \[ \begin{aligned} \mu_{S_{t}\vert{s_{t-1}}} = s_{t-1}+\epsilon_{t} \end{aligned} \]
The \(\epsilon_{t}\) can be a Notmal distribution similar to the one above,with mean 0 and \(\sigma_{S}\) what makes the information about \(S_{t}\) dependent on information on \(S_{t-1}\) is that we are adding it as a constant to calculate \(\mu_{S_{t}\vert{s_{t-1}}}\).
In order for the demonstration we will need to make up some numbers for a sample of individuals. These numbers need to be joint descriptions.
X=1/0 Middle Aged/Not Middle Aged
Y=1/0 Buys product/Does not Buy Product
| X=0 | X=1 | |
|---|---|---|
| Y=0 | 22 | 33 |
| Y=1 | 44 | 55 |
X=2/1/0 Senior Citizen/Middle Aged/Not Middle Aged
Y=1/0 Buys product/Does not Buy Product
| X=0 | X=1 | X=2 | |
|---|---|---|---|
| Y=0 | 22 | 33 | 66 |
| Y=1 | 44 | 55 | 77 |
Probability that a randomly selected individual in the sample is a senior citizen is \(P(X=2)=P(X=2,Y=0)+P(X=2,Y=1)\).
The sum of joint probabilities of a randomly selected individual being a senior citizen and purchasing a product and the probability of a randomly selected individual to bea senior citizen and not purchasing a product.
There are a total of 22+33+44+55+66+77=297 individuals in our sample.
\[ \begin{aligned} & P(Y=2)=&P(X=2,Y=2)&+&P(X=2,Y=1)\\ & P(Y=2)=&\frac{66}{297}&+&\frac{77}{297}= 0.78 \end{aligned} \]
Think about the consecutive coin tosses again. \[ \begin{split} & P(X_{1}=1,X_{2}=1|p=0.5)= \\ & P(X_{1}=1|p=0.5) \times P(X_{2}=1|p=0.5) \end{split} \] where p is the common probability of the outcome of a coin toss being heads.
For intro statistics sometimes we have to cut corners to make presentation simpler. Even though all probabilities are really conditional we will use the following argument.
If two random variables say \(X_{1}\) and \(X_{2}\) are independent of each other. Their joint probability can be found by multiplying each other.
\[ P(X_{1}=x_{1},X_{2}=x_{2})=P(X_{1}=x_{1}) \times P(X_{2}=x_{2}) \]
Let us work with events (special random variables)
What is the probability of observing 3 heads in a row in independent coin tosses. Use H to represent heads and T for tails. \(P(H)=0.5,P(T)=0.5\)
Below, the subscript identify the order of the throw. We would like to calculate the probability of \(H_{1},H_{2},H_{3}\) \[P(H_{1},H_{2},H_{3})=P(H_{1}) \times P(H_{2}) \times P(H_{3})\]
\[P(H_{1},H_{2},H_{3})=P(H_{1}) \times P(H_{2}) \times P(H_{3})\] \[0.5 \times 0.5 \times 0.5=0.125\]
Maddie asks, does that lead to the idea that the following sequences have all the same probability?
\(P(H_{1},H_{2},H_{3},H_{4},H_{5},H_{6},H_{7},H_{8},H_{9},H_{10},H_{11})\)
\(P(T_{1},T_{2},T_{3},T_{4},T_{5},T_{6},T_{7},T_{8},T_{9},T_{10},T_{11})\)
\(P(H_{1},T_{2},H_{3},H_{4},T_{5},T_{6},H_{7},T_{8},H_{9},T_{10},H_{11})\)
Yes all those sequences have exactly the same probability of occurring.
However if the question you were interested in was what is the probability that you are going to have 11 heads in 11 coin tosses or 6 heads in 11 coin tosses you will not get the same probability.
Why is that?
There is a difference between specifying a particular sequence of heads and tails in N throws of coins and asking for the probability of a number of heads in N throw of coins without being worried about the order in which they were observed.
If I ask for y (number of) heads in N coin tosses you will have to think about in how many ways can this happen? Once you get the answer to that question you can multiply it with the probability of a particular sequence of y heads in N coin tosses to get the answer. The combinatorial function \(n \choose y\) will calculate the number of ways.
Y=1 if person has a virus 0 otherwise, X=1 if a test returns a positive result and 0 otherwise.
If we have \(P(X=x \vert Y=y)\) can we obtain \(P(Y=y \vert X=x)\)
Perhaps. First let us understand the difference between these conditional probability constructs.
\[ \begin{aligned} P(X=x \vert Y=y)=\frac{P(X=x,Y=y)}{P(Y=y)}; \\ P(Y=y \vert X=x)= \frac{P(Y=y,X=x)}{P(X=x)}; \end{aligned}\]
Both of the conditional probabilities have the same numerator (the order of the variables in the joint probability does not matter). The denominator is where the difference is.
If you remember from the properties of the probability you can sum joint probabilities to get the marginal probability.
\[\begin{aligned} P(X=1)=P(X=1,Y=1)+P(X=1,Y=0) \\ P(Y=1)=P(X=1,Y=1)+P(X=0,Y=1) \end{aligned} \] - In general \(P(X=1)=\sum_{i=1}^{i=K}=P(X=1,Y=y_{i})\)
\[ \begin{aligned} & P(Y=1 \vert X=1)=\frac{P(Y=1,X=1)}{P(X=1)}= \\ & \frac{P(Y=1,X=1)}{P(X=1,Y=1)+P(X=1,Y=0)} \\ & \frac{P(X=1 \vert Y=1) \times P(Y=1)}{P(X=1 \vert Y=1) \times P(Y=1)+P(X=1 \vert Y=0) \times P(Y=0)} \\ \end{aligned} \]
\[\text{In General,}P(Y=i \vert X=m) \frac{P(X=m \vert Y=i) \times P(Y=i)}{ \sum_{j=1}^{j=K}P(X=m \vert Y=j) \times P(Y=j)} \]
Are you surprised at probability of being actually infected given that the test returns a positive result is 16.5\(\%\).
\(P(Y=1 \vert X=1 )=0.98\) what is the driver behind the low percentage of 16.5\(\%\)?
Try to problem by changing the virus prevalence rate to 0.001 and then to 0.1. How do the results change? Why?
We need to come up with ways that will assign probabilities to uncertain outcomes without us breaking our minds for these calculations hence the need for named probability distributions.
Probability distributions are functions that assign uncertainty to random variables’ values which obey the laws of probability.
If you take any arbitrary set of two values from a random variable, and a middle point can always be found within the sample space you have a continuous variable. Otherwise you do have a discrete one.
\[ \begin{split} & P(X=x) = p^{x}\times(1-p)^{1-x}\\ & P(X=1) = p^{1}\times(1-p)^{1-1}=p^{1}\times (1-p)^{0}=p\\ & P(X=0) = p^{0}\times(1-p)^{1-0}=p^{1}\times (1-p)^{1}=1-p \end{split} \] - The function \(P(X=x)\) assigns probability to the two values that X can take. If it takes the value 1, \(P(X=1)=p\) and \(P(X=0)=1-p\).
The letter p is the parameter of the distribution.
If X has a Bernoulli distribution p is the probability of success and (1-p) is the probability of failure.
-Furthermore if X has a Bernoulli distribution. \[ \begin{aligned} E(X)=&p\\ Var(X)=&p\times(1-p)\\ \end{aligned} \]
\[ P(Y=y) = \binom{n}{y} \times p^{y} \times (1-p)^{n-y} \] - where n is the total number of Bernoulli trials, p is probability of success and y is the total number of successes.
\[ E(Y)=n\times p , Var(Y)=n \times p \times (1-p) \]
\[ \binom{n}{y}=\frac{n!}{y!\times(n-y)!} \]
Assume you are inspecting 4 widgets. The probability of finding a defective widget is 0.2. If this probability did not change what is the probability that you will find 3 defective items.
Y is the sum of 4 bernoulli trials, each leading to either defective \((X_{i}=1)\) or nondefective \((X_{i}=0)\) where \(i\) is 1 \(\ldots\) 4.
\[ P(Y=3 \vert n=4, p =0.2)= \binom{4}{3} \times 0.2^{3} \times (1-0.2)^{4-3} \]
\[ \binom{4}{3} = \frac{4!}{3!\times (4-3)!}=\frac{4 \times 3 \times 2 \times 1}{3 \times 2 \times 1 \times 1}=4 \] - This simply implies that there are 4 different ways you can choose 3 successes (and one failure by default) out of 4 tries.
| Row | Widget 1 | Widget 2 | Widget 3 | Widget 4 |
|---|---|---|---|---|
| 1 | 1 | 1 | 1 | 0 |
| 2 | 1 | 1 | 0 | 1 |
| 3 | 1 | 0 | 1 | 1 |
| 4 | 0 | 1 | 1 | 1 |
| Row | Widget 1 | Widget 2 | Widget 3 | Widget 4 |
|---|---|---|---|---|
| 2 | 1 | 1 | 0 | 1 |
This is joint probability and can be represented with notation as \(p(X_{1}=1,X_{2}=1,X_{3}=0,X_{4}=1)\)
How do we calculate this joint probability?
\[ \begin{split} & p(X_{1}=1,X_{2}=1,X_{3}=0,X_{4}=1)\\ & p(X_{i}=1)=0.2 ~\&~ (1-p(X_{i}=1))=0.8\\ & p \times p \times (1-p) \times p = \\ & 0.2 \times 0.2 \times 0.8 \times 0.2 \\ & 0.2^{3} \times 0.8 = 0.0064 \end{split} \]
| Row | \(X_{1}\) | \(X_{2}\) | \(X_{3}\) | \(X_{4}\) | P(Row) |
|---|---|---|---|---|---|
| 1 | 1 | 1 | 1 | 0 | 0.0064 |
| 2 | 1 | 1 | 0 | 1 | 0.0064 |
| 3 | 1 | 0 | 1 | 1 | 0.0064 |
| 4 | 0 | 1 | 1 | 1 | 0.0064 |
| \(\sum_{1}^{4}\) | \(4 \times 0.0064\) |
In how many different ways can you have 2 defective items among 5? Among the 5 inspected items, if there are 2 defective items there will be exactly 3 non-defective items.
D is defective N is non-defective. Subscript is the order in which the item is observed. You would need to count each possible scenario:
\[ \begin{aligned} D_{1},D_{2},N_{3},N_{4},N_{5}\\ D_{1},N_{2},D_{3},N_{4},N_{5}\\ \cdots \cdots \cdots \cdots \cdots \\ N_{1},N_{2},N_{3},D_{4},D_{5}\\ \end{aligned} \]
\[ N_{1},N_{2},N_{3},D_{4},D_{5} \]
So we can simply multiply 10 which is number of ways you can have 2 successes out of 5 trials with 0.00729 which is the probability of having one such way to get 0.0729 which is the probability of having 2 successes out of 5 trials.
This in effect is doing the following operation:
\[ P(N_{1},N_{2},N_{3},D_{4},D_{5})+P(N_{1},N_{2},D_{3},N_{4},D_{5})+\\ \cdots+P(D_{1},D_{2},N_{3},N_{4},N_{5})=0.0729 \]
\[ \binom{n}{y}p^{y}(1-p)^{(n-y)}=\\ \binom{5}{2}\times0.1^{2}\times(1-0.1)^{(5-2)}=\\ 10\times0.00729=0.0729 \]
\[ =binom.dist(y,n,p,CUMULATIVE) \]
\[ \begin{aligned} & P(Y \le 2 \vert n=5,p=0.1)=\\ & P(Y=2 \vert n=5,p=0.1)+ \\ & P(Y=1 \vert n=5,p=0.1)+\\ & P(Y=0 \vert n=5,p=0.1) = 0.9914 \end{aligned} \]
How is \(P(Y \le 2 \vert n=5,p=0.1)\) different from \(P(Y = 2 \vert n=5,p=0.1)\)
In the cumulative probability, \(P(Y = 2 \vert n=5,p=0.1)\) is included but so are other probabilities.
If you are to stop a manufacturing process when you have inspected at most 2 defective items among 5 items you have inspected that means you will stop the process if you see not just 2 defective items but even if you observe 1 or 0 defective items!
Of course stopping the process with no defective items makes no sense but that is how I formulated the statement.
You are going to inspect 5 items from lot A and 10 items from lot B. If the process is under control the probability of having a defective item in lot A is 0.1 and it is 0.2 in lot B. You have a quality focus in the organization so you decide to stop the process if you observe at least 1 defective item in lot A and at least 1 defective item in lot B. Find the probability that the process is going to be stopped.
Define your random variables first!
A \(\equiv\) Number of defective items in lot A \(A=\{0,1,\ldots5\}\).
B \(\equiv\) Number of defective items in lot B \(B=\{0,1,\ldots10\}\).
We can define these random variables as Binomial distributed because there is a fixed number of trial. The probability of success is known and does not have a reason to change from trial to trial. We will assume that the random variables A and B are independent from each other.
\[P(A>0,B>0 \vert n_{A}=5,p_{A}=0.1,n_{B}=10,p_{B}=0.2)=\] \[P(A>0 \vert n_{A}=5 ,p_{A}=0.1 ) \times P(B>0 \vert n_{B}=10 ,p_{B}=0.2) \] The decomposition can only happen because we assume A and B are independent.
Sum of all probabilities associated with the values of a random variable has to add up to 1. \(P(\Omega)=1\).
In our example this would simply mean \(P(A=0)+P(A=1)\cdots+P(A=5)=1\)
This means to evaluate \(P(A>0)\) I can either as mentioned already, \(P(A=1)+P(A=2)+\cdots +P(A=5)\) or better yet
\(P(A=0)+P(A=1)\cdots+P(A=5)=1\) therefore \(P(A=1)\cdots+P(A=5)=1-P(A=0)\)
\[P(A>0)=1-P(A=0)=\\ 1-binom.dist(0,5,0.1,FALSE)=\\ 1-0.59=0.41 \] - The last parameter FALSE/TRUE does not matter in this particular case since A can not be less than 0.
\[P(B>0)=1-P(B=0)=\\1-binom.dist(0,10,0.2,FALSE)=\\1-0.11=0.89\]
\[P(A>0,B>0)=(1-P(A=0))\times(1-P(B=0))\\ 0.41 \times 0.89=0.37 \] - There is a \(37\%\) probability that the process is going to be stopped.
So far the Bernoulli and Binomial random variables had an upper limit. Bernoulli can at most be 1 (single trial) which of course leads to Binomial distribution to be at most N. However there are cases where there is no theoretical upper limit to the random variable.
Number of patients arriving to the E.R.
Number of calls to a call center.
Number of eggs a chicken hatches.
\[ P(X = x) = \frac{e^{-\lambda} \lambda^x}{x!}, \quad x = 0,1,2,\ldots \] - Assumptions of the Poisson Distribution
Each event occurs (i.e. customer arrival) occurs independently.
\(\lambda\) does not change and is equal to mean and variance of r.v.. Which means the mean and variance is constant.
No simultaneous events. Two customers do not arrive exactly at the same time etc…
The mean and variance is equal to \(\lambda\)
Usually a random variable which is assumed to have a Poisson Distribution will have an associated rate. For instance 3 customers per hour arrives on average to a store. Per 8 hours the expectation is \(3 \times 8=24\) arriving on average to the store.
Fun fact if a random variable has Binomial distribution as number of Bernoulli trials becomes “large” the random variable approaches Poisson distribution.
I will hire a new employee if the number of customers that purchase an item is more than 5 within an hour. We can assume that the customers purchase an item with a Poisson distribution. The standard deviation of the customers that purchase an item within an hour is 1.5. What is the probability that I will hire a new employee.
First define your random variable.
X= number of customers that purchase an item within the hour.
Define the question with probability notation
We can not evaluate every probability for X greater than 5. Nobody has time for that. So instead?
\[P(X > 5 \vert \lambda=2.25) = 1- P(X \le 5 \vert \lambda=2.25) \] \[1-poisson.dist(5,2.25,TRUE) = 0.03\]
The quality control procedure inspects 8 items in lot 1, 10 items from lot 2. If there is at least 2 items that are defective in lot 1 or at least 3 items that are defective in lot 2. You stop the manufacturing process. What is the probability that you will stop the process. Probability of a random item to be defective in lot 1 is 0.3. Probability of a random item to be defective in lot 2 is 0.25.
Start by defining the random variables and identify prob.
Define the question with notation and the random variables you have defined.
Let Y represent the number of items in lot A that are defective. \(Y=\{0,\cdots,8\},p_{y}=0.3,n_{y}=8\)
Let Z represent the number of items in lot B that are defective. \(Z=\{0,\cdots,10\},p_{z}=0.25,n_{z}=10\)
The question redefined with notation and r.v.s \[ \begin{split} & P(Y \ge 2 \cup Z \ge 3) \\ & P(Y \ge 2)+P(Z \ge 3) - P(Y \ge 2, Z \ge 3)\\ \end{split} \]
Since the probability of all values of Y has to add up to 1. \[ \begin{split} & P(Y \le 1) + P(Y \ge 2)) =1 \\ & P(Y \ge 2)=1-P(Y \le 1) \end{split} \]
Find each component in the equation separately.
\[ \begin{split} & P(Z \le 2) + P(Z \ge 3)) =1 \\ & P(Z \ge 3)=1-P(Z \le 2) \end{split} \]
\[ P(Y \ge 2) \times P(Z \ge 3)= 0.7447 \times 0.4744=0.3533\] - Putting it all together
\[ \begin{split} & P(Y \ge 2) &+& & P(Z \ge 3) &-& & P(Y \ge 2 , Z \ge 3)&= \\ & 0.7447 &+& & 0.4744 &-& & 0.3533 &= 0.8658 \end{split} \]
What values do you think were involved to assume these cutoff values that lead to process stoppage?
Think short term. Think long term.
Think about what you are trying to optimize and why.
Off course we have a multitude more of distributions than the ones below.
Discrete random variables (Values it can take):
Bernoulli (0,1),
Binomial (0,1,\(\ldots\),n)
Poisson (0,1,\(\ldots,+\infty\))
Continuous random variable:
Normal Distribution \((-\infty,+\infty)\)
The mean (\(\mu\)) divides the distribution into 2 equivalent halves.
50\(\%\) of the data is above (to the right) of the mean and 50\(\%\) is below the mean.
The standard deviation (\(\sigma\)) is an index of how much the values of the random variable deviates around the mean (same interpretation as variance)
\[\sigma_{X}=\sqrt{\frac{\Sigma_{i=1}^{i=N}(x-\mu_{x})^{2}}{N}} \]
f(X=x) gives you the height of the distribution
The height of the distribution is the likelihood/density (NOT PROBABILITY).
The probability of a single value, \(X=x\) is a infinitely small value.
We can not calculate meaningful probabilities for single values.
But we can calculate probabilities for X’s intervals.
\[ P(X>x \vert \mu, \sigma)= \int_{50}^{\infty} \frac{1}{\sigma*\sqrt(2*\pi)} \times exp^{- \frac{(x - \mu)^2}{2*\sigma^2}}\]
\[ P(X>x \vert \mu, \sigma)= \int_{50}^{90} \frac{1}{\sigma*\sqrt(2*\pi)} \times exp^{- \frac{(x - \mu)^2}{2*\sigma^2}}\]
Any value of a random variable with a normal distribution can have its value mapped to the x-axis of the standard normal distribution.
The x-axis values of the standard normal distribution are called z-scores.
\[z=\frac{(x-\mu)}{\sigma} \]
Because it signifies the number of standard deviations a particular value of X is away from \(\mu_{X}\). This in turn implies that the probability that is associated with a random variable X can be associated with the random variable Z.
\(P(X<x)=P(Z=z)\) for correct x to z mapping.
\[z=\frac{(x-\mu_{x})}{\sigma_{x}} \]
\[z_{90}=\frac{90-70}{10}=2 \] \[z_{50}=\frac{50-70}{10}=-2 \]
We can use these numbers in order to calculate the probability we are interested
Given that \(X \ sim N(70,10^{2})\) \[P(z_{50}<Z<z_{90})=P(50<X<90) \] \[P(-2<Z<2)=P(50<X<90) \]
\(P(Z<2)=0.98\)
\(P(Z<-2)=0.02\)
X and Z have a direct relationship that is easy to interpret.
Prob is associated with a Z score that in turn can allow you to obtain an X value that you are interested in.
A stats professor just did an exam and saw that her students had an average of 60 and standard deviation of 12 (assume these population parameters). She wants to give a B to the top \(20\%\) of her class. What is the cutoff value for a B?
In Excel the function \(norm.s.inv\) returns you a z-score. You feed into it the probability to the left of the point you find. 0.20 is the area to the right of the point you would like to find. Therefore we need to subtract 0.20 from 1 in order to find the area to the left of the point we are interested in.
In R the same exact value is obtained with the function qnorm requiring the exact same parameter for exactly the same reason.
\[\text{Excel }=norm.s.inv(0.8)=\text{R }qnorm(0.8)=0.84 \]
What we find using the norm.s.inv or qnorm function is how many standard deviations is the X value away from the mean and in which direction. The positive values indicate above the mean and the negative values indicate below the mean.
For this particular question we know that the B is 0.84 \(\sigma\) above the mean. \(60+0.84\times 12=70.08\)
A machine cuts a widget’s length on average 7 cm with standard deviation 0.01.
A machine cuts a widget’s width on average 6 cm with standard deviation 0.01.
Define a random variable L number of cm of length cut.
Define a random variable W number of cm of width cut.
The engineering specs for the widgets is 6.975 and 7.025 on the length and 5.985 and 6.02 on the width.
What is the probability that the widget length is going to fit the specs? P(6.975<L<7.025)=?
What is the probability that the widget width is going to fit the specs? P(5.985<W<6.02)=?
What is the probability that the widget is going to fit the specs? P(6.975<L<7.025,5.985<W<6.02)=?
\[\begin{aligned} & z=\frac{(630-545)}{45}=1.89\\ & P(T>630 \vert \mu=545, \sigma=45)=P(Z>1.89) \end{aligned} \]
If 10 calls have been received what is the probability that there will be more than 2 calls that have to wait more than 630 seconds? Why is this number different from just calculating the answer to the first question and squaring it?
To answer this question we will have to define a new random variable. W can be defined as the number of calls that wait more than 630 seconds in 10 calls. If we assume that everything is stable, we can model W with a binomial distribution. Recall that we have already identified that the probability that a single individual has to wait more than 630 seconds is 0.03. This is your probability of success.
\[P(W>2 \vert n=10, p =0.03)=P(W=3)+\ldots+P(W=10)=? \]
\[\begin{aligned} 1= P(W>2 \vert n=10, p =0.03)+P(W \le 2 \vert n=10, p =0.03) \\ P(W>2 \vert n=10, p =0.03) = 1-P(W \le 2 \vert n=10, p =0.03) \end{aligned} \]
\[\begin{aligned} & Excel =BINOM.DIST(2,10,0.03,TRUE)=\\ & R= pbinom(2,10,0.03)=0.997 \end{aligned}\]
Information: P(X<=0)=0.05 , P(X<=1)=0.20, P(X<=2)=0.42, P(X<=3)=0.65 P(X<=4)=0.82, P(X<=5)=0.92, P(X<=6)=0.97.
What random variable do you need to define for a question that will answer probability of having to divert a patient? X = Number of patients arriving per hour. Note that number of beds you open up is not a random variable, it is a decision variable.
Write the following question in rotational form. Probability of having a diversion event not being more than 0.1 (Hint: Yes you can answer this question with a Bernoulli distribution but ignore that). \(P(X \ge x)=0.10\)
How many beds should you open so that you will not have more than 10\(\%\) chance of diverting a patient while maximizing profit? 5 beds.
How many beds should you open up to maximize profit disregarding the 10\(\%\) constraint? Each patient arriving can be thought of as a Bernoulli trial, insured or uninsured. Expected profit/cost from a patient regardless of number of beds can be calculated as:
Expected Profits from a patient: \[ \begin{aligned} & p*(Profit)+ (1-p)*Cost= \\ & 0.78*5000+0.22*(-2000)=3,460\\ \end{aligned} \]