Resources :
Please have a look at re-emphasizing the core concepts of CLT, standard error, hypothesis testing, confidence interval and p-values. Please skim through them (~15 minutes) before attempting the assignment to refresh your memories.
Please find the Open Intro Statistics textbook (OpenStat_textbook.pdf) in our Dropbox folder - skimming over Chapter 5, 6 and 7 may be helpful to see the standard error formulas for some of the questions. I will explicitly redirect you to the textbook for some questions.
I have 4 user defined functions below - you do not have to use them, but may find them useful to graphically draw out what is happening in the question.
# Clear the workspace
rm(list = ls()) # Clear environment
gc() # Clear unused memory
## used (Mb) gc trigger (Mb) limit (Mb) max used (Mb)
## Ncells 531511 28.4 1183970 63.3 NA 669265 35.8
## Vcells 979646 7.5 8388608 64.0 16384 1840452 14.1
cat("\f") # Clear the console
We write a function which takes in two arguments (numbers here), runs some computations (basic inequality) on them and prints an output based on the computation result -
myp=function(p, alpha){
if(p<alpha){print('REJECT Ho')}else{print('FAIL 2 REJECT')}
}
Test our function to make sure it is performing as intended -
myp(.01, .05) # p is less than alpha
## [1] "REJECT Ho"
myp(.1, .05) # p is greater than alpha
## [1] "FAIL 2 REJECT"
Now, lets write a bit more complex function (takes in many arguments) that is designed to shade the standard normal distribution as the default option for a 5% double sided hypothesis test and can be adapted for other purposes too. You can chnage the arguments of mu, sig, pcts, color,…
shadenorm = function(below=NULL, above=NULL, pcts = c(0.025,0.975), mu=0, sig=1, numpts = 500, color = "gray", dens = 40, justabove= FALSE, justbelow = FALSE, lines=FALSE,between=NULL,outside=NULL){
if(is.null(between)){
below = ifelse(is.null(below), qnorm(pcts[1],mu,sig), below)
above = ifelse(is.null(above), qnorm(pcts[2],mu,sig), above)
}
if(is.null(outside)==FALSE){
below = min(outside)
above = max(outside)
}
lowlim = mu - 4*sig # min point plotted on x axis
uplim = mu + 4*sig # max point plotted on x axis
x.grid = seq(lowlim,uplim, length= numpts)
dens.all = dnorm(x.grid,mean=mu, sd = sig)
if(lines==FALSE){
plot(x.grid, dens.all, type="l", xlab="X", ylab="Density") # label y and x axis
}
if(lines==TRUE){
lines(x.grid,dens.all)
}
if(justabove==FALSE){
x.below = x.grid[x.grid<below]
dens.below = dens.all[x.grid<below]
polygon(c(x.below,rev(x.below)),c(rep(0,length(x.below)),rev(dens.below)),col=color,density=dens)
}
if(justbelow==FALSE){
x.above = x.grid[x.grid>above]
dens.above = dens.all[x.grid>above]
polygon(c(x.above,rev(x.above)),c(rep(0,length(x.above)),rev(dens.above)),col=color,density=dens)
}
if(is.null(between)==FALSE){
from = min(between)
to = max(between)
x.between = x.grid[x.grid>from&x.grid<to]
dens.between = dens.all[x.grid>from&x.grid<to]
polygon(c(x.between,rev(x.between)),c(rep(0,length(x.between)),rev(dens.between)),col=color,density=dens)
}
}
# TEST THE FUCTION
shadenorm(mu = 0, sig = 1, pcts = c(0.025,0.975))
# shadenorm(mu = 20, sig = 6, pcts = c(0.025,0.975))
shadet = function(below=NULL, above=NULL, pcts = c(0.025,0.975), df=1, numpts = 500, color = "gray", dens = 40, justabove= FALSE, justbelow = FALSE, lines=FALSE,between=NULL,outside=NULL){
if(is.null(between)){
below = ifelse(is.null(below), qt(pcts[1],df), below)
above = ifelse(is.null(above), qt(pcts[2],df), above)
}
if(is.null(outside)==FALSE){
below = min(outside)
above = max(outside)
}
lowlim = -4
uplim = 4
x.grid = seq(lowlim,uplim, length= numpts)
dens.all = dt(x.grid,df)
if(lines==FALSE){
plot(x.grid, dens.all, type="l", xlab="X", ylab="Density")
}
if(lines==TRUE){
lines(x.grid,dens.all)
}
if(justabove==FALSE){
x.below = x.grid[x.grid<below]
dens.below = dens.all[x.grid<below]
polygon(c(x.below,rev(x.below)),c(rep(0,length(x.below)),rev(dens.below)),col=color,density=dens)
}
if(justbelow==FALSE){
x.above = x.grid[x.grid>above]
dens.above = dens.all[x.grid>above]
polygon(c(x.above,rev(x.above)),c(rep(0,length(x.above)),rev(dens.above)),col=color,density=dens)
}
if(is.null(between)==FALSE){
from = min(between)
to = max(between)
x.between = x.grid[x.grid>from&x.grid<to]
dens.between = dens.all[x.grid>from&x.grid<to]
polygon(c(x.between,rev(x.between)),c(rep(0,length(x.between)),rev(dens.between)),col=color,density=dens)
}
}
# TEST THE FUCTION
shadet(df = 4, pcts = c(0.025,0.975)) # see the area under the tails are further away from the mean 0..
# shadet(df = 120, pcts = c(0.025,0.975)) # t dist converges to normal when we have high degrees o freedom..
shadechi = function(below=NULL, above=NULL, pcts = c(0.025,0.975), df=1, numpts = 500, color = "gray", dens = 40, justabove= FALSE, justbelow = FALSE, lines=FALSE,between=NULL,outside=NULL){
if(is.null(between)){
below = ifelse(is.null(below), qchisq(pcts[1],df), below)
above = ifelse(is.null(above), qchisq(pcts[2],df), above)
}
if(is.null(outside)==FALSE){
below = min(outside)
above = max(outside)
}
lowlim = 0
uplim = qchisq(.99,df)
x.grid = seq(lowlim,uplim, length= numpts)
dens.all = dchisq(x.grid,df)
if(lines==FALSE){
plot(x.grid, dens.all, type="l", xlab="X", ylab="Density")
}
if(lines==TRUE){
lines(x.grid,dens.all)
}
if(justabove==FALSE){
x.below = x.grid[x.grid<below]
dens.below = dens.all[x.grid<below]
polygon(c(x.below,rev(x.below)),c(rep(0,length(x.below)),rev(dens.below)),col=color,density=dens)
}
if(justbelow==FALSE){
x.above = x.grid[x.grid>above]
dens.above = dens.all[x.grid>above]
polygon(c(x.above,rev(x.above)),c(rep(0,length(x.above)),rev(dens.above)),col=color,density=dens)
}
if(is.null(between)==FALSE){
from = min(between)
to = max(between)
x.between = x.grid[x.grid>from&x.grid<to]
dens.between = dens.all[x.grid>from&x.grid<to]
polygon(c(x.between,rev(x.between)),c(rep(0,length(x.between)),rev(dens.between)),col=color,density=dens)
}
}
# TEST THE FUCTION
shadechi(df = 2, pcts=c(.05)) # change pcts and see what happen
shadechi(df = 18, pcts=c(.05)) # change df and see what happens
Using traditional methods, it takes 109 hours to receive a basic driving license. A new license training method using Computer Aided Instruction (CAI) has been proposed. A researcher used the technique with 190 students and observed that they had a mean of 110 hours. Assume the standard deviation is known to be 6. A level of significance of 0.05 will be used to determine if the technique performs differently than the traditional method. Make a decision to reject or fail to reject the null hypothesis. Show all work in R.
Given: \(\mu=, n=, \bar{x}=, \sigma=, \alpha=\).
To Do: Determine if the technique performs differently than the traditional method. Burden of proof falls on alternative hypothesis -
Ho: \(\mu\), Ha: $$
Ho: in words ???
Ha: in word ???
Two sided test (look at alternative hypothesis).
\(\alpha = .05\)
Distribution: Z (known SD)
???
Z = (110-109 ) / (6/sqrt(190)) # compute test statistic
Z
## [1] 2.297341
p_value = 2 * ( 1-pnorm(q = Z,
mean = 0,
sd = 1)
) # p value associated with the two sided hypothesis test - probability of finding as or more extreme outcomes than what the sample suggests
p_value
## [1] 0.0215993
alpha = .05 # given to us in question
# visually, this is happening (you do not have to draw the curves)
shadenorm( mu = 109, sig = 6/sqrt(190), pcts = c(0.025,0.975), color = "red") # shades significance level gates
lines(x=rep(110,10), y=seq(0,1,length.out=10), col='blue') # mark point estimate from sample
# Algorithm / myp function that we define gives us the same result
myp(p = p_value, alpha = alpha )
## [1] "REJECT Ho"
# Lets call comparing p value with alpha this method 1. Other methods should give us the same result too.
# Method 2 - comparing test statistic with critical values (we should expect more extreme test statistic than critical value as we know we reject the null here)...these are on standard normal...
test_value <- Z
critical_value <- qnorm(p = .975,mean = 0,sd = 1 )
test_value
## [1] 2.297341
critical_value
## [1] 1.959964
Since the pvalue is less than alpha, we reject the null hypothesis - unlikely to see the sample mean we saw if the null is true.
Our environment is very sensitive to the amount of ozone in the upper atmosphere. The level of ozone normally found is 5.3 parts/million (ppm). A researcher believes that the current ozone level is at an insufficient level. The mean of 5 samples is 5.0 parts per million (ppm) with a standard deviation of 1.1. Does the data support the claim at the 0.05 level? Assume the population distribution is approximately normal.
Given: \(???\).
To Do: Researcher believes that the current ozone level is at an insufficient level - does the data support the claim at the 0.05 level ?
Ho: $$, Ha: $$
Ho: ???
Ha:???
$$
Distribution: ???
myp()
ACCEPT OR REJECT NULL???
Our environment is very sensitive to the amount of ozone in the upper atmosphere. The level of ozone normally found is 7.3 parts/million (ppm). A researcher believes that the current ozone level is not at a normal level. The mean of 51 samples is 7.1 ppm with a variance of 0.49. Assume the population is normally distributed. A level of significance of 0.01 will be used. Show all work and hypothesis testing steps.
Given: \(???\).
To Do: Researcher believes that the current ozone level is not at normal level. Thus, set a double sided hypothesis.
Ho: $$, Ha: $$
Ho:???
Ha:???
\(\alpha = .05\)
Distribution: t (unknown SD)
myp()
Accept/REJECT
A publisher reports that 36% of their readers own a laptop. A marketing executive wants to test the claim that the percentage is actually less than the reported percentage. A random sample of 100 found that 29% of the readers owned a laptop. Is there sufficient evidence at the 0.02 level to support the executive’s claim? Show all work and hypothesis testing steps.
Given: $= , n = , = ,= $.
To Do: Executive wants to test the claim that the percentage is actually less than the reported percentage. Thus, set a single sided hypothesis.
Ho: \(\pi ???\), Ha: \(\pi ???\)
Ho:Mean proportion of readers that own a laptop is greater that or equal to .36
Ha:Mean proportion of readers that own a laptop is less than .36
\(\alpha = .02\)
Distribution: Z (proportion)
myp()
DECION
A hospital director is told that 31% of the treated patients are uninsured. The director wants to test the claim that the percentage of uninsured patients is less than the expected percentage. A sample of 380 patients found that 95 were uninsured. Make the decision to reject or fail to reject the null hypothesis at the 0.05 level. Show all work and hypothesis testing steps.
Given: \(???\).
To Do: Researcher believes that the current ozone level is not at normal level. Thus, set a double sided hypothesis.
Ho: $$, Ha: $$
Ho:
Ha:
$$
Distribution: Z (proportion)
myp()
Strongly reject the null hypothesis that 31% or less people are insured.
OMITTED.
A standardized test is given to a sixth-grade class. Historically the mean score has been 112 with a standard deviation of 24. The superintendent believes that the standard deviation of performance may have recently decreased. She randomly sampled 22 students and found a mean of 102 with a standard deviation of 15.4387. Is there evidence that the standard deviation has decreased at the = 0.1 level? Show all work and hypothesis testing steps.
Given: $n , = ,= $.
To Do: Is there evidence that the standard deviation has decreased?
Ho: $$, Ha: $$
Ho:
Ha:
$= $
Distribution: Chi Squared (Random variable is the probability distribution of the sum of the squared errors, and it occurs naturally by squaring a normal distribution )
myp()
#Ho: sigma>=24, Ha: sigma<=24
ACCEPT/REJECT
The Satterthwaite formula for 2 sample t-test degrees of freedom is \(\dfrac{(\dfrac{s_1^2}{n_1}+\dfrac{s_2^2}{n_2})^2}{\dfrac{1}{n_1-1}(\dfrac{s_1^2}{n_1})^2+\dfrac{1}{n_2-1}(\dfrac{s_2^2}{n_2})^2}\), where \(s_1\) and \(n_1\) is the standard deviation and sample size of group 1, and \(s_2\) and \(n_2\) is the standard deviation and sample size of group 2.
Thus, you may for 8 and 9 select the smaller of the two degrees of freedom corresponding to each group.
A medical researcher wants to compare the pulse rates of smokers and non-smokers. He believes that the pulse rate for smokers and non-smokers is different and wants to test this claim at the 0.1 level of significance. The researcher checks 32 smokers and finds that they have a mean pulse rate of 87, and 31 non-smokers have a mean pulse rate of 84. The standard deviation of the pulse rates is found to be 9 for smokers and 10 for non-smokers. Let \(\mu_1\) be the true mean pulse rate for smokers and \(\mu_2\) be the true mean pulse rate for non-smokers. Show all work and hypothesis testing steps.
Let smoker group be indexed by 1, non-smoker group by 2.
Given: $n_1 = , _1 = , n_2 = , _2 = , _1 = , _2 = , = $.
To Do: Test if the pulse rate for smokers and non-smokers is different at the 0.1 level of significance. Thus, double sided test.
Ho: \(\mu_1 - \mu_2 = 0\), Ha: \(\mu_1 - \mu_2 \neq 0\)
Ho: ???
Ha: ???
\(\alpha = ???\)
Distribution: t
myp()
# Ho: Mu1-mu2=0, Ha: Mu1-Mu2<>0
mu1 <-
mu2 <-
alpha <-
# dist = t
n1 <-
n2 <-
df1 <- n1-1
df2 <- n2-1
sd1 <-
sd2 <-
var1 <-
var2 <-
num_point_estimate_diff <- (mu1 - mu2 ) # point estimate difference
den_Se <- sqrt( var1/n1 + var2/n2 ) # Se formula - Standard Error using sample standard deviations rather than population standard deviations
t <- num_point_estimate_diff / den_Se
numdf <- (var1/n1 + var2/n2)^2 # Satterthwaite
dendf <- (var1/n1)^2 / df1 + (var2/n2)^2 / df2 # Satterthwaite
df <- numdf / dendf # Satterthwaite - can be replaced with smaller of df1 or df2
shadet(df = df, pcts = c(.05,.95))
lines(rep(t,10), seq(0,1,length.out=10),col='red')
?pt # distribution function for the t distribution with df degrees of freedom
p_value = 2 * ( 1 - pt(q = t, df = numdf/dendf)) # Satterthwaite ## [1] 0.2160473
p_value
myp(p_value,alpha)
p_value_robust <- 2 * ( 1 - pt(q = t, df = min(df1, df2))) # smaller of the numerator and denominator degree of freedom
p_value_robust # a bit different p value, but the same end decision rule !!! ## [1] 0.220848
myp(p_value_robust,alpha)
???
Given two independent random samples with the following results: \(n_1, \bar{x}_1, \sigma_1, n_2, \bar{x}_2, \sigma_2\)
Use this data to find the 95% confidence interval for the true difference between the population means. Assume that the population variances are not equal and that the two populations are normally distributed.
To Do: Create a 95% confidence interval for true difference between the population means.
Ho: ${x}_1 - {x}_2 = ??? $, Ha: \(\bar{x}_1 - \bar{x}_2 ???\)
\(\alpha =\)
Distribution: t
myp()
alpha =
xbar1 = # mean
xbar2 = # mean
n1 = # sample size
n2 = # sample size
df1 = n1-1 # degrees of freedom
df2 = n2-1 # degrees of freedom
s1 = # sd
s2 = # sd
var1 = s1^2 # variance
var2 = s2^2 # variance
# Satterthwaite DF - can be replaced with smaller of df1 or df2
numdf = ( var1 / n1 + var2 / n2 )^2
dendf = ( var1 / n1 )^2 / df1 + (var2 / n2 )^2 / df2
df = numdf / dendf
df
delta = xbar1 - xbar2 # point estimate difference
delta
t = qt(p = .975, df = df) # two sided hypothesis test at 5% level of significance, p = vector of probabilities
t
Se = sqrt( var1/n1 + var2/n2 ) # Se formula - Standard Error using sample standard deviations rather than population standard deviations
Se
interval = c( delta - t * Se , delta + t * Se )
interval
CI is
Two men, A and B, who usually commute to work together decide to conduct an experiment to see whether one route is faster than the other. The men feel that their driving habits are approximately the same, so each morning for two weeks one driver is assigned to route I and the other to route II. The times, recorded to the nearest minute, are shown in the following table. Using this data, find the 98% confidence interval for the true mean difference between the average travel time for route I and the average travel time for route II.
Let \(d1 =\) (route I travel time) − (route II travel time).
Assume that the populations of travel times are normally distributed for both routes. Show all work and hypothesis testing steps.
To Do: Find the 98% confidence interval for the true mean difference between the average travel time for route I and the average travel time for route II.
Notation: Let \(d1 =\) (route I travel time) − (route II travel time).
Small Sample -> T distribution (Standard normal with fat tails)
r1 = c (32, 27, 34, 24, 31, 25, 30, 23, 27, 35)
r2 = c (28, 28, 33, 25, 26, 29, 33, 27, 25, 33)
delta = r1 - r2
# xbar +/- t * Se
xbar = mean(delta)
xbar
## [1] 0.1
n = 10
df = n-1
t = qt(p = .99, df = df)
Se = sd(delta) / sqrt( length(delta)
)
CI is ( , ).
The U.S. Census Bureau conducts annual surveys to obtain information on the percentage of the voting-age population that is registered to vote. Suppose that 391 employed persons and 510 unemployed persons are independently and randomly selected, and that 195 of the employed persons and 193 of the unemployed persons have registered to vote. Can we conclude that the percentage of employed workers (p1) who have registered to vote, exceeds the percentage of unemployed workers (p2) who have registered to vote? Use a significance level of 0.05 for the test. Show all work and hypothesis testing steps.
Q: Can we conclude that the percentage of employed workers (p1) who have registered to vote, exceeds the percentage of unemployed workers (p2) who have registered to vote?
Notation: Employed group is group 1 ; Unemployed group is group 2.
Given mean, sample size of both groups, alpha, single sided test, use proportion formula (for Se) from framing of question.
# Ho: pi1 - pi2 <= 0, Ha: pi1-pi2 > 0 (Percentage of employed workers p1 who have registered to vote, exceeds the percentage of unemployed workers p2 who have registered to vote)
# alpha=.05
# Z
n1 =
n2 =
x1 =
x2 =
p1 = x1/n1 # Estimate
p2 = x2/n2 # Estimate
p1-p2
Se = sqrt (p1 * (1-p1) / n1 + p2 * (1-p2) / n2) # textbook formula
Se
pbar = (x1 + x2) / (n1 + n2) # alternative way to get to the same SE
qbar = 1 - pbar
correction = 1 / n1 + 1 / n2
Se_robust = sqrt( pbar * qbar * correction )
# Pretty much the same
Se_robust # [1] 0.03328424
Se # [1] 0.03317529
Z = ( p1 - p2) / Se # textbook formula
Z
shadenorm(mu = 0, sig = Se, pcts = c(0.0,0.95))
lines(rep(p1-p2,10), seq(0,20,length.out=10),col='red')
FAIL TO / reject the null???
If we assume equal variance of the two groups, then we used the “pooled” method. If we do not assume equal variances of the groups, then we use the “Satterthwaite” method.↩︎