C. Donovan
Power is intimately related to type 2 errors (false negatives)
It is our ability to detect differences when they actually exist
We look to address:
We've done much testing now.
We control Type 1 in our tests by setting our threshold \( p \)-value (type I error = \( \alpha \)).
There is nothing explicitly controlling type II error (\( \beta \)) in our tests.
Power is \( 1-\beta \) - the probability we correctly reject \( H_0 \)
Raises various questions:
Power calculations are often done to determine how much data to collect.
Armed with these, we can advise on \( n \) to meet specifications
Power calculations are often done to determine how much data to collect.
(Example modified from Larsen & Marx, 2001)
NB we've been given \( \sigma \) as known. If this were estimated, we'd have to use \( t \)-distributions.
qnorm(0.95)
[1] 1.644854
\[ 25 + 1.644854 \times \frac{2.4}{\sqrt{30}} \]
25+1.644854*2.4/sqrt(30)
[1] 25.72074
[I'll indulge in some laptop sketching again]
The decision boundary for \( \alpha = 0.05 \) (red)
The false negative area if \( \mu = 25.75 \) (blue)
The decision boundary for \( \alpha = 0.05 \) (red)
The false negative area if \( \mu = 25.75 \) (blue)
Calculate the blue false negative area in R:
pnorm(27.72, mean = 27.75, sd = 2.4/sqrt(30), lower.tail = T)
[1] 0.4727076
What happens if \( n \) is doubled?
decisionBound <- 25+1.644854*2.4/sqrt(60)
decisionBound
[1] 25.50964
What happens if \( n \) is doubled?
pnorm(decisionBound, mean = 25.75, sd = 2.4/sqrt(60))
[1] 0.2189452
pnorm(decisionBound, mean = 25*1.04, sd = 2.4/sqrt(60))
[1] 0.05675267
qnorm(0.9)
[1] 1.281552
decisionBound <- 25+qnorm(0.9)*2.4/sqrt(30)
decisionBound
[1] 25.56155
pnorm(decisionBound, mean = 25*1.03, sd = 2.4/sqrt(30))
[1] 0.3335682
With this in mind, consider multiple comparison adjustments
Implications for cynical exploitations:
We need this a lot:
\[ \mathbf{y} = \mathbf{X}\boldsymbol{\beta} + \mathbf{e} \]
leading to things like this
\[ \hat{\mathbf{y}} = \mathbf{X}\boldsymbol{\hat{\beta}} = \begin{bmatrix} 1 & x_{1,1} & x_{2,1}\\ 1 & x_{1,2} & x_{2,2}\\ \vdots & \vdots & \vdots\\ 1 & x_{1,n} & x_{2,n}\\ \end{bmatrix} \begin{bmatrix} \hat{\beta_0}\\ \hat{\beta_1}\\ \hat{\beta_2}\\ \end{bmatrix} \]
We've looked at
“Assessment”