M. Drew LaMar
October 19, 2020
Question: Do populations differ in the variability of measurements?
Remember, it isn't always about inferring central tendency!
There are two main tests:
Example 12.4
The brook trout is a species native to eastern North America that has been introduced into streams in the West for sport fishing. Biologists followed the survivorship of a native species, chinook salmon, in a series of 12 streams that either had brook trout introduced or did not (Levin et al. 2002). Their goal was to determine whether the presence of brook trout effected the survivorship of the salmon. In each stream, they released a number of tagged juvenile chinook and then recorded whether or not each chinook survived over one year.
Load data and sneak-a-peek:
'data.frame': 12 obs. of 4 variables:
$ troutTreatment : chr "present" "absent" "present" "present" ...
$ nReleased : int 820 467 960 700 959 545 1029 769 27 998 ...
$ nSurvivors : int 166 180 136 153 178 103 326 173 7 120 ...
$ proportionSurvived: num 0.202 0.385 0.142 0.219 0.186 0.189 0.317 0.225 0.259 0.12 ...
Compute variances in both groups:
chinook %>%
group_by(troutTreatment) %>%
summarize(variance = var(proportionSurvived))
# A tibble: 2 x 2
troutTreatment variance
<chr> <dbl>
1 absent 0.0107
2 present 0.000883
var.test(proportionSurvived ~ troutTreatment,
data = chinook)
F test to compare two variances
data: proportionSurvived by troutTreatment
F = 12.165, num df = 5, denom df = 5, p-value = 0.01589
alternative hypothesis: true ratio of variances is not equal to 1
95 percent confidence interval:
1.702272 86.936360
sample estimates:
ratio of variances
12.16509
library(car)
leveneTest(chinook$proportionSurvived,
group = chinook$troutTreatment,
center = mean)
Levene's Test for Homogeneity of Variance (center = mean)
Df F value Pr(>F)
group 1 10.315 0.009306 **
10
---
Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
How to compare between two groups with only confidence intervals?
Example 12.5: Mommy's baby, Daddy's maybe
Question: Do babies look more like their fathers or their mothers?
Example 12.5: Mommy's baby, Daddy's maybe
Question: Do babies look more like their fathers or their mothers?
Christenfeld and Hill (1995) predicted that babies more resemble their fathers, due to the hypothesis that this resemblance affords an evolutionary advantage of increased paternal care. They tested this by obtaining pictures of a series of babies and their mothers and fathers. Particpants shown picture of child, and either three possible mothers or three possible fathers (one is correct).
Conclusion: Authors concluded that since fathers turned up statistically significant and mothers did not, that babies more resembled their fathers than their mothers.
Discuss: What’s the mistake here?
Mistake: Misinterpretation of statistical significance
Fallacy: If one test in Group 1 shows with statistical significance that \( \mu_{1} > \mu_{0} \), and the same test in Group 2 does
not show \( \mu_{2} > \mu_{0} \), then this shows with statistical significance that \( \mu_{1} > \mu_{2} \).
Fallacy: If one test in Group 1 shows with statistical significance that \( \mu_{1} > \mu_{0} \), and the same test in Group 2 does
not show \( \mu_{2} > \mu_{0} \), then this shows with statistical significance that \( \mu_{1} > \mu_{2} \).
Fallacy: If \( \bar{Y}_{1} > \bar{Y}_{2} \), then \( \mu_{1} > \mu_{2} \).
Mistake:Relying on point estimates rather than interval estimates
Conclusion: Comparisons between two groups should always be made directly using the appropriate statistical test, not indirectly by comparing both to the same null hypothesized value.
Four options for handling violations of assumptions:
Need to detect deviations first
To check for normality, first (as always) look at your data. Histograms work best here.
The following data come from a normal distribution:
They don't look normal, but they:
Examples of data from non-normal distributions:
Definition: The
normal quantile plot compares each observation in the sample with its quantile expected from the standard normal distribution. Points should fall roughly along a straight line if the data come from a normal distribution.
x <- sort(rnorm(20)) # (1)
p <- (1:20)/21 # (2)
q <- qnorm(p, lower.tail = TRUE) # (3)
plot(q ~ x, xlab="Measurements", ylab="Normal quantiles") # (4)
x <- sort(rnorm(20)) # (1)
p <- (1:20)/21 # (2)
q <- qnorm(p, lower.tail = TRUE) # (3)
plot(q ~ x, xlab="Measurements", ylab="Normal quantiles") # (4)
Fast way (note: axes are flipped by default!)
qqnorm(x, datax = TRUE)
Question: Are marine reserves effective in preserving marine wildlife?
Experimental design
Halpern (2003) matched 32 marine reserves to a control location, which was either the site of the reserve before it became protected or a similar unprotected site nearby. They then evaluated the “biomass ratio,” which is the ratio of total masses of all marine plants and animals per unit area of reserve in the protected and matched unprotected areas.
Experimental design
Halpern (2003) matched 32 marine reserves to a control location, which was either the site of the reserve before it became protected or a similar unprotected site nearby. They then evaluated the “biomass ratio,” which is the ratio of total masses of all marine plants and animals per unit area of reserve in the protected and matched unprotected areas.
Discuss: Observational or experimental? Paired or unpaired? Interpret response measure in terms of effect of protection.
Answer: Observational. Paired (matching). Biomass ratio = 1 (no effect); > 1 (beneficial effect); < 1 (detrimental effect).
Practice Problem #4: Interpret the following normal quantile plots.
Definition: A
Shapiro-Wilk test evaluates the goodness of fit of a normal distribution to a set of data randomly sampled from a population.
\( H_{0} \): The data are sampled from a population having a normal distribution.
\( H_{A} \): The data are sampled from a population not having a normal distribution.
Cautions:
marine <- read.csv("/Users/mdlama/Dropbox/Work/Teaching/College of William and Mary/Fall 2018/Datasets/chapter13/chap13e1MarineReserve.csv")
hist(marine$biomassRatio)
marine <- read.csv("/Users/mdlama/Dropbox/Work/Teaching/College of William and Mary/Fall 2018/Datasets/chapter13/chap13e1MarineReserve.csv")
shapiro.test(marine$biomassRatio)
Shapiro-Wilk normality test
data: marine$biomassRatio
W = 0.81751, p-value = 8.851e-05
Conclusion: Combination of graphical, testing, and common sense.
Definition: A statistical procedure is
robust if the answer it gives is not sensitive to violations of assumptions of the method.
Main takeaway point: This is a case-by-case basis that depends on the statistical test and data (see book for discussion).
Definition: A
data transformation changes each measurement by the same mathematical formula.
Common transformations:
Other transformations:
Hypothesis testing
marine <- read.csv("/Users/mdlama/Dropbox/Work/Teaching/College of William and Mary/Fall 2018/Datasets/chapter13/chap13e1MarineReserve.csv")
shapiro.test(log(marine$biomassRatio))
Shapiro-Wilk normality test
data: log(marine$biomassRatio)
W = 0.93795, p-value = 0.06551
Hypothesis testing
hist(log(marine$biomassRatio))
Original statistical hypotheses:
\( H_{0} \): The mean of the biomass ratio of marine reserves is one (\( \mu = 1 \))
\( H_{A} \): The mean of the biomass ratio of marine reserves is not one (\( \mu \neq 1 \))
Transformed statistical hypotheses:
\( H_{0} \): The mean of the log biomass ratio of marine reserves is zero (\( \mu^{\prime} = 0 \))
\( H_{A} \): The mean of the log biomass ratio of marine reserves is not zero (\( \mu^{\prime} \neq 0 \))
t.test(log(marine$biomassRatio), mu=0)
One Sample t-test
data: log(marine$biomassRatio)
t = 7.3968, df = 31, p-value = 2.494e-08
alternative hypothesis: true mean is not equal to 0
95 percent confidence interval:
0.3470180 0.6112365
sample estimates:
mean of x
0.4791272
Estimation
The 95% confidence interval for the log transformed data is
\[ 0.347 < \mu^{\prime} < 0.611. \]
For a 95% confidence interval of the untransformed data, we have
\[ e^{0.347} < \mathrm{geometric \ mean} < e^{0.611}, \]
or
\[ 1.41 < \mathrm{geometric \ mean} < 1.84. \]
Discuss: Conclusion?
Definition: A
nonparametric method makes fewer assumptions than standardparametric methods do about the distribution of the variables.
Property: Nonparametric methods are usually based on the
ranks of the data points (medians, quartiles, etc.)
Property: Nonparametric tests are typically
less powerful than parametric tests.
Definition: The
sign test compares the median of a sample to a constant specified in the null hypothesis. It makes no assumptions about the distribution of the measurements in the population.
Definition: The
Mann-Whitney \( U \)-test compares the distributions of two groups. It does not require as many assumptions as the two-sample \( t \)-test.
Algorithm:
Sign test has very little power. If \( n \leq 5 \), then can't use sign test.