class: middle background-image: url(data:image/png;base64,#LTU_logo.jpg) background-position: top left background-size: 30% # STM1001 [Topic 6](https://bookdown.org/content/f9d035ed-86ea-4779-ad01-31acc973f0dd/) Lecture ## `\(t\)`-tests for two-sample hypothesis testing ### La Trobe University This lecture complements the [Topic 6 readings](https://bookdown.org/content/f9d035ed-86ea-4779-ad01-31acc973f0dd/) --- # Topic 6: Related Links ## Readings [Topic 6 readings](https://bookdown.org/content/f9d035ed-86ea-4779-ad01-31acc973f0dd/) ## Notation [Topics 5 and 6: Hypothesis testing and `\(t\)`-tests](https://bookdown.org/a_shaker/STM1001_Topic_0/notation-summary.html#topics-5-and-6-hypothesis-testing-and-t-tests) --- # Topic 6: `\(t\)`-tests for two-sample hypothesis testing **Overview** <iframe src="https://bookdown.org/content/f9d035ed-86ea-4779-ad01-31acc973f0dd/" width="100%" height="400px" data-external="1"></iframe> --- # Today's Lecture Having learnt about the one-sample `\(t\)`-test in the previous topic, today we will be learning about two more types of `\(t\)`-tests: -- * The ***independent samples t-test*** (or two-sample `\(t\)`-test) -- * The ***paired t-test*** (or dependent samples `\(t\)`-test) -- We will learn about these tests via examples, including the following steps: -- * Visualising the data -- * Checking the assumptions -- * Carrying out the test -- To conclude, we will discuss ***effect sizes***, which help to determine the relative ***size*** of any differences found (as distinguished from ***statistical significance***). -- Remember, if you need to refresh your understanding of any Maths concepts, or notation introduced recently, you can check the [Maths and Notation Summary Guide](https://bookdown.org/a_shaker/STM1001_Topic_0/). --- name: stat class: middle background-image: url(data:image/png;base64,#slide_1.png) background-size: 110% --- name: stat class: middle background-image: url(data:image/png;base64,#slide_9.png) background-size: 100% --- # *t*-test versions The `\(t\)`-test framework we introduced in the previous topic is very flexible, and can be adapted to a variety of scenarios. -- There are three main versions of the `\(t\)`-test: -- 1. The ***one-sample t-test*** is used when we have one group, and assess one measurement from each individual in the group. -- * Our focus is typically on comparing results from the group (e.g. the sample mean) to a fixed reference value -- 2. The ***independent samples t-test*** (or two-sample `\(t\)`-test) is used when we have two independent groups, and assess one measurement from each individual in each group. -- * Our focus is typically on comparing the two groups to check for similarities or differences -- 3. The ***paired t-test*** (or dependent samples `\(t\)`-test) is used when we have taken two measurements of the same characteristic from each individual in a group, typically at two time points. -- * Our focus is typically on comparing the two sets of observations --- # Students' Eye Colour vs Sleep Time Example Suppose we made the following claim: *We believe that on average, STM1001 students with brown eyes spend either more or less time sleeping than people who do not have brown eyes* Further suppose that a sample of STM1001 students have been asked: -- * *In the past 24 hours, how many minutes did you spend sleeping?*, and -- * *What is your eye colour?* -- We can test this claim using a hypothesis test, just like we did with the sleep example in the previous lecture. -- However, since we have two independent groups of individuals here (students with brown eyes and students who don't have brown eyes), we will need to conduct a new version of the `\(t\)`-test, known as ***the independent samples t-test*** (aka the two-sample `\(t\)`-test). --- # Eye Colour/Sleep Example Hypotheses First, we need to set up our hypotheses: `$$H_0:\mu_1 = \mu_2\;\;\text{versus}\;\;H_1:\mu_1 \neq \mu_2,$$` where: -- * `\(\mu_1\)` denotes the population mean number of minutes STM1001 students with brown eyes spend sleeping per day -- * `\(\mu_2\)` denotes the population mean number of minutes STM1001 students who do not have brown eyes spend sleeping per day -- Note that, just like in the last topic, we are using `\(H_0\)` to denote the null hypothesis, and `\(H_1\)` to denote the alternate hypothesis. -- Note that if `\(\mu_1 = \mu_2\)`, this means that the difference between `\(\mu_1\)` and `\(\mu_2\)` is zero. So the above hypothesis could equivalently be written as: `$$H_0:\mu_1 - \mu_2 = 0\;\;\text{versus}\;\;H_1:\mu_1 - \mu_2 \neq 0.$$` --- # Independent samples `\(t\)`-test What does it mean to have ***two independent groups***, which is a requirement of conducting an independent-samples `\(t\)`-test? -- * One way of thinking of it would be that individuals can only be in one group or the other; individuals cannot be in both groups simultaneously -- * E.g. for this example, we assume a person belongs to the 'brown eyes' or 'other' group, but not both -- * This means the two groups are ***independent***, and appropriate for the independent-samples `\(t\)`-test --- # Variables: Independent samples `\(t\)`-test What type of variables are required for the independent samples `\(t\)`-test? -- .content-box-blue[ .center[ An **independent samples** *t*-test will always involve two variables: ] 1. The ***dependent*** variable, sometimes also called the *response* variable. This should be a numeric, continuous variable. 2. The ***independent*** variable. This should be a categorical variable with only ***two categories***. ] -- * So our ***dependent*** variable is minutes of sleep * Our ***independent*** variable is eye colour --- # Assumptions: Independent samples `\(t\)`-test The independent samples `\(t\)`-test assumptions are similar to those of the one-sample `\(t\)`-test, *with one addition*: ***equal variances*** between groups, also known as ***homogeneity of variance***. -- .content-box-blue[ .center[ **Independent samples *t*-test Assumptions:** ] 1. The data are numeric. 2. Observations are independent of one another * (that is, the sample is a simple random sample and each individual within the population has an equal chance of being selected) 3. The sample mean, `\(\overline{X}\)`, is normally distributed. 4. The two groups have equal variances * We can check this assumption using a statistical test that we will cover in more detail shortly ] -- * Normally, if the sd of one group is more than twice the value of the other group's sd, the equal variance assumption has been violated --- # Data Visualisation & Assumption Checking Before carrying out a hypothesis test, it is always a good idea to look at some descriptive statistics and plots. -- * This give us an idea what to expect when we carry out the test, and also check the assumptions -- | | | | | |:--------------|:---------------|:--------|:------| |**Eye Colour** |**Sample size** |**Mean** |**SD** | |Brown |51 |459.53 |154.37 | |Other |40 |441.88 |113.76 | --- <img src="data:image/png;base64,#Topic_6_Lecture_files/figure-html/unnamed-chunk-3-1.svg" width="80%" style="display: block; margin: auto;" /> --- # Data Visualisation & Assumption Checking From the descriptive statistics and plots, we can observe the following: 1. The boxplots and sample means indicate that the average sleep looks similar between groups. This can be deceiving though, and remember we are conducting a hypothesis test in order to make an inference about the population, not the just the sample. -- When we carry out the `\(t\)`-test, we will see whether or not there is a ***statistically significant*** difference. -- 2. From the boxplots, the data appear to be similarly spread out, with slightly more variation in the Brown group. The SD's are also fairly similar to each other (neither one is double the other). This indicates the equal variances assumption has (probably) not been violated. -- 3. The sample sizes for the *Brown Eyes* and *Other* groups are `\(n_{brown\, eyes} = 51\)` and `\(n_{other} =40\)` respectively. This will be useful knowledge later when checking for normality. --- # Levene's test for Equality of Variances To more formally assess equality of variance between groups, we can use the statistical test ***Levene's test for equality of variances***. -- .content-box-blue[ .center[ **Levene's test for Equality of Variances:** ] `$$H_0 : \text{The groups have equal variances}$$` `$$H_1 : \text{The groups do not have equal variances}$$` * If *p* < 0.05, equal variances cannot be assumed * If *p* > 0.05, equal variances can be assumed ] -- Since we start out by assuming the groups have equal variances, the test tells us to only reject this assumption if we get a small `\(p\)`-value. -- That is, a small `\(p\)`-value indicates the groups do not have equal variances. --- # Levene's test for Equality of Variances Let's carry out the Levene's test for the eye colour/sleep data: ``` Levene's Test for Homogeneity of Variance (center = median) Df F value Pr(>F) group 1 0.6788 0.4122 89 ``` As we can see, we have `\(p = 0.4122\)`. Since `\(p > 0.05\)`, equal variances can be assumed. Given our observations from the box plots and standard deviations, this is not a surprising result. -- #### What happens if the equal variances assumption is violated? * There are two versions of the independent samples `\(t\)`-test: one that assumes equal variances, and one that does not * If equal variances can be assumed, we use the version of the `\(t\)`-test that assumes equal variances * If equal variances cannot be assumed, we use the version of the `\(t\)`-test that does NOT assume equal variances We will have a chance to practise this in future computer labs. --- # Checking for normality Recall that we can consider histograms, Normal Q-Q plots, and the Shapiro-Wilk test to check for normality -- * For the independent samples `\(t\)`-test, this needs to be done for ***both groups individually*** --- <img src="data:image/png;base64,#Topic_6_Lecture_files/figure-html/unnamed-chunk-5-1.svg" width="70%" style="display: block; margin: auto;" /> Based on the histogram and Normal Q-Q plots, do you have any concerns regarding normality for either group? --- # Checking for normality Given there is some doubt following inspection of the histogram and Normal Q-Q plots, the Shapiro-Wilk test results can provide further guidance: **Shapiro-Wilk test for Brown eyes group:** ``` Shapiro-Wilk normality test data: sleep$Minutes[sleep$Eye_colour == "Brown"] W = 0.89414, p-value = 0.0002696 ``` **Shapiro-Wilk test for not Brown eyes group:** ``` Shapiro-Wilk normality test data: sleep$Minutes[sleep$Eye_colour == "Other"] W = 0.9426, p-value = 0.04234 ``` -- Given we have `\(p < 0.001\)` and `\(p = 0.0423\)` for each group respectively, it appears the normality assumption has been violated. -- However... --- # Checking for normality * Recall that the sample size for each group is 51 and 40 respectively. -- * Also recall that the underlying distribution does not have to be normally distributed to satisfy the assumption - it is the sample mean that should be normally distributed -- * Given `\(n > 30\)` for both groups, we can therefore apply the Central Limit Theorem and conclude that the **normality assumption has been met** We are now ready to carry out the independent samples `\(t\)`-test. --- # Independent samples `\(t\)`-test output ``` r Two Sample t-test data: sleep$Minutes by sleep$Eye_colour t = 0.60549, df = 89, p-value = 0.5464 alternative hypothesis: true difference in means between group Brown and group Other is not equal to 0 95 percent confidence interval: -40.28041 75.58923 sample estimates: mean in group Brown mean in group Other 459.5294 441.8750 ``` --- # Independent samples `\(t\)`-test output ``` r Two Sample t-test data: sleep$Minutes by sleep$Eye_colour t = 0.60549, df = 89, `p-value = 0.5464` alternative hypothesis: true difference in means between group Brown and group Other is not equal to 0 95 percent confidence interval: -40.28041 75.58923 sample estimates: mean in group Brown mean in group Other 459.5294 441.8750 ``` * The ** `\(p\)`-value** is 0.5464, which is greater than 0.05, so we cannot reject `\(H_0\)`. That is, do not we have enough evidence to conclude that the difference between groups is statistically significant. --- # Independent samples `\(t\)`-test output ``` r Two Sample t-test data: sleep$Minutes by sleep$Eye_colour t = 0.60549, df = 89, p-value = 0.5464 alternative hypothesis: true difference in means between group Brown and group Other is not equal to 0 `95 percent confidence interval:` `-40.28041 75.58923` sample estimates: mean in group Brown mean in group Other 459.5294 441.8750 ``` * The **95% confidence interval** for `\(\mu_1 - \mu_2\)` is (-40.28, 75.59), meaning we are 95% confident that the population mean sleep of the Brown group is between -40.28 and 75.59 minutes higher than that of the Other group. Since this interval includes 0, we cannot reject `\(H_0\)`. --- # Independent samples `\(t\)`-test output ``` r Two Sample t-test data: sleep$Minutes by sleep$Eye_colour t = 0.60549, df = 89, p-value = 0.5464 alternative hypothesis: true difference in means between group Brown and group Other is not equal to 0 `95 percent confidence interval:` `-40.28041 75.58923` sample estimates: mean in group Brown mean in group Other 459.5294 441.8750 ``` * To understand why ***we cannot reject `\(H_0\)` based on the confidence interval result***, recall: `\(H_0\)` represents the null hypothesis that `\(\mu_1 = \mu_2\)`. If `\(\mu_1\)` and `\(\mu_2\)` were equal, the difference between them would be 0. Since we are 95% confident that the true difference between `\(\mu_1\)` and `\(\mu_2\)` is between (-40.28, 75.59), a range which includes 0, we do not have enough evidence to conclude that the difference is statistically significant and we cannot reject `\(H_0\)`. --- # Independent samples `\(t\)`-test output ``` r Two Sample t-test data: sleep$Minutes by sleep$Eye_colour `t = 0.60549`, df = 89, p-value = 0.5464 alternative hypothesis: true difference in means between group Brown and group Other is not equal to 0 95 percent confidence interval: -40.28041 75.58923 sample estimates: mean in group Brown mean in group Other 459.5294 441.8750 ``` * The **test statistic** is `\(t = 0.6055\)`. --- # Independent samples `\(t\)`-test output ``` r Two Sample t-test data: sleep$Minutes by sleep$Eye_colour `t = 0.60549`, `df = 89`, p-value = 0.5464 alternative hypothesis: true difference in means between group Brown and group Other is not equal to 0 95 percent confidence interval: -40.28041 75.58923 sample estimates: mean in group Brown mean in group Other 459.5294 441.8750 ``` * The **test statistic** is `\(t = 0.6055\)` * The **degrees of freedom** value is 89. --- # Independent samples `\(t\)`-test output ``` r Two Sample t-test data: sleep$Minutes by sleep$Eye_colour `t = 0.60549`, `df = 89`, p-value = 0.5464 alternative hypothesis: true difference in means between group Brown and group Other is not equal to 0 95 percent confidence interval: -40.28041 75.58923 `sample estimates:` `mean in group Brown mean in group Other ` `459.5294` `441.8750` ``` * The **test statistic** is `\(t = 0.6055\)` * The **degrees of freedom** is 89 * The **sample means** are `\(459.53\)` and `\(441.88\)` respectively. --- # Other types of alternative hypotheses So far, we have only considered the two-sided hypothesis test version for our independent samples `\(t\)`-test, i.e.: `$$H_0:\mu_1 = \mu_2\;\;\text{versus}\;\;H_1:\mu_1 \neq \mu_2,$$` -- or equivalently, `$$H_0:\mu_1 - \mu_2 = 0\;\;\text{versus}\;\;H_1:\mu_1 - \mu_2 \neq 0.$$` -- One-sided tests are also possible. -- For example, `$$H_0:\mu_1 - \mu_2 = 0\;\;\text{versus}\;\;H_1:\mu_1 - \mu_2 < 0.$$` or `$$H_0:\mu_1 - \mu_2 = 0\;\;\text{versus}\;\;H_1:\mu_1 - \mu_2 > 0.$$` --- # Other types of alternative hypotheses We can also test for `\(\mu_0 = \mu_1 - \mu_2\)` values **other than 0**. For example: -- `$$H_0:\mu_1 - \mu_2 = 5\;\;\text{versus}\;\;H_1:\mu_1 - \mu_2 \neq 5,$$` -- or `$$H_0:\mu_1 - \mu_2 = 5\;\;\text{versus}\;\;H_1:\mu_1 - \mu_2 < 5,$$` -- or `$$H_0:\mu_1 - \mu_2 = 5\;\;\text{versus}\;\;H_1:\mu_1 - \mu_2 > 5.$$` --- name: menti class: middle background-image: url(data:image/png;base64,#menti.jpg) background-size: 115% # Kahoot ## Go to [www.kahoot.it](https://www.kahoot.it) and use ## the code provided --- # The paired `\(t\)`-test The paired `\(t\)`-test (or *dependent samples `\(t\)`-test*) can be used when we have two samples, with the same set of subjects in each sample. -- * For each subject, we have two measurements of the same characteristic: often at different time-points or under different conditions -- When conducting a paired `\(t\)`-test, we typically use the following hypotheses: `$$H_0:\mu_D = 0\;\;\text{versus}\;\;H_1:\mu_D \neq 0,$$` where: * `\(\mu_D\)` is defined as the true (population) mean difference between the two time-points/conditions considered -- If `\(\mu_D = 0\)`, this effectively means there's no difference at the population level between the two time-points/conditions for the characteristic being considered. --- # Variables: Paired `\(t\)`-test What type of variables are required for the paired `\(t\)`-test? -- .content-box-blue[ .center[ A ***paired t-test*** will always involve two variables: ] 1. The ***dependent*** variable, sometimes also called the *response* variable. This should be a numeric, continuous variable. 2. The ***independent*** variable. This should be a categorical variable with only ***two categories*** which represent before/after categories, or two different conditions. ] --- background-image: url(data:image/png;base64,#Roasted_coffee_beans_translucent.jpg) background-size: cover # Coffee Break Example As an example aplication of the paired `\(t\)`-test, let's check if coffee breaks affect productivity. -- * We will consider a data set called `Coffee` from the R package `BSDA` (Arnholt and Evans, 2023; Kitchens, 2003) -- * In this example, workers' productivity scores were measured with and without coffee breaks <!-- public domain image https://commons.wikimedia.org/wiki/File:Roasted_coffee_beans.jpg --> -- So for this example: * Our ***dependent*** variable is productivity score * Our ***independent*** variable is coffee break (with or without) --- # Coffee Break Example: Hypotheses * We will test whether the mean difference in productivity with and without a coffee break is statistically significant different by carrying out a paired `\(t\)`-test with the following hypotheses: `$$H_0:\mu_D = 0\;\;\text{versus}\;\;H_1:\mu_D \neq 0,$$` where: * `\(\mu_D\)` is defined as the true (population) mean difference in productivity with and without a coffee break --- # Assumptions: Paired samples `\(t\)`-test The assumptions of the paired `\(t\)`-test are exactly the same as those of the one-sample `\(t\)`-test (introduced in the last lecture): .content-box-blue[ .center[ **Paired samples *t*-test Assumptions:** ] 1. The data are numeric 2. Observations are independent of one another * (that is, the sample is a simple random sample and each individual within the population has an equal chance of being selected) 3. The sample mean, `\(\overline{X}\)`, is normally distributed. ] --- # Assumptions: Paired samples `\(t\)`-test When checking assumptions for the paired `\(t\)`-test, the variable of interest is the ***paired differences*** rather than the 'before' and 'after' variables themselves. -- For example, consider the following snapshot of the `Coffee` data set: ``` without with differences 1 23 28 5 2 35 38 3 3 29 29 0 4 33 37 4 5 43 42 -1 6 32 30 -2 ``` -- The paired `\(t\)`-test is testing whether the average of the ***paired differences*** is equal to zero. For this reason, it is this variable that is of interest when checking assumptions. --- # Data Visualisation & Assumption Checking We will again begin by considering descriptive statistics and plots: <br> <br> | | | | | |:--------------------|:---------------|:--------|:------| |**Variable** |**Sample size** |**Mean** |**SD** | |Without Coffee Break |9 |34.89 |6.39 | |With Coffee Break |9 |35.89 |5.58 | |Paired differences |9 |1 |2.55 | --- <img src="data:image/png;base64,#Topic_6_Lecture_files/figure-html/unnamed-chunk-18-1.svg" width="80%" style="display: block; margin: auto;" /> --- # Data Visualisation & Assumption Checking From the descriptive statistics and plots, we can observe the following: 1. The box plots and sample means respectively indicate that both the median and mean productivity score **with** a coffee break are higher. When we carry out the paired `\(t\)`-test, we can check whether or not this difference is ***statistically significant*** -- 2. The mean paired difference is 1, with a standard deviation of 2.55. The paired `\(t\)`-test will be testing to see whether this result is significantly different from 0. -- 3. The sample size of the paired differences is 9. This will be useful knowledge when checking for normality. --- # Checking for normality We will again consider histograms, Normal Q-Q plots, and the Shapiro-Wilk test to check for normality. -- * As mentioned earlier, we will check for normality of the *differences* between the two conditions --- # Checking for normality <img src="data:image/png;base64,#Topic_6_Lecture_files/figure-html/unnamed-chunk-19-1.svg" style="display: block; margin: auto;" /> Based on the histogram and Normal Q-Q plots, do you have any concerns regarding normality for either group? --- # Checking for normality There is potentially some skewness present. However, given the small sample size, it is difficult to tell whether or not these data are sampled from a normal distribution. -- Let's see what the Shapiro-Wilk test shows: **Shapiro-Wilk test for paired differences:** ``` Shapiro-Wilk normality test data: Coffee$differences W = 0.89448, p-value = 0.2217 ``` -- Given we have `\(p = 0.2217\)` which is greater than 0.05, we do not have evidence against normality (recall the null hypothesis for the Shapiro-Wilk test is that the data are normal. -- Therefore, for the purposes of this example, **we will assume that the normality assumption has not been violated.** -- We are now ready to carry out the paired `\(t\)`-test. --- # Coffee Break Example: Paired `\(t\)`-test output ``` r Paired t-test data: Coffee$with and Coffee$without t = 1.1767, df = 8, p-value = 0.2731 alternative hypothesis: true difference in means is not equal to 0 95 percent confidence interval: -0.9597267 2.9597267 sample estimates: mean of the differences 1 ``` --- # Coffee Break Example: Paired `\(t\)`-test output ``` r Paired t-test data: Coffee$with and Coffee$without `t = 1.1767`, `df = 8`, p-value = 0.2731 alternative hypothesis: true difference in means is not equal to 0 95 percent confidence interval: -0.9597267 2.9597267 `sample estimates:` `mean of the differences` `1` ``` * The **test statistic** is `\(t = 1.1767\)` * The **degrees of freedom** is `\(8\)` * The **sample mean** of the paired differences is `\(1\)`. --- # Coffee Break Example: Paired `\(t\)`-test output ``` r Paired t-test data: Coffee$with and Coffee$without t = 1.1767, df = 8, `p-value = 0.2731` alternative hypothesis: true difference in means is not equal to 0 95 percent confidence interval: -0.9597267 2.9597267 sample estimates: mean of the differences 1 ``` * The ** `\(p\)`-value** is 0.2731, which is greater than 0.05, so we cannot reject `\(H_0\)`. That is, we do not have enough evidence to conclude that the difference in productivity is different depending on coffee break --- # Coffee Break Example: Paired `\(t\)`-test output ``` r Paired t-test data: Coffee$with and Coffee$without t = 1.1767, df = 8, p-value = 0.2731 alternative hypothesis: true difference in means is not equal to 0 `95 percent confidence interval:` `-0.9597267 2.9597267` sample estimates: mean of the differences 1 ``` * The **95% confidence interval** for `\(\mu_D\)` is (-0.96, 2.96), meaning we are 95% confident that the population mean difference in productivity with a coffee break compared to without is between 0.96 and -2.96. Since this interval includes 0, we cannot reject `\(H_0\)`. --- # Paired and one-sample `\(t\)`-tests link It turns out that the paired and one-sample `\(t\)`-tests are actually equivalent! -- * See [this topic's readings](https://bookdown.org/content/f9d035ed-86ea-4779-ad01-31acc973f0dd/2.4-link-between-paired-and-one-sample-t-tests.html) if you are interested to learn why -- * You will also have a chance to try this for yourself in the computer lab --- #Independent samples vs. Paired `\(t\)`-test While independent samples `\(t\)`-tests and paired `\(t\)`-tests both involve using data from two groups, we have to be careful to only apply them in appropriate contexts. -- * One key concept to remember for an **independent** samples `\(t\)`-test is that we must have ***two independent groups*** -- * This would normally mean that it is not possible for the same person (or subject/observation if the study is not about people) to be in both groups (e.g. brown eyes vs. not brown eyes) -- * On the other hand, a **paired** `\(t\)`-test tests for a difference when we have the **same set of subjects in each sample** -- * This would mean we have ***two dependent groups*** -- * One way of thinking of it would be that all individuals must be in both groups --- # Effect sizes When we conduct our tests, we want to verify that our results are not only statistically significant but also clinically meaningful. -- * An ***effect size*** tells us the ***relative size*** of any difference observed -- * The ***effect size*** can be thought of as the ***standardised difference in mean*** -- * For `\(t\)`-tests, ***Cohen's `\(d\)`*** (Cohen, 1988) is typically used and can be calculated as $$ d = \displaystyle \frac{\bar{x} - \mu_0}{s},$$ where * `\(\bar{x}\)` denotes the sample mean * `\(\mu_0\)` denotes the mean under the null hypothesis * `\(s\)` denotes the sample standard deviation. --- # Effect sizes Cohen (1988) provided a guide to quantify the magnitude of an effect size: .content-box-blue[ .center[ **Guidelines for quantifying magnitude of Cohen's *d* effect sizes for *t*-tests:** ] * `\(|d| < 0.2\)`: "negligible" * `\(0.2 \leq |d| < 0.5\)`: "small" * `\(0.5 \leq |d| < 0.8\)`: "medium" * `\(|d| \geq 0.8\)`: "large" ] -- Thankfully, most computer packages will do the hard work for us. We will consider some examples in the upcoming computer labs. --- # References Arnholt, A. T. and B. Evans (2023). _BSDA: Basic Statistics and Data Analysis_. R package version 1.2.2. URL: [https://CRAN.R-project.org/package=BSDA](https://CRAN.R-project.org/package=BSDA). Cohen, J. (1988). _Statistical power analysis for the behavioral sciences_. 2nd edition. New York: Academic Press. Kitchens, L. J. (2003). _Basic Statistics and Data Analysis_. Pacific Grove, CA,: Brooks/Cole, a division of Thomson Learning. --- background-image: url(data:image/png;base64,#computerlab.jpg) background-position: bottom background-size: 75% class: center # See you in the computer labs! --- class: middle <font color = "grey"> These notes have been prepared by Amanda Shaker and Rupert Kuveke. The copyright for the material in these notes resides with the authors named above, with the Department of Mathematics and Statistics and with La Trobe University. Copyright in this work is vested in La Trobe University including all La Trobe University branding and naming. Unless otherwise stated, material within this work is licensed under a Creative Commons Attribution-Non Commercial-Non Derivatives License <a href = "https://creativecommons.org/licenses/by-nc-nd/4.0/" target="_blank"> BY-NC-ND. </a> </font>