class: middle background-image: url(data:image/png;base64,#LTU_logo.jpg) background-position: top left background-size: 30% # STM1001 [Topic 6](https://amandashaker-stm1001-topic-6.share.connect.posit.cloud/) Lecture ## `\(t\)`-tests for two-sample hypothesis testing ### La Trobe University This lecture complements the [Topic 6 readings](https://amandashaker-stm1001-topic-6.share.connect.posit.cloud/) --- # Topic 6: Related Links ## Readings [Topic 6 readings](https://amandashaker-stm1001-topic-6.share.connect.posit.cloud/) ## Notation [Topics 5 and 6: Hypothesis testing and `\(t\)`-tests](https://amandashaker-stm1001-topic-0.share.connect.posit.cloud/notation-summary.html#topics-5-and-6-hypothesis-testing-and-t-tests) --- # Topic 6: `\(t\)`-tests for two-sample hypothesis testing **Overview** <iframe src="https://amandashaker-stm1001-topic-6.share.connect.posit.cloud/" width="100%" height="400px" data-external="1"></iframe> --- # Today's Lecture Having learnt about the one-sample `\(t\)`-test in the previous topic, today we will be learning about two more types of `\(t\)`-tests: -- * The ***independent samples t-test*** (or two-sample `\(t\)`-test) -- * The ***paired t-test*** (or dependent samples `\(t\)`-test) -- We will learn about these tests via examples, including the following steps: -- * Visualising the data -- * Checking the assumptions -- * Carrying out the test -- To conclude, we will discuss ***effect sizes***, which help to determine the relative ***size*** of any differences found (as distinguished from ***statistical significance***). -- Remember, if you need to refresh your understanding of any Maths concepts, or notation introduced recently, you can check the [Maths and Notation Summary Guide](https://amandashaker-stm1001-topic-0.share.connect.posit.cloud/index.html). --- name: stat class: middle background-image: url(data:image/png;base64,#slide_1.png) background-size: 110% --- name: stat class: middle background-image: url(data:image/png;base64,#slide_9.png) background-size: 100% --- # Three types of *t*-tests In the previous topic, we learnt about the ***one-sample `\(t\)`-test***. Today, we will cover two more types of `\(t\)`-tests. -- The three versions of the `\(t\)`-test are: -- 1. The ***one-sample `\(t\)`-test*** is used when we have one group, and assess one measurement from each individual in the group. We compare results from the group (i.e. the sample mean) to a fixed reference value -- 2. The ***independent samples `\(t\)`-test*** (or two-sample `\(t\)`-test) is used when we have two independent groups, and assess one measurement from each individual in each group. We compare the two groups to check for similarities or differences -- 3. The ***paired `\(t\)`-test*** (or dependent samples `\(t\)`-test) is used when we have taken two measurements of the same characteristic from each individual in a group, typically at two time points. We compare the two sets of observations --- # Students' Eye Colour vs Sleep Time Example Suppose we made the following claim: *We believe that on average, STM1001 students with brown eyes spend either more or less time sleeping than people who do not have brown eyes* Further suppose that a sample of STM1001 students have been asked: -- * *In the past 24 hours, how many minutes did you spend sleeping?*, and -- * *What is your eye colour?* -- We can test this claim using a hypothesis test, just like we did with the sleep example in the previous lecture. -- However, since we have two independent groups of individuals here (students with brown eyes and students who don't have brown eyes), we will need to use the ***independent samples `\(t\)`-test*** (aka the two-sample `\(t\)`-test). --- # Eye Colour/Sleep Example Hypotheses * First, we need to set up our hypotheses: `$$H_0:\mu_1 = \mu_2\;\;\text{versus}\;\;H_1:\mu_1 \neq \mu_2,$$` where: -- * `\(\mu_1\)` denotes the population mean number of minutes STM1001 students with brown eyes spend sleeping per day * `\(\mu_2\)` denotes the population mean number of minutes STM1001 students who do not have brown eyes spend sleeping per day -- Note that if `\(\mu_1 = \mu_2\)`, this means that the difference between `\(\mu_1\)` and `\(\mu_2\)` is zero. So the above hypothesis could equivalently be written as: `$$H_0:\mu_1 - \mu_2 = 0\;\;\text{versus}\;\;H_1:\mu_1 - \mu_2 \neq 0.$$` --- # Independent samples `\(t\)`-test What does it mean to have ***two independent groups***, as we need to have to carry out an independent-samples `\(t\)`-test? * One way of thinking of it would be that individuals can only be in one group or the other: not both. -- * E.g. for this example, we assume a person belongs to the 'brown eyes' or 'other' group, but not both. * This means the two groups are ***independent***, and appropriate for the independent-samples `\(t\)`-test. --- # Independent samples `\(t\)`-test What type of variables are required for the independent samples `\(t\)`-test? -- .content-box-blue[ .center[ An independent samples *t*-test will always involve two variables: ] 1. The ***dependent*** variable, sometimes also called the *response* variable. This should be a numeric, continuous variable. 2. The ***independent*** variable. This should be a categorical variable with only ***two categories***. ] -- * So our ***dependent*** variable is minutes of sleep * Our ***independent*** variable is eye colour --- # Assumptions: Independent samples `\(t\)`-test The assumptions for the independent samples `\(t\)`-test are similar to those we saw for the one-sample `\(t\)`-test, with one addition: ***equal variances*** between groups, also known as ***homogeneity of variance***. In summary: .content-box-blue[ .center[ **Independent samples *t*-test Assumptions:** ] 1. The data are numeric 2. Observations are independent of one another (that is, the sample is a simple random sample and each individual within the population has an equal chance of being selected) 3. The sample mean, `\(\overline{X}\)`, is normally distributed 4. Equal variances between groups (aka homogeneity of variances). ] -- * Normally, if the standard deviation of one group is more than twice that of the other group, the equal variance assumption has been violated * There is also a statistical test we can use to help us check the equal variances assumption. We will cover this in more detail shortly --- # Visualising the data and checking assumptions * Before carrying out a hypothesis test, it is always a good idea to look at some descriptive statistics and plots -- * This give us an idea what to expect when we carry out the test, and also check the assumptions -- | | | | | |:--------------|:---------------|:--------|:------| |**Eye Colour** |**Sample size** |**Mean** |**SD** | |Brown |51 |459.53 |154.37 | |Other |40 |441.88 |113.76 | --- <img src="data:image/png;base64,#Topic_6_Lecture_files/figure-html/unnamed-chunk-3-1.svg" width="80%" style="display: block; margin: auto;" /> --- # Visualising data & checking assumptions From the descriptive statistics and plots, we can observe the following: 1. The boxplots and sample means indicate that the average sleep looks similar between groups. When we carry out the `\(t\)`-test, we will see whether or not there is a ***statistically significant*** difference -- 2. From the boxplots, the data appear to be similarly spread out, with slightly more variation in the Brown group. The SD's are also fairly similar to each other (neither one is double the other). This indicates the equal variances assumption has (probably) not been violated. -- 3. The sample size in the Brown and Other groups are 51 and 40 respectively. This will be useful knowledge later when checking for normality. --- # Levene's test for equality of variances To more formally assess equality of variance between groups, we can use the ***Levene's test for equality of variances***. -- Consider the following null and alternative hypotheses: * `\(H_0 : \text{The groups have equal variances}\)` * `\(H_1 : \text{The groups do not have equal variances}.\)` -- Since we start out by assuming the groups have equal variances, the test tells us to only reject this assumption if we get a small `\(p\)`-value. That is, a small `\(p\)`-value indicates the groups do not have equal variances. To summarise: .content-box-blue[ .center[ **Levene's test for equality of variances:** ] * If `\(p\)` < 0.05, equal variances cannot be assumed * If `\(p\)` > 0.05, equal variances can be assumed ] --- # Levene's test for equality of variances Let's carry out the Levene's test for the sleep / eye colour data: ``` Levene's Test for Homogeneity of Variance (center = median) Df F value Pr(>F) group 1 0.6788 0.4122 89 ``` As we can see, we have `\(p = 0.4122\)`. Since `\(p > 0.05\)`, equal variances can be assumed. Given our observations from the box plots and standard deviations, this is not a surprising result. -- #### What happens if the equal variances assumption is violated? * There are two versions of the independent samples `\(t\)`-test: one that assumes equal variances, and one that does not * If equal variances can be assumed, we use the version of the `\(t\)`-test that assumes equal variances * If equal variances cannot be assumed, we use the version of the `\(t\)`-test that does NOT assume equal variances. We will have a chance to practise this in the computer lab. --- # Checking for normality * Recall that we can consider histograms, Normal Q-Q plots, and the Shapiro-Wilk test to check for normality -- * For the independent samples `\(t\)`-test, this needs to be done for both groups individually --- <img src="data:image/png;base64,#Topic_6_Lecture_files/figure-html/unnamed-chunk-5-1.svg" width="70%" style="display: block; margin: auto;" /> Based on the histogram and Normal Q-Q plots, do you have any concerns regarding normality for either group? --- # Checking for normality Given there is some doubt following inspection of the histogram and Normal Q-Q plots, the Shapiro-Wilk test results can provide further guidance: **Shapiro-Wilk test for Brown eyes group:** ``` Shapiro-Wilk normality test data: sleep$Minutes[sleep$Eye_colour == "Brown"] W = 0.89414, p-value = 0.0002696 ``` **Shapiro-Wilk test for not Brown eyes group:** ``` Shapiro-Wilk normality test data: sleep$Minutes[sleep$Eye_colour == "Other"] W = 0.9426, p-value = 0.04234 ``` -- Given we have `\(p < 0.001\)` and `\(p = 0.0423\)` for each group respectively, it appears the normality assumption has been violated. -- However... --- # Checking for normality * Recall that the sample size for each group is 51 and 40 respectively. -- * Also recall that the underlying distribution does not have to be normally distributed to satisfy the assumption - it is the sample mean that should be normally distributed -- * Given `\(n > 30\)` for both groups, we can therefore apply the Central Limit Theorem and conclude that the **normality assumption has been met** We are now ready to carry out the independent samples `\(t\)`-test. --- # Independent samples `\(t\)`-test output ``` r Two Sample t-test data: sleep$Minutes by sleep$Eye_colour t = 0.60549, df = 89, p-value = 0.5464 alternative hypothesis: true difference in means between group Brown and group Other is not equal to 0 95 percent confidence interval: -40.28041 75.58923 sample estimates: mean in group Brown mean in group Other 459.5294 441.8750 ``` --- # Independent samples `\(t\)`-test output ``` r Two Sample t-test data: sleep$Minutes by sleep$Eye_colour t = 0.60549, df = 89, `p-value = 0.5464` alternative hypothesis: true difference in means between group Brown and group Other is not equal to 0 95 percent confidence interval: -40.28041 75.58923 sample estimates: mean in group Brown mean in group Other 459.5294 441.8750 ``` * The ** `\(p\)`-value** is 0.5464, which is greater than 0.05, so we cannot reject `\(H_0\)`. That is, do not we have enough evidence to conclude that the difference between groups is statistically significant. --- # Independent samples `\(t\)`-test output ``` r Two Sample t-test data: sleep$Minutes by sleep$Eye_colour t = 0.60549, df = 89, p-value = 0.5464 alternative hypothesis: true difference in means between group Brown and group Other is not equal to 0 `95 percent confidence interval:` `-40.28041 75.58923` sample estimates: mean in group Brown mean in group Other 459.5294 441.8750 ``` * The **95% confidence interval** for `\(\mu_1 - \mu_2\)` is (-40.28, 75.59), meaning we are 95% confident that the population mean sleep of the Brown group is between -40.28 and 75.59 minutes higher than that of the Other group. Since this interval includes 0, we cannot reject `\(H_0\)`. --- # Independent samples `\(t\)`-test output ``` r Two Sample t-test data: sleep$Minutes by sleep$Eye_colour t = 0.60549, df = 89, p-value = 0.5464 alternative hypothesis: true difference in means between group Brown and group Other is not equal to 0 `95 percent confidence interval:` `-40.28041 75.58923` sample estimates: mean in group Brown mean in group Other 459.5294 441.8750 ``` * To understand why ***we cannot reject `\(H_0\)` based on the confidence interval result***, recall: `\(H_0\)` represents the null hypothesis that `\(\mu_1 = \mu_2\)`. If `\(\mu_1\)` and `\(\mu_2\)` were equal, the difference between them would be 0. Since we are 95% confident that the true difference between `\(\mu_1\)` and `\(\mu_2\)` is between (-40.28, 75.59), a range which includes 0, we do not have enough evidence to conclude that the difference is statistically significant and we cannot reject `\(H_0\)`. --- # Independent samples `\(t\)`-test output ``` r Two Sample t-test data: sleep$Minutes by sleep$Eye_colour `t = 0.60549`, df = 89, p-value = 0.5464 alternative hypothesis: true difference in means between group Brown and group Other is not equal to 0 95 percent confidence interval: -40.28041 75.58923 sample estimates: mean in group Brown mean in group Other 459.5294 441.8750 ``` * The **test statistic** is `\(t = 0.6055\)` --- # Independent samples `\(t\)`-test output ``` r Two Sample t-test data: sleep$Minutes by sleep$Eye_colour `t = 0.60549`, `df = 89`, p-value = 0.5464 alternative hypothesis: true difference in means between group Brown and group Other is not equal to 0 95 percent confidence interval: -40.28041 75.58923 sample estimates: mean in group Brown mean in group Other 459.5294 441.8750 ``` * The **test statistic** is `\(t = 0.6055\)` * The **degrees of freedom** is 89 --- # Independent samples `\(t\)`-test output ``` r Two Sample t-test data: sleep$Minutes by sleep$Eye_colour `t = 0.60549`, `df = 89`, p-value = 0.5464 alternative hypothesis: true difference in means between group Brown and group Other is not equal to 0 95 percent confidence interval: -40.28041 75.58923 `sample estimates:` `mean in group Brown mean in group Other ` `459.5294` `441.8750` ``` * The **test statistic** is `\(t = 0.6055\)` * The **degrees of freedom** is 89 * The **sample means** are `\(459.53\)` and `\(441.88\)` respectively. --- # Other types of alternative hypotheses So far, we have only considered the two-sided hypothesis test: `$$H_0:\mu_1 = \mu_2\;\;\text{versus}\;\;H_1:\mu_1 \neq \mu_2,$$` -- or equivalently, `$$H_0:\mu_1 - \mu_2 = 0\;\;\text{versus}\;\;H_1:\mu_1 - \mu_2 \neq 0.$$` -- One-sided tests are also possible. -- For example, `$$H_0:\mu_1 - \mu_2 = 0\;\;\text{versus}\;\;H_1:\mu_1 - \mu_2 < 0.$$` or `$$H_0:\mu_1 - \mu_2 = 0\;\;\text{versus}\;\;H_1:\mu_1 - \mu_2 > 0.$$` --- # Other types of alternative hypotheses We can also test for `\(\mu_0 = \mu_1 - \mu_2\)` values **other than 0**. For example: -- `$$H_0:\mu_1 - \mu_2 = 5\;\;\text{versus}\;\;H_1:\mu_1 - \mu_2 \neq 5,$$` -- or `$$H_0:\mu_1 - \mu_2 = 5\;\;\text{versus}\;\;H_1:\mu_1 - \mu_2 < 5,$$` -- or `$$H_0:\mu_1 - \mu_2 = 5\;\;\text{versus}\;\;H_1:\mu_1 - \mu_2 > 5.$$` --- name: menti class: middle background-image: url(data:image/png;base64,#menti.jpg) background-size: 115% # Kahoot ## Go to [www.kahoot.it](https://www.kahoot.it) and use ## the code provided --- # The paired `\(t\)`-test * The paired `\(t\)`-test (or *dependent samples `\(t\)`-test*) can be used when we have two samples, with the same set of subjects in each sample -- * For each subject, we have two measurements of the same characteristic: often at different time-points or under different conditions -- * For example, do you think coffee breaks affect productivity? -- * We will consider a data set called `Coffee` from the R package `BSDA` (Arnholt and Evans, 2023; Kitchens, 2003) -- * In this example, workers' productivity scores were measured with and without coffee breaks --- # The paired `\(t\)`-test * We will test whether the mean difference in productivity with and without a coffee break is statistically significant by carrying out a paired `\(t\)`-test with the following hypotheses: `$$H_0:\mu_D = 0\;\;\text{versus}\;\;H_1:\mu_D \neq 0,$$` where: * `\(\mu_D\)` is defined as the true (population) mean difference in productivity with and without a coffee break. --- # Paired `\(t\)`-test What type of variables are required for the paired `\(t\)`-test? -- .content-box-blue[ .center[ A paired *t*-test will always involve two variables: ] 1. The ***dependent*** variable, sometimes also called the *response* variable. This should be a numeric, continuous variable. 2. The ***independent*** variable. This should be a categorical variable with only ***two categories*** which represent before/after categories, or two different conditions. ] -- * So our ***dependent*** variable is productivity * Our ***independent*** variable is coffee break (with or without) --- # Assumptions: Paired samples `\(t\)`-test The assumptions of the paired `\(t\)`-test are exactly the same as those of the one-sample `\(t\)`-test: .content-box-blue[ .center[ **Paired samples *t*-test Assumptions:** ] 1. The data are numeric 2. Observations are independent of one another (that is, the sample is a simple random sample and each individual within the population has an equal chance of being selected) 3. The sample mean, `\(\overline{X}\)`, is normally distributed. ] --- # Assumptions: Paired samples `\(t\)`-test When checking assumptions for the paired `\(t\)`-test, the variable of interest is the ***paired differences*** rather than the 'before' and 'after' variables themselves. -- For example, consider the following snapshot of the `Coffee` data set: ``` without with differences 1 23 28 5 2 35 38 3 3 29 29 0 4 33 37 4 5 43 42 -1 6 32 30 -2 ``` -- The paired `\(t\)`-test is testing whether the average of the ***paired differences*** is equal to zero. For this reason, it is this variable that is of interest when checking assumptions. --- # Visualising the data and checking assumptions We will again begin by considering descriptive statistics and plots: <br> | | | | | |:------------------|:---------------|:--------|:------| |**Variable** |**Sample size** |**Mean** |**SD** | |Without |9 |34.89 |6.39 | |With |9 |35.89 |5.58 | |Paired differences |9 |1 |2.55 | --- <img src="data:image/png;base64,#Topic_6_Lecture_files/figure-html/unnamed-chunk-18-1.svg" width="80%" style="display: block; margin: auto;" /> --- # Visualising data & checking assumptions From the descriptive statistics and plots, we can observe the following: 1. The boxplots and sample means indicate that the average productivity **with** a coffee break is higher. When we carry out paired the `\(t\)`-test, we will see whether or not this difference is ***statistically significant*** -- 2. The mean paired difference is 1, with a standard deviation of 2.55. The paired `\(t\)`-test will be testing to see whether this is significantly different from 0. -- 3. The sample size of the paired differences is 9. This will be useful knowledge when checking for normality. --- # Checking for normality * We will again consider histograms, Normal Q-Q plots, and the Shapiro-Wilk test to check for normality -- * As mentioned earlier, we will check for normality of the *differences* to do so --- # Checking for normality <img src="data:image/png;base64,#Topic_6_Lecture_files/figure-html/unnamed-chunk-19-1.svg" style="display: block; margin: auto;" /> Based on the histogram and Normal Q-Q plots, do you have any concerns regarding normality for either group? --- # Checking for normality There is potentially some skewness present. However, given the small sample size, it is difficult to tell whether or not these data are sampled from a normal distribution. -- Let's see what the Shapiro-Wilk test shows: **Shapiro-Wilk test for paired differences:** ``` Shapiro-Wilk normality test data: Coffee$differences W = 0.89448, p-value = 0.2217 ``` -- Given we have `\(p = 0.2217\)` which is greater than 0.05, we do not have evidence against normality. Therefore, for the purposes of this example, **we will assume that the normality assumption has not been violated.** We are now ready to carry out the paired `\(t\)`-test. --- # Coffee Break Example: Paired `\(t\)`-test output ``` r Paired t-test data: Coffee$with and Coffee$without t = 1.1767, df = 8, p-value = 0.2731 alternative hypothesis: true difference in means is not equal to 0 95 percent confidence interval: -0.9597267 2.9597267 sample estimates: mean of the differences 1 ``` --- # Coffee Break Example: Paired `\(t\)`-test output ``` r Paired t-test data: Coffee$with and Coffee$without `t = 1.1767`, `df = 8`, p-value = 0.2731 alternative hypothesis: true difference in means is not equal to 0 95 percent confidence interval: -0.9597267 2.9597267 `sample estimates:` `mean of the differences` `1` ``` * The **test statistic** is `\(t = 1.1767\)` * The **degrees of freedom** is `\(8\)` * The **sample mean** of the paired differences is `\(1\)`. --- # Coffee Break Example: Paired `\(t\)`-test output ``` r Paired t-test data: Coffee$with and Coffee$without t = 1.1767, df = 8, `p-value = 0.2731` alternative hypothesis: true difference in means is not equal to 0 95 percent confidence interval: -0.9597267 2.9597267 sample estimates: mean of the differences 1 ``` * The ** `\(p\)`-value** is 0.2731, which is greater than 0.05, so we cannot reject `\(H_0\)`. That is, we do not have enough evidence to conclude that the difference in productivity is different depending on coffee break --- # Coffee Break Example: Paired `\(t\)`-test output ``` r Paired t-test data: Coffee$with and Coffee$without t = 1.1767, df = 8, p-value = 0.2731 alternative hypothesis: true difference in means is not equal to 0 `95 percent confidence interval:` `-0.9597267 2.9597267` sample estimates: mean of the differences 1 ``` * The **95% confidence interval** for `\(\mu_D\)` is (-0.96, 2.96), meaning we are 95% confident that the population mean difference in productivity with a coffee break compared to without is between 0.96 and -2.96. Since this interval includes 0, we cannot reject `\(H_0\)`. --- # Link between paired and one-sample `\(t\)`-tests * It turns out that the paired and one-sample `\(t\)`-tests are actually equivalent! -- * See [this topic's readings](https://amandashaker-stm1001-topic-6.share.connect.posit.cloud/2.4-link-between-paired-and-one-sample-t-tests.html) if you are interested to learn why -- * You will also have a chance to try this for yourself in the computer lab --- #Independent samples vs. Paired `\(t\)`-test While independent samples `\(t\)`-tests and paired `\(t\)`-tests both involve using data from two groups, we have to be careful to only apply them in appropriate contexts. -- * One key concept to remember for an **independent** samples `\(t\)`-test is that we must have ***two independent groups*** -- * This would normally mean that it is not possible for the same person (or subject/observation if the study is not about people) to be in both groups (e.g. brown eyes vs. not brown eyes) -- * On the other hand, a **paired** `\(t\)`-test tests for a difference when we have the **same set of subjects in each sample** -- * This would mean we have ***two dependent groups*** -- * One way of thinking of it would be that all individuals must be in both groups --- # Effect sizes When we conduct our tests, we want to verify that our results are not only statistically significant but also clinically meaningful (i.e. that there is a meaningful 'real world' difference). -- * An ***effect size*** tells us the ***relative size*** of any difference observed -- * The ***effect size*** can be thought of as the ***standardised difference in mean*** -- * For `\(t\)`-tests, ***Cohen's `\(d\)`*** (Cohen, 1988) is typically used and can be calculated as $$ d = \displaystyle \frac{\bar{x} - \mu_0}{s},$$ where * `\(\bar{x}\)` denotes the sample mean * `\(\mu_0\)` denotes the mean under the null hypothesis * `\(s\)` denotes the sample standard deviation. --- # Effect sizes Cohen (1988) provided a guide to quantify the magnitude of an effect size: .content-box-blue[ .center[ **Guidelines for quantifying magnitude of Cohen's *d* effect sizes for *t*-tests:** ] * `\(|d| < 0.2\)`: "negligible" * `\(0.2 \leq |d| < 0.5\)`: "small" * `\(0.5 \leq |d| < 0.8\)`: "medium" * `\(|d| \geq 0.8\)`: "large" ] -- Thankfully, most computer packages will do the hard work for us. We will consider some examples in the computer lab. --- # References Arnholt, A. T. and B. Evans (2023). _BSDA: Basic Statistics and Data Analysis_. R package version 1.2.2. URL: [https://CRAN.R-project.org/package=BSDA](https://CRAN.R-project.org/package=BSDA). Cohen, J. (1988). _Statistical power analysis for the behavioral sciences_. 2nd edition. New York: Academic Press. Kitchens, L. J. (2003). _Basic Statistics and Data Analysis_. Pacific Grove, CA,: Brooks/Cole, a division of Thomson Learning. --- background-image: url(data:image/png;base64,#computerlab.jpg) background-position: bottom background-size: 75% class: center # See you in the computer labs! --- class: middle <font color = "grey"> These notes have been prepared by Amanda Shaker and Rupert Kuveke. The copyright for the material in these notes resides with the authors named above, with the Department of Mathematics and Statistics and with La Trobe University. Copyright in this work is vested in La Trobe University including all La Trobe University branding and naming. Unless otherwise stated, material within this work is licensed under a Creative Commons Attribution-Non Commercial-Non Derivatives License <a href = "https://creativecommons.org/licenses/by-nc-nd/4.0/" target="_blank"> BY-NC-ND. </a> </font>