We use linear regression to analyse linear patterns between a continuous response variable and one or more continuous predictor variable(s). In contrast to correlation analysis, which focuses on the strength of a linear relationship, linear regression assumes a causal relationship between the predictor(s) and the response variable.
The model predictions can be plotted as a regression line which is determined by the intercept (y-value at x = 0) and the slope (change in y per unit increase in x). These parameters are often referred to as \(\beta_{0}\) and \(\beta_{1}\).\[ y_i = \beta_{0} + \beta_{1}x_{i} + \epsilon_{i} \quad \quad \quad \epsilon \sim N(0,\sigma^2) \]
\(x = predictor~variable~(continuous~explanatory~variable)\)
\(\beta_{0}~=~intercept\)
\(\beta_{1}~=~slope\)
\(\epsilon = model~errors~estimated~by~the~residuals\)
\(\epsilon \sim N(0,\sigma^2)~reads:the~residuals~are~assumed~to~approx.~follow~a~normal~distribution~with~a~\mathbf{mean~of~zero}~and\)
\(a~\mathbf{constant~variance}~of~\sigma^2\)