Simple linear regression is the subset of general linear regression dealing with a single dependent variable and a single explanatory variable. As its name suggests, it is in many ways the simplest and most straightforward implementation of linear regression.
Mathematically, for explanatory variable x and dependent variable y
\(y = \beta_0 + \beta_1x + \epsilon\)
\(\text{where} \hspace{.1cm} \beta_0 \hspace{.1cm} \text{is a constant, }\) \(\beta_1 \hspace{.1cm} \text{is the regression coefficient, and}\) \(\epsilon \hspace{.1cm} \text{is the error term}\)
Generally with simple linear regression, ordinary least squares (OLS) is used to measure the relative accuracy of resulting regression line. This is done by squaring the vertical distance between the predicted regression value and a given point of the data set, with the goal that the smallest summation of these values is the most accurate regression estimation line.