The Scientific Method: A Process for Learning
In the Middle Ages, science was deduced from principles set down many
centuries earlier by authorities such as Aristotle. The idea that
scientific theories should be tested against real world data
revolutionized thinking. This way of thinking known as the scientific
method sparked the Renaissance.
The scientific method rests on the following premises:
A scientific hypothesis can never be shown to be absolutely
true.
However, it must potentially be disprovable.
It is a useful model until it is established that it is not
true.
Always go for the simplest hypothesis, unless it can be shown to
be false.
This last principle, elaborated by William of Ockham in the 13th
century, is now known as Ockham’s razor and is firmly
embedded in science. It keeps science from developing fanciful overly
elaborate theories. Thus, the scientific method directs us through an
improving sequence of models, as previous ones get falsified. The
scientific method generally follows the following procedure:
Ask a question or pose a problem in terms of the current
scientific hypothesis.
Gather all the relevant information that is currently available.
This includes the current knowledge about parameters of the
model.
Design an investigation or experiment that addresses the question
from Step 1. The predicted outcome of the experiment should be
one thing if the current hypothesis is true, and something else if the
hypothesis is false.
Gather data from the experiment.
Draw conclusions given the experimental results. Revise the
knowledge about the parameters to take the current results into
account.
The scientific method searches for cause-and-effect
relationships between an experimental variable and an outcome variable.
In other words, how changing the experimental variable results in a
change to the outcome variable. Scientific modeling develops
mathematical models of these relationships. Both of them need to isolate
the experiment from outside factors that could affect the experimental
results. All outside factors that can be identified as possibly
affecting the results must be controlled. It is no coincidence that the
earliest successes for the method were in physics and chemistry where
the few outside factors could be identified and controlled. Thus, there
were no lurking variables. All other relevant variables could
be identified and could then be physically controlled by being held
constant. That way they would not affect results of the experiment, and
the effect of the experimental variable on the outcome variable could be
determined. In biology, medicine, engineering, technology, and the
social sciences it is not that easy to identify the relevant factors
that must be controlled. In those fields a different way to control
outside factors is needed, because they cannot be identified beforehand
and physically controlled.
The Role of Statistics in the Scientific Method
Statistical methods of inference can be used when there is random
variability in the data. The probability model for the data is justified
by the design of the investigation or experiment. This can extend the
scientific method into situations where the relevant outside factors
cannot even be identified. Since we cannot identify these outside
factors, we cannot control them directly. The lack of direct control
means the outside factors will be affecting the data. There is a danger
that the wrong conclusions could be drawn from the experiment due to
these uncontrolled outside factors.
The important statistical idea of randomization has
been developed to deal with this possibility. The unidentified outside
factors can be averaged out by randomly assigning each unit to
either treatment or control group. This contributes variability to the
data. Statistical conclusions always have some uncertainty or error due
to variability in the data. We can develop a probability model of the
data variability based on the randomization used. Randomization not only
reduces this uncertainty due to outside factors, it also allows us to
measure the amount of uncertainty that remains using the probability
model. Randomization lets us control the outside factors statistically,
by averaging out their effects.
Underlying this is the idea of a statistical population,
consisting of all possible values of the observations that could be
made. The data consists of observations taken from a sample of the
population. For valid inferences about the population parameters from
the sample statistics, the sample must be representative of the
population. Amazingly, choosing the sample randomly is the most
effective way to get representative samples!
Main Approaches to Statistics
There are two main philosophical approaches to statistics. The first
is often referred to as the frequentist approach.
Sometimes it is called the classical approach.
Procedures are developed by looking at how they perform over all
possible random samples. The probabilities do not relate to the
particular random sample that was obtained. In many ways this indirect
method places the cart before the horse.
The alternative approach that we take in this course is the
Bayesian approach. It applies the laws of probability
directly to the problem. This offers many fundamental advantages over
the more commonly used frequentist approach.
Frequentist Approach to Statistics
The frequentist approach to statistics is based on the following
ideas:
Parameters, the numerical characteristics of the population, are
fixed but unknown constants.
Probabilities are always interpreted as long-run relative
frequency.
Statistical procedures are judged by how well they perform in the
long run over an infinite number of hypothetical repetitions of the
experiment.
Probability statements are only allowed for random quantities. The
unknown parameters are fixed, not random, so probability statements
cannot be made about their value. Instead, a sample is drawn from the
population, and a sample statistic is calculated. The probability
distribution of the statistic over all possible random samples from the
population is determined and is known as the sampling distribution of
the statistic. A parameter of the population will also be a parameter of
the sampling distribution. The probability statement that can be made
about the statistic based on its sampling distribution is converted to a
confidence statement about the parameter. The confidence is based on the
average behavior of the procedure over all possible samples.
Bayesian Approach to Statistics
Bayesian approach to statistics put forward the ideas that:
Since we are uncertain about the true value of the parameters, we
will consider them to be random variables.
The rules of probability are used directly to make inferences
about the parameters.
Probability statements about parameters must be interpreted as
degree of belief. The prior distribution must be
subjective. Each person can have his/her own prior, which contains the
relative weights that person gives to every possible parameter value. It
measures how plausible the person considers each parameter
value to be before observing the data.
We revise our beliefs about parameters after getting the data by
using Bayes’ theorem. This gives our posterior distribution which gives
the relative weights we give to each parameter value after analyzing the
data. The posterior distribution comes from two sources: the prior
distribution and the observed data.
This has a number of advantages over the conventional frequentist
approach. Bayes’ theorem is the only consistent way to modify our
beliefs about the parameters given the data that actually occurred. This
means that the inference is based on the actual occurring data, not all
possible data sets that might have occurred but did not! Allowing the
parameter to be a random variable allows us make probability statements
about it, posterior to the data.
This contrasts with the conventional approach where inference
probabilities are based on all possible data sets that could have
occurred for the fixed parameter value. Given the actual data, there is
nothing random left with a fixed parameter value, so one can only make
confidence statements, based on what could have occurred.
Bayesian statistics also has a general way of dealing with a nuisance
parameter. A nuisance parameter is one which we do not want to make
inference about, but we do not want them to interfere with the
inferences we are making about the main parameters. Frequentist
statistics does not have a general procedure for dealing with them.
Bayesian statistics is predictive, unlike conventional frequentist
statistics. This means that we can easily find the conditional
probability distribution of the next observation given the sample
data.