2023-09-26
Next lecture:
- Frequency distributions
- Probability using R frequency distribution functions
Expectations for today:
-Understand how probability applies to the study of political (or any) science.
- Understand basic assumptions/rules about probability.
- Be able to do relatively simple probability computations by hand. Code will be provided to them in R
Hypothesis testing: checking to see if our theory about how the world works is false or more likely than not to be true
correlation: the relationship between the X and Y variables
Probability and Frequency Distributions set the foundation for hypothesis testing
Our ultimate goal is to know whether the correlation represented by the equation and line is meaningful
- We define meaningful as being *significant* to a certain *probability*
Probability and Frequency Distributions set the foundation for hypothesis testing
Our ultimate goal is to know whether the correlation represented by the equation and line is meaningful
- We define meaningful as being *significant* to a certain *probability*
- We need *probability theory* to get there
Probability and Frequency Distributions set the foundation for hypothesis testing
Our ultimate goal is to know whether the correlation represented by the equation and line is meaningful
- We define meaningful as being *significant* to a certain *probability*
- We need *probability theory* to get there
- Why? There is an *element of randomness* in our data
Probability and Frequency Distributions set the foundation for hypothesis testing
Our ultimate goal is to know whether the correlation represented by the equation and line is meaningful
- We define meaningful as being *significant* to a certain *probability*
- We need *probability theory* to get there
- Why? There is an *element of randomness* in our data
- Probability *models* randomness or chance
Probability and Frequency Distributions set the foundation for hypothesis testing
Our ultimate goal is to know whether the correlation represented by the equation and line is meaningful
- We define meaningful as being *significant* to a certain *probability*
- We need *probability theory* to get there
- Why? There is an *element of randomness* in our data
- Probability *models* randomness or chance
Statistical Significance is essentially the probability that our answer is due to randomness and not truly meaningful
Mathematical probability: the ratio of the cases of interest to the whole number of cases possible.
P(A) = \(\frac{possible_A}{all \; possible \; outcomes}\)
This is often phrased as ratio of “favorable outcomes” to possible outcomes, where favorable means outcomes where A occurs.
Unfortunate phrasing. Does not mean a “good” outcome.For example, in the study of war it might mean the outcomes where war occurs.
If we call heads “favorable”
If we call heads “favorable”
If we call heads “favorable”
If we call heads “favorable”
The ratio of
heads on a fair coin: 1
to the whole number of cases possible:
heads on a fair coin + tails on a fair coin = 2
\(\frac{1}{2}\)
Other examples work the same way just with more possible cases and different ratios
Other examples work the same way just with more possible cases and different ratios
Other examples work the same way just with more possible cases and different ratios
Other examples work the same way just with more possible cases and different ratios
Other examples work the same way just with more possible cases and different ratios
We do something similar when trying to infer something about the way the world works from statistical data
Probability in statistical inference: Proportional times an outcome occurs if the experiment (or observation/data collection) is repeated a large number of times. (This should go on your exam notes!)
This is the frequentist interpretation. There is an important alternative definition you should be aware of, but that you are not responsible for in this class, the Bayesian interpretation
Bayesian probability suggests that in statistical inference, probability is a method of updating our beliefs based on new evidence. This is based on Bayes Rule, which has been called “the most important rule in data science.”
bayes rule
P(A|B) - This is read: The probability of A given B.
P(A) is our prior (existing) probability of A given no information about B.
In other words, if we know that B is true (given B) what is the probability that A happens.
https://www.freecodecamp.org/news/bayes-rule-explained/
Simplified explanation involving error rates in cancer tests: https://betterexplained.com/articles/an-intuitive-and-short-explanation-of-bayes-theorem/
Good explanation using set notation: https://www.investopedia.com/terms/p/prior_probability.asp
A & B are opposites and complements.
S = {1,2,3,4,5,6,7,8,9,10}
A = {1,2,3,4,5}
B = {6,7,8,9,10}
P (A) = 5 / 10
P(B) = 5 / 10
P(A \(\cup\) B) =
S = {1,2,3,4,5,6,7,8,9,10}
A = {1,2,3,4,5,6}
B = {5,6,7,8,9,10}
P (A) = 6 / 10
P(B) = 6 / 10
P(A \(\cup\) B) =
B \(\cap\) A = {5,6}
P(B \(\cap\) A) = 2 / 10
Author: Tom Hanna
Website: tomhanna.me
License: This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.</>
GOVT2306, Fall 2023, Instructor: Tom Hanna