## Installing package into '/cloud/lib/x86_64-pc-linux-gnu-library/4.3'
## (as 'lib' is unspecified)
## Installing package into '/cloud/lib/x86_64-pc-linux-gnu-library/4.3'
## (as 'lib' is unspecified)
Basketball players who make several baskets in succession are described as having a hot hand. Fans and players have long believed in the hot hand phenomenon, which refutes the assumption that each shot is independent of the next. However, a 1985 paper by Gilovich, Vallone, and Tversky collected evidence that contradicted this belief and showed that successive shots are independent events. This paper started a great controversy that continues to this day, as you can see by Googling hot hand basketball.
We do not expect to resolve this controversy today. However, in this lab we’ll apply one approach to answering questions like this. The goals for this lab are to (1) think about the effects of independent and dependent events, (2) learn how to simulate shooting streaks in R, and (3) to compare a simulation to actual data in order to determine if the hot hand phenomenon appears to be real.
In this lab, we will explore and visualize the data using the
tidyverse suite of packages. The data can be found in the
companion package for OpenIntro labs, openintro.
Let’s load the packages.
Your investigation will focus on the performance of one player: Kobe Bryant of the
Los Angeles Lakers. His performance against the Orlando Magic in the 2009 NBA Finals
earned him the title Most Valuable Player and many spectators
commented on how he appeared to show a hot hand. The data file we’ll use
is called kobe_basket.
## Rows: 133
## Columns: 6
## $ vs <fct> ORL, ORL, ORL, ORL, ORL, ORL, ORL, ORL, ORL, ORL, ORL, ORL…
## $ game <int> 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1…
## $ quarter <fct> 1, 1, 1, 1, 1, 1, 1, 1, 1, 2, 2, 2, 2, 2, 2, 2, 2, 2, 3, 3…
## $ time <fct> 9:47, 9:07, 8:11, 7:41, 7:03, 6:01, 4:07, 0:52, 0:00, 6:35…
## $ description <fct> Kobe Bryant makes 4-foot two point shot, Kobe Bryant misse…
## $ shot <chr> "H", "M", "M", "H", "H", "M", "M", "M", "M", "H", "H", "H"…
This data frame contains 133 observations and 6 variables, where
every row records a shot taken by Kobe Bryant. The shot
variable in this dataset indicates whether the shot was a hit
(H) or a miss (M).
Just looking at the string of hits and misses, it can be difficult to gauge whether or not it seems like Kobe was shooting with a hot hand. One way we can approach this is by considering the belief that hot hand shooters tend to go on shooting streaks. For this lab, we define the length of a shooting streak to be the number of consecutive baskets made until a miss occurs.
For example, in Game 1 Kobe had the following sequence of hits and misses from his nine shot attempts in the first quarter:
\[ \textrm{H M | M | H H M | M | M | M} \]
You can verify this by viewing the first 9 rows of the data in the data viewer.
Within the nine shot attempts, there are six streaks, which are separated by a “|” above. Their lengths are one, zero, two, zero, zero, zero (in order of occurrence).
In this specific context, the streak is only iterated when we have makes. In this case, a ‘streak of 1’ means that Kobe made a shot. A ‘streak of 0’ would be an observation of Kobe missing the shot.
Counting streak lengths manually for all 133 shots would get tedious,
so we’ll use the custom function calc_streak to calculate
them, and store the results in a data frame called
kobe_streak as the length variable.
We can then take a look at the distribution of these streak lengths.
The typical streak for Kobe during this data sample was a 0, aka a miss. Streaks of 2 and 3 are surprisingly, relatively equally common. The ratio between 0:1 and 2:3 seem to be roughly in line. The longest streak observed within the data is a streak of 4 consecutive makes.
We’ve shown that Kobe had some long shooting streaks, but are they long enough to support the belief that he had a hot hand? What can we compare them to?
To answer these questions, let’s return to the idea of independence. Two processes are independent if the outcome of one process doesn’t effect the outcome of the second. If each shot that a player takes is an independent process, having made or missed your first shot will not affect the probability that you will make or miss your second shot.
A shooter with a hot hand will have shots that are not independent of one another. Specifically, if the shooter makes his first shot, the hot hand model says he will have a higher probability of making his second shot.
Let’s suppose for a moment that the hot hand model is valid for Kobe. During his career, the percentage of time Kobe makes a basket (i.e. his shooting percentage) is about 45%, or in probability notation,
\[ P(\textrm{shot 1 = H}) = 0.45 \]
If he makes the first shot and has a hot hand (not independent shots), then the probability that he makes his second shot would go up to, let’s say, 60%,
\[ P(\textrm{shot 2 = H} \, | \, \textrm{shot 1 = H}) = 0.60 \]
As a result of these increased probabilites, you’d expect Kobe to have longer streaks. Compare this to the skeptical perspective where Kobe does not have a hot hand, where each shot is independent of the next. If he hit his first shot, the probability that he makes the second is still 0.45.
\[ P(\textrm{shot 2 = H} \, | \, \textrm{shot 1 = H}) = 0.45 \]
In other words, making the first shot did nothing to effect the probability that he’d make his second shot. If Kobe’s shots are independent, then he’d have the same probability of hitting every shot regardless of his past shots: 45%.
Now that we’ve phrased the situation in terms of independent shots, let’s return to the question: how do we tell if Kobe’s shooting streaks are long enough to indicate that he has a hot hand? We can compare his streak lengths to someone without a hot hand: an independent shooter.
While we don’t have any data from a shooter we know to have independent shots, that sort of data is very easy to simulate in R. In a simulation, you set the ground rules of a random process and then the computer uses random numbers to generate an outcome that adheres to those rules. As a simple example, you can simulate flipping a fair coin with the following.
## [1] "tails"
The vector coin_outcomes can be thought of as a hat with
two slips of paper in it: one slip says heads and the other
says tails. The function sample draws one slip
from the hat and tells us if it was a head or a tail.
Run the second command listed above several times. Just like when flipping a coin, sometimes you’ll get a heads, sometimes you’ll get a tails, but in the long run, you’d expect to get roughly equal numbers of each.
If you wanted to simulate flipping a fair coin 100 times, you could
either run the function 100 times or, more simply, adjust the
size argument, which governs how many samples to draw (the
replace = TRUE argument indicates we put the slip of paper
back in the hat before drawing again). Save the resulting vector of
heads and tails in a new object called sim_fair_coin.
To view the results of this simulation, type the name of the object
and then use table to count up the number of heads and
tails.
## [1] "tails" "tails" "heads" "tails" "tails" "tails" "heads" "tails" "heads"
## [10] "tails" "heads" "heads" "tails" "tails" "heads" "tails" "tails" "tails"
## [19] "heads" "tails" "tails" "heads" "tails" "heads" "tails" "tails" "tails"
## [28] "heads" "heads" "tails" "heads" "tails" "tails" "heads" "heads" "heads"
## [37] "heads" "heads" "heads" "tails" "heads" "heads" "tails" "heads" "tails"
## [46] "heads" "heads" "heads" "heads" "heads" "tails" "heads" "heads" "tails"
## [55] "tails" "tails" "heads" "heads" "heads" "heads" "heads" "tails" "tails"
## [64] "heads" "heads" "heads" "tails" "heads" "heads" "heads" "tails" "tails"
## [73] "heads" "heads" "heads" "tails" "tails" "tails" "heads" "heads" "heads"
## [82] "tails" "heads" "tails" "heads" "heads" "heads" "tails" "heads" "tails"
## [91] "tails" "heads" "tails" "tails" "tails" "tails" "heads" "tails" "tails"
## [100] "tails"
## sim_fair_coin
## heads tails
## 52 48
Since there are only two elements in coin_outcomes, the
probability that we “flip” a coin and it lands heads is 0.5. Say we’re
trying to simulate an unfair coin that we know only lands heads 20% of
the time. We can adjust for this by adding an argument called
prob, which provides a vector of two probability
weights.
prob=c(0.2, 0.8) indicates that for the two elements in
the outcomes vector, we want to select the first one,
heads, with probability 0.2 and the second one,
tails with probability 0.8. Another way of thinking about
this is to think of the outcome space as a bag of 10 chips, where 2
chips are labeled “head” and 8 chips “tail”. Therefore at each draw, the
probability of drawing a chip that says “head”” is 20%, and “tail” is
80%.
## [1] "tails" "tails" "tails" "tails" "tails" "tails" "tails" "heads" "heads"
## [10] "tails" "tails" "heads" "tails" "tails" "tails" "tails" "tails" "tails"
## [19] "heads" "tails" "heads" "heads" "tails" "heads" "tails" "tails" "tails"
## [28] "heads" "tails" "heads" "tails" "tails" "tails" "tails" "tails" "tails"
## [37] "heads" "tails" "heads" "heads" "tails" "tails" "tails" "tails" "heads"
## [46] "tails" "tails" "heads" "tails" "tails" "tails" "tails" "tails" "tails"
## [55] "tails" "tails" "tails" "tails" "tails" "tails" "tails" "heads" "tails"
## [64] "tails" "heads" "tails" "tails" "tails" "tails" "tails" "tails" "tails"
## [73] "tails" "tails" "tails" "tails" "tails" "tails" "tails" "tails" "heads"
## [82] "tails" "tails" "tails" "tails" "tails" "tails" "tails" "tails" "tails"
## [91] "tails" "heads" "tails" "tails" "heads" "tails" "tails" "tails" "tails"
## [100] "tails"
## sim_unfair_coin
## heads tails
## 19 81
I set the random seed ‘092297’ and then I ran the unfair coin simulation and I ended up with a count of 19 heads and 81 tails.
A note on setting a seed: Setting a seed will cause R to select the same sample each time you knit your document. This will make sure your results don’t change each time you knit, and it will also ensure reproducibility of your work (by setting the same seed it will be possible to reproduce your results). You can set a seed like this:
## i set the seed above because when I knit + run it does them in order, and we want it for be done beforehandThe number above is completely arbitrary. If you need inspiration, you can use your ID, birthday, or just a random string of numbers. The important thing is that you use each seed only once in a document. Remember to do this before you sample in the exercise above.
In a sense, we’ve shrunken the size of the slip of paper that says
“heads”, making it less likely to be drawn, and we’ve increased the size
of the slip of paper saying “tails”, making it more likely to be drawn.
When you simulated the fair coin, both slips of paper were the same
size. This happens by default if you don’t provide a prob
argument; all elements in the outcomes vector have an equal
probability of being drawn.
If you want to learn more about sample or any other
function, recall that you can always check out its help file.
Simulating a basketball player who has independent shots uses the same mechanism that you used to simulate a coin flip. To simulate a single shot from an independent shooter with a shooting percentage of 50% you can type
To make a valid comparison between Kobe and your simulated independent shooter, you need to align both their shooting percentage and the number of attempted shots.
sample function so
that it reflects a shooting percentage of 45%? Make this adjustment,
then run a simulation to sample 133 shots. Assign the output of this
simulation to a new object called sim_basket.independent_outcomes <- c("H", "M")
sim_basket <- sample(independent_outcomes, size = 133, replace = TRUE, prob = c(0.45, 0.55))
table(sim_basket)## sim_basket
## H M
## 59 74
The above code reflects a simulation of a shooter with a FG% of 45% attempting 133 shots. This particular simulation has 60 makes and 73 misses, for a percentage of 45.1%
Note that we’ve named the new vector sim_basket, the
same name that we gave to the previous vector reflecting a shooting
percentage of 50%. In this situation, R overwrites the old object with
the new one, so always make sure that you don’t need the information in
an old vector before reassigning its name.
With the results of the simulation saved as sim_basket,
you have the data necessary to compare Kobe to our independent
shooter.
Both data sets represent the results of 133 shot attempts, each with the same shooting percentage of 45%. We know that our simulated data is from a shooter that has independent shots. That is, we know the simulated shooter does not have a hot hand.
calc_streak, compute the streak lengths of
sim_basket, and save the results in a data frame called
sim_streak.Insert your answer here
The result ends up being more or less the same as what we saw for Kobe. One notable difference is that in this simulation our independent shooter made to a streak of 5, which is one more than what Kobe did in his sample of 133 shots. In general, the same trends hold where the shots are slightly more likely to miss than any other result. One way to interpret the parity between length 1 and length 2 which is observed in this case is that made shots were equally likely to come ‘alone’ (sandwiched by misses), as they are to be a part of a ‘proper streak’ (I would describe a streak of one as a ‘degenerate streak’ and any streak of length greater than 2 as a proper streak.)
By definition, you would expect the independent shooter to produce the same distribution of streak lengths every time.
I would say that the model of an independent shooter is quite comparable to what we see with Kobe. Even though this trial was severely limited in its scope, we cannot see that there seems to be any evidence to support the ‘hot hand’ theory based on what we found in this sample, although I think if we asked NBA players we would get a different answer!
I also wanted to share this video I came across a few weeks ago about an algorithm called Hyperloglog.
https://www.youtube.com/watch?v=lJYufx0bfpw)
Hyperloglog is a probabilistic counting algorithm that very popular for counting things at scale. The basic idea is that, someone like Kobe or the independent shooter would be likely to go on longer streaks if we open up the range that we are looking at. The longest streak is going to be directly proportional to the number of shots in the range. In this case, HLL looks at the 0s and 1s in binary code and looks at the longest streak of 0s to begin a number in the dataset. This ‘streak’ allows to companies to very accurately estimate large counts that would otherwise require a lot more computational infrastructure to maintain.