Bayesian Inference Pt. 1: Bayes Rule
Up until now, most of the statistical work we have looked at would fall under the category of frequentist statistics: we’ve made inferential conclusions about populations based solely on data.
But we don’t necessarily need to rely on just a sample of data to make a conclusion; we can include our prior knowledge and beliefs on the subject using what is called Bayesian statistics. The next couple of posts will focus on Bayesian methodology, allowing us to see how we can include our own prior beliefs in models we create.
Quick Probability Review
Most people deal with probability in their day-to-day lives, whether you’re a statistician or not. Thinking about things in a probabilistic manner is just human nature. When we think about the Warriors winning this year’s championship, for example, we’re more likely to gauge it as a probability than a yes or no outcome (although it’s probably close to a 100% probability).
We don’t always think of the probability of one event, though… we sometimes combine events together. For example, what is the probability of the Warriors winning the championship this year AND next year (again, probably close to 100%). Here we start to get into joint and conditional probability.
Conditional probability, represented as , is the probability that event X will happen, given event happens. For example, if the Warriors win the championship, what is the probability that they sign another superstar?
Joint probability, represented as , is the probability that event X and event happen. So this would be the probability that the Warriors win a championship and sign a superstar.
The joint probability of two independent events (say someone scoring 20 points in one game and another player scoring 15 points in a separate game) is found by multiplying those events’ probabilities:
If one of the events is dependent on the other, we multiple the probability of the first event and the conditional probability of the second given event one has happened:
We can restructure this to find the formula for the conditional probability:
Bayes Rule
Bayesian statistics has become extremely popular as computer power has improved, but its origins actually go way back to the 18th century and Reverend Thomas Bayes.
Bayes rule is another form of the conditional probability formula. Remember that the conditional probability of an event can be found with:
We could replace the numerator in the conditional probability formula with to get:
What the following equation states is that the conditional probability of given is equal to the product of the conditional probability of given and the probability of , divided by the probability of .
Let’s put it into an example in basketball terms. Let’s say we wanted to know the probability of a college player being drafted given that they are at least 7 feet tall. We can write out what each term means in the equation:
- : Probability a player is drafted
- : Probability a player is at least 7 ft.
- : Probability a player is at least 7 ft. given that they are drafted
- : Probability a player is drafted given that they are at least 7 ft.
Note that the numbers I use for this are not based on any data; we’ll make them up along the way just to further the example.
Let’s address first; even if you are a college player, the chances of being drafted are extremely low. There are only 60 spots each year, and a handful of those will go to guys who didn’t even play in college (Euroleague, G-League). Let’s say the probability of being drafted is 1% (and even that is being generous). This is our prior belief; it’s what we assume going into the problem.
Now let’s think of . This is the probability that a college player is at least 7 ft. This is pretty rare, but not as rare as being drafted. We’ll say that the probability is 8%.
Now what about ? There aren’t a ton of 7 footers in the draft pool because of a bunch of reasons (shift to position-less basketball, small amount of guys talented enough, injury issues, etc.). Despite this, there’s still a handful that get drafted… let’s say about 12% of drafted players are at least 7 feet. This is the data we are using to update our prior belief.
With all of these numbers, we can now plug them into the equation:
Therefore, our probability of a player getting drafted given that they are at least 7 ft. is 1.5%. That may not seem like a huge jump, but remember how hard it is to get drafted in the first place! We’ve taken our prior belief of how likely it is for a college player to be drafted and updated it with new information. Now we know that players that are at least 7 ft. are more likely to be drafted.
In the next lesson, we’ll move this into an inference environment, and see how we can expand on the basic Bayesian formula.
If interested, I like the following explanations provided on the blogs linked here and here.