A baseball season is the amalgamation of a lot of little events. Each pitch fits into a plate appearance which fits into an inning which fits into a game which fits into a series which fits into a season. That’s a lot of little data points flowing into an overall end result. We care a lot about which players will have good seasons and careers. It matters to us that we can distinguish between good players and bad players, but doing so requires that we understand which chunks of data are meaningful and which aren’t.

Enter sample size. You’ve heard this phrase plenty over the years when talking about baseball statistics and it’s usually a conversation ender rather than a conversation starter. Someone cites a stat and then another person says it doesn’t matter because the sample size is too small. What does that mean and how should we properly think about sample size in baseball?

If you’re just looking for the numbers, skip ahead by clicking here .

Overview:

Each little moment in baseball is essentially random. Not random in the sense that all outcomes are equally likely and subject completely to chance, but random in the sense that the most likely outcome doesn’t happen every time. If the best hitter in baseball faced the worst pitcher 100 times, he would very likely strike out a couple of times and hit into a double play or two. He wouldn’t always hit a home run even if it was Coors Field and the pitcher was throwing meatballs. Think about the home run derby. MLB players can’t simply hit home runs on demand even when the pitcher is trying to help.

When dealing with pitches flying 90+ miles per hour and split second movements, a whole bunch of randomness gets thrown into the pot. This means that any one plate appearance might have a funky result, meaning that you need to see lots of events to get a clear picture of what is going on. You know this. One time Don Kelly took Yu Darvish deep.

So of course, we know that a single plate appearance isn’t a convincing amount of data. Even the least sabermetrically-minded person agrees with that concept. That single plate appearance is a valid data point, but it’s not enough information to inform your opinion very fully. Instead, you need more and more data points until you have enough for them to “stabilize.” Remember that word because the way we’re going to come back to it in a very specific way in a moment.

Essentially, we want to make sure we have enough observations that the random noise gets cancelled out. Don Kelly hit a home run against Yu Darvish one time, but how many Kelly versus Davrish at bats do we need before we can accurately access their abilities? It’s more than one for sure, but the actual number you need depends on the skill you’re trying to analyze. For example, strikeout rate starts to communicate useful information in fewer than 100 PA while BABIP for a pitcher can take three years. The difference is the nature of the skill and the number of factors that influence the outcome of the play. With respect to strikeout rate, we’re only talking about the batter and pitcher’s ability to make or allow contact (or let strikes go by). When you’re talking about BABIP, you’re adding in quality of contact, direction, weather, defensive ability, luck, etc. That means there’s more room for noise and things with less noise in the actual data generating process stabilize more quickly. There are also diminishing returns. Having 20 PA is better than five PA, but having 520 PA is only a little better than having 505 PA.

Stabilization? Reliability?

So let’s go back to this idea of stabilization. This is the word you hear a lot in conversation about baseball statistics. Conceptually, it’s an ironclad idea. You want to know how many data points it takes for the current information to provide an accurate assessment of the player in question. However, there is no one point at which something stabilizes. Things become stable over time, at a given speed. So after five PA, you know more about a hitter’s walk rate than after one PA, but you don’t know as much as you do after 150 PA or 600 PA. A statistic doesn’t stabilize, it becomes more stable.

The word “stabilize” got into the baseball lexicon after some work by Russell Carleton (aka Pizza Cutter), who looked to see how many PA you need for a given statistic to reach the point where the correlation between that sample and another sample of the same size is 0.7 (i.e. R^2 of .49). That’s the colloquial definition of stabilize and despite Carleton’s constant warnings, most of us picked up the word “stabilize” and ran with it even if it’s not the most useful term. He has done updated work here and here, and has some thoughts about how we talk about stabilization and reliability here.

Regardless, the key is that 100 PA is better than 50 PA no matter the statistic, but smaller samples are more useful for some statistics than others. It’s always better to have more data, but the rate at which the data becomes useful varies based on the statistic. A good way to think about this is by visualizing a curve.

This is from work by Sean Dolinar and Jonah Pemstein which takes a methodology similar to Carleton’s but doesn’t focus on the .49 R^2 (technically Cronbach’s Alpha for the technically inclined). Instead, they plot the reliability measure for each 10 PA increment to better show the nature of stabilization. As you can see, K% crosses the .49 threshold much more quickly that the other statistics, which is consistent with our common understanding, but this also shows that the first 200 PA are much more important than the second 200 PA for understanding K%, something you wouldn’t necessarily see in the Carleton studies. Here is a link to their update work which allows you to visualize many different statistics using this method. The same tool is available below.

Practical Use:

For practical purposes, you really want to know the difference between a sample that’s meaningful and one that isn’t. There isn’t a point at which it becomes useful data all of a sudden, but there are quantities that are clearly one or the other. For example, 400 PA is enough to tell you a lot about strikeout rate and 30 PA is not enough to tell you much about HR rate. The real trick is how you update your beliefs between those two extremes.

Every April, at least one previously bad hitter has an awesome month. They have a .380 wOBA over three weeks and lots of people rush to suggest they are a breakout candidate who did something during the offseason to improve. It’s important to note that this may be true or it may not be true. All we know is that they hit .380 wOBA over three weeks, let’s call it 85 PA.

Are those 85 PA enough to lead us to totally change our opinion about this bad hitter to the point where we now think they are fundamentally different in the box? Using our sample size rules of thumb, the answer is no. A bad hitter can easily have a .380 wOBA over 85 PA without actually being a different hitter, just due to random chance. A couple lucky bounces and a well timed cluster of hits and his numbers look great even if he’s no different than he was before.

Those 85 PA give you some idea that he might be improving, but they are not sufficient to change your mind completely. A true .380 wOBA hitter should hit .380 over more stretches than a .310 wOBA hitter, but a .310 wOBA hitter can hit .380 over a month no problem.

The goal of these “stabilization” or “reliability” numbers is to prevent you from reacting to data that is still highly susceptible to random chance. Good hitters can have bad results in small samples even when their process is fine. The larger the sample gets, the more we can discount this random noise component and zero in on factors that are within the player’s control. That’s the point of sample size; preventing us from ascribing too much meaning to small chunks of data.



Rules of Thumb:

Here is a tool that allows you to explore those reliability graphs mentioned earlier. If it’s not loading on this page for you, here is a link you can use to find it.

<a href=’#’><img alt=’ ‘ src=’http://public.tableau.com/static/images/ML/MLBStatReliability_0/Batters/1_rss.png’ style=’border: none’ /></a>

If you’re only looking for Carleton’s .49 cut points, they are listed below.

“Stabilization” Points for Offense Statistics:

60 PA: Strikeout rate

120 PA: Walk rate

240 PA: HBP rate

290 PA: Single rate

1610 PA: XBH rate

170 PA: HR rate

910 AB: AVG

460 PA: OBP

320 AB: SLG

160 AB: ISO

80 BIP: GB rate

80 BIP: FB rate

600 BIP: LD rate

50 FBs: HR per FB

820 BIP: BABIP

“Stabilization” Points for Pitching Statistics:

70 BF: Strikeout rate

170 BF: Walk rate

640 BF: HBP rate

670 BF: Single rate

1450 BF: XBH rate

1320 BF: HR rate

630 BF: AVG

540 BF: OBP

550 AB: SLG

630 AB: ISO

70 BIP: GB rate

70 BIP: FB rate

650 BIP: LD rate

400 FB: HR per FB

2000 BIP: BABIP

Links to Further Reading:

525,600 Minutes: How Do You Measure a Player in a Year? – Statistically Speaking / Pizza Cutter

On the Reliability of Pitching Stats – Statistically Speaking / Pizza Cutter

When Samples Become Reliable – FanGraphs

The Beginner’s Guide to Sample Size – FanGraphs

A New Way to Look at Sample Size – FanGraphs

A Long-Needed Update on Reliability – FanGraphs

Should I Worry About My Favorite Pitcher – Baseball Propsectus

It’s A Small Sample Size After All – Baseball Prospectus

Reliably Stable (You Keep Using That Word) – Baseball Propsectus

-Neil Weinberg, Updated August 2017