Imagine you’re looking for variables that predict the cross-section of expected returns. No search process is perfect. As you work, you will inevitably uncover both tradable anomalies as well as spurious correlations. To figure out which are which, you regress returns on each variable that you come across:

This is helpful because “any predictive regression can be expressed as a portfolio sort” and vice versa. So, a statistically significant suggests a profitable stock-picking strategy.

But, what qualifies as a “statistically significant” test result? If a variable doesn’t actually predict returns, then the probability that its will have a -stat greater than is *by definition*:

But, what if you didn’t just consider one variable on its own? Instead, suppose that you ran separate regressions. If all variables were just spurious correlations, the probability that at least one of these s has a -stat greater than would be much larger than :

Finding a becomes meaningless as you run more and more regressions. e.g., if , then you should expect to see a -stat larger than more than of the time.

When people in academic finance talk about the problem of “data mining”, this is what they’re referring to. It seems patently obvious that having such a high false-positive rate is a bad thing. And, at first glance, it seems like there’s an easy way to fix the problem: just use a larger cutoff for statistical significance. e.g., researchers have suggested using a *t*-stat greater than 3.00 rather than to account for the fact that we’ve proposed thousands of candidate variables. But, is our obsession with minimizing the false-positive rate really the right approach? Do we always want to choose our statistical tests so that they have the lowest possible false-positive rate? Not necessarily. And, this post describes two reasons why.

## Reason #1

We don’t care about the false-positive rate for its own sake. What we really want to know is: “Conditional on observing a significant test result, how likely is it that we’ve found an honest-to-goodness anomaly?” Using Bayes’ theorem, we can write this conditional probability as

Clearly, if we underestimate the false-positive rate, , then we’re going to overestimate this conditional probability because we’re going to be dividing by a smaller number on the right-hand side.

But, isn’t the only term on the right-hand side of the equation! We care about more than just the false-positive rate when updating our priors. e.g., if we knew there were never any anomalies, then we could guarantee that every single significant result was a false positive. So, we would always conclude that regardless of how much we underestimated the false-positive rate.

Let’s sharpen this insight with a little algebra. First, suppose that the unconditional probability of finding a tradable anomaly is :

Next, suppose that the probability of observing a significant test result for a spurious correlation is for while the probability of observing a significant test result for a tradable anomaly is :

Thus, represents the base rate of observing anomalies, and represents the additional false-positive rate introduced by the data-mining problem described above. Roughly speaking, a larger means that anomalies are more common. And, a large means that regressions are easier to run.

We can express our posterior beliefs that a particular variable represents a tradable anomaly given a significant test result as

Thus, ignoring the additional false positives introduced by data mining biases our posterior beliefs by

The figure to the right plots this bias (y-axis: ) as a function of the excess false-positive rate (x-axis: ). The line is always sloping upward because the more you underestimate a test’s false-positive rate the more confident you will be that you’ve found an anomaly. However, the shape of the line dramatically changes as you play around with the slider, which adjusts . When , the plot has more or less the same upward-sloping concave shape. But, when , the shape of the plot flattens out dramatically. In other words, if the base rate is sufficiently small, then underestimating the false-positive rate doesn’t affect our posterior beliefs very much.

This observation implies something a little counterintuitive: a paper can’t simultaneously argue that A) almost all documented anomalies are in fact spurious correlations, and that B) it’s super important for other researchers to use a test procedure that they’ve proposed which minimizes the false-positive rate. Although they get made one after the other (e.g., here), these two claims aren’t internally consistent. It’s one or the other. Minimizing the false-positive rate can only matter if there’s a non-negligible chance of finding a tradable anomaly.

## Reason #2

If you’re headed to the doctor’s office for a pregnancy test, then false positives matter. Finding out you’re pregnant is a big deal. Later discovering that it was a mistake would be traumatic. But, testing for Coeliac disease is different. If you have the disease, then you really want to know. But, treating the disease only involves changing your diet. There’s no need to undergo risky surgery or take expensive medication. So, when testing for Coeliac disease, false positives aren’t such a big deal (unless you looooove bread). And, if given the choice, your doctor should choose a test for the disease that has the lowest false-negative rate, even if that means asking a few perfectly healthy patients to cut gluten out of their diets.

The same sort of logic applies to testing for tradable anomalies. If you can trade on a statistically significant anomaly using liquid actively-traded stocks, then why spend time worrying about the false-positive rate. If you find out you’re wrong, then you can quickly and painlessly exit the position. If this is the sort of world you’re operating in, then you might actually want to set up your statistical tests to minimize your false-negative rate. This is one way to interpret pithy trader sayings like “invest first, investigate later”.

The medical literature also gives a nice way of formalizing this idea using something called the number needed to treat (e.g., see here). Suppose I came to you with a bunch of variables that each seemed to predict the cross-section of expected returns. They each delivered significant excess returns in backtesting. If you choose of these variables

then you should expect your selections to contain one more tradable anomaly than if you had just picked variables at random—i.e., regardless of whether they had delivered significant excess returns in backtesting.

Now, think about a portfolio that invests the same amount of money in strategies based on each of the candidate variables you choose. This portfolio’s expected return will depend on both the profitability of your one extra tradable anomaly and the losses of your other spurious predictors:

Clearly, a higher false-positive rate means a larger . But, the formulation above illustrates why this might not be such a bad thing. If trading on your one genuine anomaly is really profitable and you can quickly identify/exit your remaining spurious positions, then who cares if is large?