What Happens When you Data Mine 2 million Fundamental Quant Strategies

////What Happens When you Data Mine 2 million Fundamental Quant Strategies

What Happens When you Data Mine 2 million Fundamental Quant Strategies

By | 2018-06-05T09:06:08+00:00 September 13th, 2017|Factor Investing, Research Insights|14 Comments

As we have mentioned before, herehere and here, there is overwhelming evidence that the number of stock anomalies in the universe is much lower than originally thought. Most of the previous research papers attempt to filter out past anomalies in the literature (generally over 300+) by applying more stringent standards, such as higher p-values or more advanced statistical tests.

A working paper we examine below, “p-hacking: Evidence from two million trading strategies” by Chordia, Goyal and Saretto, takes an alternative approach. The authors take the Compustat universe of data points, and use every variable in the dataset to create over 2 million trading strategies — explicit data-mining!

The idea behind the paper is to examine what is possible if one simply data-mined the entire universe of signals. The authors make an effort to only examine tradeable strategies by eliminating small and micro-cap stocks. In addition, the authors apply more stringent statistical standards (which I will discuss below) to identify the true anomalies in the data.

After examining all the signals, the authors find only a handlful of trading strategies that are “anomalous” and most of these strategies make no economic sense! Now the authors do assume (through their tests), that the Fama and French 5-factor model plus momentum explain the cross-section of stock returns (so all the classic characteristics we all argue about are controlled for in the study), but the author’s main contribution is that there is little to no evidence for additional anomalies.

However, many papers have already found this. So why is this paper important?

A newer topic that we are commonly asked about is machine learning. Many are intrigued by the idea — let the computer with its algorithms come up with the best trading signals using all the data. At first, this sounds great (and in certain contexts can be extremely useful). But taking a step back, we need to examine what happens if we simply examine all the data. This paper highlights that trading on every signal in the fundamental-signal universe yields almost no (additional) anomalies.(1)It may be the case that machine-learning is great at combining the already well-known anomalies that the authors assume in the paper (such as Value and Momentum); however, machine-learning may also end up enhancing frictional costs and increasing the chances of a data-mined result (despite the algorithms best efforts to avoid this problem).

Below we dig into the details of the paper.

The data

The paper examines the idea of finding anomalies in a different manner than most — it simply data mines.

Here is the high-level summary from the paper:

We consider the list of all accounting variables on Compustat and basic market variables on CRSP. We construct trading signals by considering various combinations of these basic variables and construct roughly 2.1 million different trading signals.

Two additional screens that I like from that paper are that they (1) eliminate all firms with stock prices below $3 as well as those below the 20th percentile for market capitalization and (2) require all the variables to have information to include the firm in the sample.(2)

The paper then examines the 156 variables in the Compustat library (listed in Appendix A1 of the paper) to create over 2 million trading signals. Here is how the signals are constructed, directly from the paper:

There are 156 variables that clear our filters and can be used to develop trading signals. The list of these variables is provided in Appendix Table A1. We refer to these variables as Levels. We also construct Growth rates from one year to the next for these variables. Since it is common in the literature to construct ratios of different variables we also compute all possible combinations of ratios of two levels, denoted Ratios of two, and ratios of any two growth rates, denoted Ratios of growth rates. Finally, we also compute all possible combinations that can be expressed as a ratio between the difference of two variables to a third variable (i.e., (x1 − x2)/x3). We refer to this last group as Ratios of three. We obtain a total of 2,090,365 possible signals.

Since the paper has already eliminated small and micro cap stocks from the tests, they form portfolios using a one-dimensional sort on each of the variables. The portfolios are rebalanced annually, creating long/short portfolios that go long the top decile on each measure, and short the bottom decile.

The paper tests these 2 million portfolios by (1) regressing the L/S portfolio returns against the Fama and French 5-factor model plus the momentum factor and (2) examining Fama-MacBath (FM) regressions.

The Tests and Results

Before getting into the specific results, a good exercise (especially when data-mining) is to simply examine the distribution of outcomes. Figure 1 (shown below) in the paper shows the distributions and t-stats.

Source: p-Hacking: Evidence from Two Million Trading Strategies. Accessed from SSRN on 8/28/17. The results are hypothetical results and are NOT an indicator of future results and do NOT represent returns that any investor actually attained. Indexes are unmanaged, do not reflect management or trading fees, and one cannot invest directly in an index. Additional information regarding the construction of these results is available upon request.

As viewed from the distributions, most are centered around 0.(3)The question is as follows: how robust are these trading strategies with significant alphas and Fama-MacBeth coefficients?

Examining raw returns first, the paper finds 22,237 portfolios with T-stats above 2.57 (in absolute value) — this is less than 1% of the total portfolios. Next, the paper examines the 6-factor regressions and finds that around 31% of the sample has a significant alpha at that 5% level, and 17% of the sample are significant at the 1% level. Last, examining the Fama-MacBeth regressions, the paper finds similar results — 31% of the sample has a t-stat above 1.96, and 18% of the sample has a t-stat above 2.57.

Based on these independent tests (alphas and FM regressions), the results are promising. However, the authors dig into the statistics with more advanced tests.

The reason to do this, as we discussed here before, is that as the number of ideas (in our case, 2 million) increases, the probability of Type 1 Errors increases. The authors describe this well in their paper:

Classical single hypothesis testing uses a significance level \alpha to control Type I error (discovery of false positives). In multiple hypothesis testing (MHT), using \alpha to test each individual hypothesis does not control the overall probability of false positives. For instance, if test statistics are independent and normally distributed and we set the significance level at 5%, then the rate of Type I error (i.e., the probability of making at least one false discovery) is 1 – 0.95^10 = 40% in testing ten hypotheses and over 99% in testing 100 hypotheses. There are three broad approaches in the statistics literature to deal with this problem: family-wise error rate (FWER), false discovery rate (FDR), and false discovery proportion (FDP). In this section, we describe these approaches and provide details on their implementation.

The authors test and discuss this multiple hypothesis testing framework in Section 3 of the paper and the results are documents in Table 4.(4) However, false discoveries can still occur (by definition the tests allow this). To correct for this, the authors impose economic hurdles.The hurdles are listed below:

The hurdles are listed below:

  1. The strategy must be statistically significant for both the (1) 6-factor regression and (2) the Fama-MacBeth regression.
  2. The strategy must have a Sharpe ratio above the market’s over the time period studied, as well as in both sub-samples (splitting the dataset in two)

The full results (using multiple tests) are in Table 5 of the paper.The authors summarize the findings here:

The authors summarize the findings here:

In summary, in the most optimistic scenario where we consider the least stringent BHYS approach (and, therefore, neglect to account for cross-correlation in the data), we find at most 345 economically significant strategies (52 if we impose some persistence in economic performance). In the least optimistic scenario using the FWER approach, we find 5 strategies. If we properly account for the statistical properties of the data-generating process and use the FDP approach, we are left with a handful of exceptional investment opportunities. If we adopt an all-together conservative approach and control FDP at \gamma = 1% (i.e., we accept one per cent of lucky discovery among all discoveries on average or in our sample), we reject all the two million strategies.

So high-level, very few strategies are significant using the authors’ test requirements!

A natural question is what are the strategies that survive, and do they make sense?

The authors examine some of the strategies in Table 6 of the paper (shown below):

Source: p-Hacking: Evidence from Two Million Trading Strategies. Accessed from SSRN on 8/28/17.

The 17 strategies in this Table are all different than the 447 strategies tested in Replicating Anomalies (2017) by Hou, Xue, and Zhang. For those who aren’t used to using Compustat variables, a list of the names can be found here. So examining the first strategy above (that is statistically significant) the proposed strategy is to sort stocks on Common/Ordinary Stock (cstk) minus Retained Earnings/Other Adjustments (reajo) divided by Advertising Expense (xad).

The majority of the other strategies above are just as absurd (have fun with the Compustat definition link!).


Overall, this paper examines what is possible if one simply data-mined the Compustat database. The authors impose MHT as well as 2 restrictions, that the strategies need to work for (1) both 6-factor and Fama-MacBeth regressions and (2) pass a Sharpe ratio test. After these restrictions, the authors find very few strategies that are significant. It is important to remember that the authors’ baseline assumption is that the 6-factors (Fama and French 5-factors plus momentum) are a given. However, the next time someone pitches you a machine learning algorithm, keep this paper in mind. Otherwise, you may be investing in a strategy that sorts on ratios that make no economic sense!

Let us know what you think!

PS: Note I am being a little flippant towards machine learning here, which is not to say it has no value (Google has proven that!). However, I am trying to highlight that machine learning generally comes up with ideas that have been studied in the past, such as Value, Momentum, and Quality to name a few. Machine learning may be able to optimize when to get into and out of a factor (factor-timing), but that has already been shown to be difficult; additionally, one should always consider tax/frictional consequences when investing, and models with more trading should be discounted appropriately for taxable investors.

p-hacking: Evidence from two million trading strategies

  • Tarun Chordia, Amit Goyal, and Alessio Saretto
  • A version of the paper can be found here.


We implement a data mining approach to generate about 2.1 million trading strategies. This large set of strategies serves as a laboratory to evaluate the seriousness of p-hacking and data snooping in finance. We apply multiple hypothesis testing techniques that account for cross-correlations in signals and returns to produce t-statistic thresholds that control the proportion of false discoveries. We find that the difference in rejections rates produced by single and multiple hypothesis testing is such that most rejections of the null of no outperformance under single hypothesis testing are likely false (i.e., we find a very high rate of type I errors). Combining statistical criteria with economic considerations, we find that a remarkably small number of strategies survive our thorough vetting procedure. Even these surviving strategies have no theoretical underpinnings. Overall, p-hacking is a serious problem and, correcting for it, outperforming trading strategies are rare.

  • The views and opinions expressed herein are those of the author and do not necessarily reflect the views of Alpha Architect, its affiliates or its employees. Our full disclosures are available here. Definitions of common statistics used in our analysis are available here (towards the bottom).
  • Join thousands of other readers and subscribe to our blog.
  • This site provides NO information on our value ETFs or our momentum ETFs. Please refer to this site.

Print Friendly, PDF & Email

References   [ + ]

1. It should be noted that this paper only examines fundamental anomalies. It may be possible that there are more trading signals in the price data, but one needs to assess those signals in the context of other known price anomalies, such as short-term reversion, intermediate-term continuation, and long-term reversion.
2. This 2nd restriction can be somewhat restrictive, but will (if anything) bias the sample towards larger securities.
3. The paper provides more detail on some of the significant trading strategies in the Appendix Tables
4. For those interested in the weeds of the regressions and the tests, I recommend reading section 3 of the paper

About the Author:

Jack Vogel, Ph.D., conducts research in empirical asset pricing and behavioral finance, and is a co-author of DIY FINANCIAL ADVISOR: A Simple Solution to Build and Protect Your Wealth. His dissertation investigates how behavioral biases affect the value anomaly. His academic background includes experience as an instructor and research assistant at Drexel University in both the Finance and Mathematics departments, as well as a Finance instructor at Villanova University. Dr. Vogel is currently a Managing Member of Alpha Architect, LLC, an SEC-Registered Investment Advisor, where he heads the research department and serves as the Chief Financial Officer. He has a PhD in Finance and a MS in Mathematics from Drexel University, and graduated summa cum laude with a BS in Mathematics and Education from The University of Scranton.


  1. Thomas Musselman September 13, 2017 at 12:27 pm

    It would be nice for people to try out shorter-term holds; 1 year is great for the tax rate and mutual funds but us little guys…..

  2. Ankush September 13, 2017 at 2:08 pm

    Apparently, the author of your article has absolutely no knowledge of Machine Learning. Such a shame that Machine Learning has been undermined in this article. You say, “..machine learning generally comes up with ideas that have been studied in the past..”
    Fyi: Machine Learning comes up with predictions based on concrete mathematical algorithms. (Ideas are vague)
    Secondly, in ML we construct additional features (not just historical) using not only data mining tools, as discussed in your article, but also using other important branches of mathematics such as linear algebra and calculus.
    All in all, i strongly believe that if the author of this article has little or no knowledge about Machine Learning, he better not make a mockery of it


    • Jack Vogel, PhD September 13, 2017 at 2:13 pm

      As the author, that was not the intent. Using ML has benefits, no disagreement from me.

    • Michael Harris September 15, 2017 at 1:40 pm

      Performance of machine leaning depends on the features, predictors, factors, etc. used. Constructing additional features, a.k.a feature engineering, that are suggested by the data and by any mathematical means is exactly what multiple comparisons and p-hacking are all about. However, machine learning seems to work well with features that have economic value and are based on solid foundation. But in general, the way machine learning is practiced leads to statistically insignificance models. This is because 95% of the job is to find the relevant features, and machine learning is just the 5% of the job to classify and obtain the probabilities of class 1.

      • gregorsam September 15, 2017 at 2:26 pm

        Not quite true.

        Breiman’s (sp?) Random Forests algorithm for example does not require you to engage in ‘feature engineering’. the sorting out of the relevant features is a part of the process. Similarly, by using kernels, in, eg. SVM, you minimize the need for constructing complex features from the raw data.

        The boundary between Machine Learning and Statistical methods is quite fuzzy. What does one call the Fisher’s discriminant? Ankush has a slight point but he overstates his case.

        • Michael Harris September 15, 2017 at 3:22 pm

          I don’t think I claimed that feature engineering is required. As far as Breiman’s RF, taking as a reference the Numerai contest, it had the lowest performance from about a dozen algos I used for several months. Furthermore, by using Kernels, risk of over-fitting is increased. In general risk of over-fitting is high with ML but again using ML is more of an art than the science involved.

  3. JN September 13, 2017 at 2:34 pm

    Thanks for the posting Jack! The paper is interesting in a way that it provides a direct testing of p-hacking in the anomaly space, but I agree with Ankush that relating p-hacking / data mining to machine learning is a bit unfair and confusing.

    • Jack Vogel, PhD September 13, 2017 at 2:53 pm

      Happy to post. I do agree ML has benefits; this paper simply highlights that using their methodology (to come up with 2+ million trading strategies) yields very few strategies that meet the author’s hurdles imposed on the data.

      • bubba123 September 20, 2017 at 6:06 am

        Sounds like Bogle lol

  4. Michael Harris September 15, 2017 at 1:32 pm

    Excellent analysis of the paper but I claim that statistical significance is the wrong way to go about this problem due to both Type I but also Type II errors. Determining when market conditions change is much more important at the end of the day than any statistical analysis. Here is my paper on th9is https://papers.ssrn.com/sol3/papers.cfm?abstract_id=2810170

  5. G-Man September 17, 2017 at 11:33 am

    Thanks for another great post!
    With regard to the Compustat variables how do you get EBIT (as part of the EBIT/EV valuation screen in the construction of your “value index”?). The closet thing I could find was “Operating Income After Depreciation”.
    Kind regards,

    • Jack Vogel, PhD September 18, 2017 at 8:49 am

      Close, we add back non-operating income as well. We examined multiple ways to calculate this and there were no statistical differences.

  6. Leigh Fanning
    Leigh Fanning October 5, 2017 at 3:13 pm

    2090365 strategies sounds impressive but it hardly touches the search space. Considering just the parameters, we have 2^156 parameter subsets (ok, subtract one for the empty set) from which to build formulas with. For each of these, then consider all possible linear formulas, then all possible quadratic formulas, ad infinitum. Humans can’t comprehend, but a deep learner might go a lot farther, especially powered by a quantum engine underneath. For me, the larger contributions are a data-driven approach to parameter selection, admission that many are not independent, and the strategy evaluation ideas.

Leave A Comment