1. Introduction
Imagine you are an algorithmic trader, and you have to set up a trading platform. How many signals should you try to process? How many assets should you trade? If you are like most people, your answer will be something like: “As many as I possibly can given my technological constraints.” Most people have the intuition that the only thing holding back algorithmic trading is computational speed. They think that left unchecked, computers will eventually uncover every trading opportunity no matter how small or obscure. e.g., as Scott Patterson writes in Dark Pools, these computerized trading platforms are just “tricked-out artificial intelligence systems designed to scope out hidden pockets in the market where they can ply their trades.”
This post highlights another concern. As you use computers to discover more and more “hidden pockets in the market”, the scale of these trades might grow faster than the usefulness of your information. Weak and diffuse signals might turn out to be really risky to trade on. How might this work? e.g., suppose that in order for your computers to recognize the impact of China’s GDP on US stock returns you need to feed them data for assets; however, in order for your computers to recognize the impact of recent copper discoveries in Guatemala on electronics manufacturing company returns, you need to feed your machines data on assets. Even if the precision of the signal that your computers spit out is the same, its magnitude will be smaller. Copper discoveries in Guatemala just don’t matter as much. Thus, you would take on more risk by aggressively trading the second signal because it’s weaker and you would have to take on a position in times as many assets! Thus, this post suggests that even if you might want to maximize the number of inputs your machines get, you might want to limit how broadly you apply them.
First, in Section 2 I illustrate how the scale of a trading opportunity might increase faster than the precision on your signal via an an example based on the well known Hodges’ Estimator. You should definitely check out Larry Wasserman‘s excellent post on the this topic for an introduction to the ideas. In Section 3 I then outline the basic decision problem I have in mind. In Section 4 I show how the risk associated with trading weak signals might explode as described above. Finally, in Section 4 I conclude by suggesting an empirical application of these ideas.
2. Hodges’ Estimator
Think about the following problem. Suppose that are random variables denoting (say) innovations in the quarterly dividends of a group of stocks:
(1)
You want to know if the mean of the dividend changes is non-zero. i.e., has this group of stocks realized a shock? Define the estimator as follows:
(2)
where . This estimator says: “If the average of the dividend changes is sufficiently big, I’m going to assume there has been a shock of size ; otherwise, I’ll assume that there’s been no shock.” This is an example of a Hodges-type estimator. is a consistent estimator of in the sense that:
(3)
Thus, as you examine more and more stocks from this group, you are guaranteed to discover the true mean. However, the worst case expected loss of the estimator, , is infinite!
(4)
This is true even though the worst case loss for the sample mean, , is flat:
(5)
I plot the risk associated with Hodges’ estimator in the figure below. What’s going on here? Well, as gets bigger and bigger, there remains a region around where you are quite certain that the mean is . However, at the edge of this region, there is a band of values of where if you are wrong and , then your prediction error summed across every one of the stocks you examined turns out to be quite big.
3. Decision Problem
Now, think about a decision problem that generates a really similar decision rule—namely, try to figure out how many shares, , to purchase where the stock’s payout is determined by different attributes via the coefficients :
(6)
Here, you are trying to maximize your value, , by choosing the right number of shares to hold—i.e., the risk action. Ideally you would take the action which is exactly equal to the ideal action ; however, it’s hard to figure out the exact loadings, , for every single one of the relevant dimensions. As a result, you take an action which isn’t exactly perfect:
(7)
I use the function to denote the loss in value you suffer by choosing a suboptimal asset holding:
(8)
where the rd line follows from assuming that .
So which of the different details should you pay attention to? Which of them should you ignore? One way to frame this problem would be to look for a sparse solution and only ask your computers to trade on the signals that are sufficiently important. Gabaix (2012) shows how to do this using an -program, I outline the main ideas in an earlier post. Basically, you would try to minimize your loss from taking a suboptimal action subject to an -penalty:
(9)
so that your optimal choice of actions is given by:
(10)
That’s the decision problem. So far everything is old hat.
4. Maximum Risk
The key insight in this post is that these terms don’t come down from on high. As a trader, you don’t just know what these terms are from the outset. They weren’t stitched onto the forehead of your favorite teddy bear. Instead, you have to use data on lots of stocks to estimate them as you go. In this section, I think about a world where you feed data on different stocks to your machines, and each of these assets has the appropriate action . I then investigate what happens when you use the estimator:
(11)
instead of the estimator in Equation (10) where and isn’t growing that fast as gets larger and larger. i.e., we have that:
(12)
In some senses, is still a really good estimator of . To see this, let denote the true effect for a particular attribute. Then, clearly for , we have that:
(13)
Similarly, for , we have that:
(14)
Thus, the estimate is a strictly better estimator of than the sample average. Pretty cool.
However, in another sense, it’s a terrible estimator since the maximum risk associated with trading on is unbounded:
(15)
This result means that if you use this estimator to trade on, there are parameter values for which lead your computer to take on positions that are infinitely risky and make you really unhappy! How is this possible? Well, take a look at the following decomposition of the risk associated with the estimator :
(16)
It turns out that there are choices of for which as , but at the same time for the same process. In words, this means that there are choices of which make dimension always appear to your machines as an unimportant detail no matter how many stocks you look at, but this conclusion is still risky enough that if you applied to every one of these stocks you’d get an infinitely risky portfolio.
e.g., consider the case where:
(17)
Then, it’s easy to see that:
(18)
where . But, this means that as , we have that:
(19)
Thus, we have a proof by explicit construction. The punchline is that unleashing your machines on the market might lead them into situations where the scale of the position grows faster than the precision of the signal.
5. Empirical Prediction
What does this result mean in the real world? Well, the argument in the section above relied on the fact that when you used different assets to derive a signal, you also then traded on all assets. i.e., this is why the risk function has a factor of in it. Thus, one way to get around this problem would be to commit to trading a smaller number of asset where:
(20)
even though you had to use assets to recover the signal. i.e., even if you have to grab signals from the corners of the market to create your trading strategy, you might nevertheless specialize in trading only a few assets so that in contrast to Equation (19) you would always have:
(21)
If traders are actually using signals from lots of assets, but only trading a few of them to avoid the infinite risk problem, then this theory would give a new motivation for excess comovement. i.e., you and I might feed the same data into our machines, get the same output, and choose to trade on entirely different subsets of stocks.
You must be logged in to post a comment.