Authors: Tobias Moskowitz
Now the contradiction between the strong belief in the hot hand or momentum in sports and the lack of actual evidence starts to make sense. A basketball player who shoots 50 percent will not miss an attempt and then make an attempt. A batter may hit .300, but it’s only an average. It doesn’t mean that he’ll get three hits in
every
ten at-bats. He might go 0 for 10 and then 6 for 10. Over the 600 at-bats throughout a season, however, he
probably will
get 180 hits. The larger a sample, the more accurately it represents reality.
Kobe Bryant shoots free throws much better than
Shaquille O’Neal does. For their respective careers, Kobe hits about 84 percent from the line, and Shaq only 53 percent. Take a sequence of
only five shots, however, and it’s entirely possible they’ll shoot comparably. It’s even reasonably possible that Shaq will outshoot Kobe. In fact, the chances are about 22 percent; that means if Shaq and Kobe staged a five-shot free throw shooting contest, about one out of every five times Shaq would do at least as well as Kobe and might even beat him. Over ten attempts, it’s less likely. Over 100, it’s remote.
What does this mean with regard to David Wright’s hitting slump? A career .307 hitter, Wright expects to get a hit 30 percent of the time. Three weeks into the season, after getting a hit only 23 percent of the time, his performance is perfectly consistent with his .307 average. The same would be true on the other side. He could have hit .400 the first few weeks, and fans would be ready to declare him the first player since Ted Williams to bat .400 for a season. Yet over a short period, a .307 career hitter batting .400 is perfectly consistent with random chance, too. Some athletes get this better than others and try to avoid “getting too high or too low.” Wright’s former teammate Jeff Francoeur performs a self-assessment on hitting every 50 at-bats. But even that is woefully narrow.
Being fooled by chance can create seemingly unbelievable statistics. Consider the following, all of which are true. Over the last decade,
in every single
MLB season:
These stats, surprising as they might seem on their face, hold up every year. Pitchers are not supposed to hit better than position players, much less all-stars. Players hitting under .225 aren’t supposed to have longer hitting streaks than .300 hitters. The best batters aren’t supposed to go six games—25 or so at-bats—without a hit, and on average, they don’t. But in isolated cases it happens, and it’s perfectly consistent with random chance.
We search for an explanation, but the true explanation is simple: Luck or chance or randomness causes streaks among even the best and worst players. It has nothing to do with momentum. When we consider the bigger picture and the larger numbers of players in Major League Baseball, this starts to make sense. How likely is it that
Tim Lincecum, the star pitcher for the San Francisco Giants, will outhit the mighty Albert Pujols over any stretch of the season? Not very. How likely is it that
any
pitcher will outhit Pujols over a two-week stretch? More likely. How likely is it that
at least one
pitcher will outhit
at least one
all-star position player over those weeks? Very, very likely. The larger the sample, the more you can find at least one seemingly unlikely example.
If you were predicting the likelihood of an MLB player getting a hit in his next at-bat, which of the following do you think would be the best predictor?
Most people are tempted to select (a), on the grounds that it is the most recent and therefore the most relevant number: He’s streaking and will continue riding the wave. Or, he’s slumping and still mired. But to pick (a) is to be fooled by randomness, tricked into thinking there’s momentum.
We looked at all MLB hitters over the last decade and tried predicting the outcome of their next at-bats by using each of the five choices above. It turns out (a) is the worst predictor. Why? Because it has the smallest sample size. Choice (b) was the next worse, then (c), and then (d). The best answer was (e), the choice with the largest sample size.
The same thing is true at the team level. Heading into the postseason—and barring the unusual, such as a recent horrific injury to a star—which of the following is a better predictor of playoff success?
Momentum would lead one to think that it’s (a) or (b) and, to a lesser extent, (c), yet those are actually the worst predictors. In
every single sport
(MLB, NBA, NHL, NFL, European soccer) we studied, we found (d) to be the best predictor of postseason or tournament success. The true quality of teams can be measured best in large samples. Small samples are more dominated by randomness and therefore are inherently unreliable.
Nor is this unique to sports. In the investment management industry, investors often “chase short-term returns,” flocking toward mutual funds that had a good quarter or year and fleeing from funds that didn’t. They ascribe success on the basis of a small sample of data. But as with the hot hand in sports, it turns out that one quarter or even one year of a fund’s performance has no special predictive power for the next year’s performance in the mutual fund industry. In fact, one year of performance for almost any fund is dominated by luck, not skill. Yet people usually don’t see it that way. Entire businesses have been built on selling short-term performance measures to investors to help them identify the best
funds, and funds aggressively market their recent strong performance to investors (and hide or bury bad performance when they can). But the reality is that every year the top 10 percent of funds are just as likely to be among the bottom 10 percent of funds the next year. It’s just pretty much random from year to year.
Sports gamblers, too, are fooled by momentum.
Colin Camerer, a Caltech professor of behavioral economics, found that winning and losing streaks affected point spreads. Bets placed on teams with winning streaks were more likely to lose, and bets placed on teams with losing streaks were more likely to pay off. In other words, gamblers systematically overvalued teams with winning streaks and undervalued those with losing streaks.
Just as an astute investor can take advantage of these misperceptions with potentially big gains, so can a savvy coach and player (and sports gambler). If the majority overvalues the recent winners and undervalues the recent losers, do the opposite.
The only problem is convincing people to go against their (and everyone else’s) intuition. After the initial study asserting the fallacy of the hot hand in basketball,
Red Auerbach, the revered Hall of Fame coach and then president of the Boston Celtics, was presented with the findings. Auerbach rolled his eyes and waved the air with his hand. “Who is this guy? So he makes a study. I couldn’t care less.”
Bob Knight, the volatile and decorated college coach, was similarly dismissive: “There are so many variables involved in shooting the basketball that a paper like this doesn’t really mean anything.” Amos Tversky, the famous psychologist and pioneering scholar who initiated the original research on momentum and the myth of the hot hand, once put it this way: “I’ve been in a thousand arguments over this topic. I’ve won them all, and I’ve convinced no one.”
“There are three kinds of lies: lies, damned lies, and statistics.”—Mark Twain
At some point it became almost cartoonish, as though he wasn’t shooting the basketball so much as simply redirecting his teammates’ passes into the hoop. In the first half of the second game of the 2010 NBA finals,
Ray Allen, the Boston Celtics’ veteran guard was … well, the usual clichés—“on fire,” “unconscious,” “in the zone”—didn’t do it justice. Shooting with ruthless accuracy, Allen drained seven three-pointers, most of them bypassing the rim and simply finding the bottom of the net. Swish. Swish. Swish-swish-swish. In all, he scored 27 points in the first half. Celtics reserve
Nate Robinson giddily anointed Allen “the best shooter in the history of the NBA.”
As Allen fired away, the commentators unleashed a similarly furious barrage of stats, confirmed by the graphics on the screen. The shooting was cast in the most glowing terms possible. Allen, viewers were told at one point, had made his last four shots. When he missed a two-pointer (turns out he was only three for nine
on two-point attempts), the stats suddenly focused only on the three-pointers.
It was inevitable that Allen would cool off. And he did in the second half, making only one three-pointer, although his eight treys for the game became a new NBA finals record and his 32 total points enabled Boston to beat the Los Angeles Lakers 103–94. But he
really
cooled off in his next game. This time he was ruthless in his
in
accuracy, missing all 13 of his shots, including eight three-point attempts, as Boston lost 91–84. As Allen clanged shot after shot, the commentators were quick to note this whiplash-inducing reversal of fortune, framing it in the most damning terms possible. At one point viewers were told that between the two games, Allen had missed 17 straight attempts.
Inasmuch as sports fans are tricked by randomness, the media share in the blame. Statistics and data are the forensic evidence of sports, but like all pieces of evidence, they can be mishandled and tampered with. We are bombarded by stats when we watch games, but the data are chosen selectively and often focus on small samples and short-term numbers. When we’re told that a player has reached base in “four of his last five at-bats,” we should assume right away that it’s four of his last six. Otherwise, rest assured, we’d have been told that the streak was five out of six. Clearly, a team that “has lost three in a row” has dropped only three of its last four—and possibly three of five or three of six or … otherwise it would have been reported as a four-game losing streak.