unnecessary labor, and has the ability to make judgments towards those ends.”
At first glance, this assumption seem awfully reasonable. Who among us isn’t self-interested? And who wouldn’t avoid unnecessary labor, given the chance? (Why clean your apartment unless you know that guests are coming?)
But as the architect Mies van der Rohe famously said, “God is in the details.” We are indeed good at dodging unnecessary labor, but true rationality is an awfully high standard, frequently well beyond our grasp. To be truly rational, we would need, at a minimum, to face each decision with clear eyes, uncontaminated by the lust of the moment, prepared to make every decision with appropriately dispassionate views of the relevant costs and benefits. Alas, as we’ll see in a moment, the weight of the evidence from psychology and neuroscience suggests otherwise. We
Appreciating what we as a species can and can’t do well — when we are likely to make sound decisions and when we are likely to make a hash of them — requires moving past the idealization of economic man and into the more sticky territory of human psychology. To see why some of our choices appear perfectly sensible and others perfectly foolish, we need to understand how our capacity for choice
I’ll start with good news. On occasion, human choices can be entirely rational. Two professors at NYU, for example, studied what one might think of as the world’s simplest touch-screen video game — and found that, within the parameters of that simple task, people were almost as rational (in the sense of maximizing reward relative to risk) as you could possibly imagine. Two targets appear (standing still) on a screen, one green, one red. In this task, you get points if you touch the green circle; you lose a larger number of points if you touch the red one. The challenge comes when the circles overlap, as they often do, and if you touch the intersection between the circles, you get both the reward and the (larger) penalty, thus accruing a net loss. Because people are encouraged to touch the screen quickly, and because nobody’s hand-eye coordination is perfect, the optimal thing to do is to point somewhere
The bad news is that such exquisite rationality may well be the exception rather than the rule. People are as good as they are at the pointing-at-circles task because it draws on a mental capacity — the ability to reach for things — that is truly ancient. Reaching is close to a reflex, not just for humans, but for every animal that grabs a meal to bring it closer to its mouth; by the time we are adults, our reaching system is so well tuned, we never even think about it. For instance, in a strict technical sense, every time I reach for my cup of tea, I make a set of choices. I decide that I want the tea, that the potential pleasure and the hydration offered by the beverage outweigh the risk of spillage. More than that, and even less consciously, I
But economics is not supposed to be a theory of how people reach for coffee mugs; it’s supposed be a theory of how they spend their money, allocate their time, plan for their retirement, and so forth — it’s supposed to be, at least in part, a theory about how people make
And often, the closer we get to conscious decision making, a more recent product of evolution, the worse our decisions become. When the NYU professors reworked their grasping task to make it a more explicit word problem, most subjects’ performance fell to pieces. Our more recently evolved deliberative system is, in this particular respect, no match for our ancient system for muscle control. Outside that rarefied domain, there are loads of circumstances in which human performance predictably defies any reasonable notion of rationality.
Suppose, for example, that I give you a choice between participating in two lotteries. In one lottery, you have an 89 percent chance of winning $1 million, a 10 percent chance of winning $5 million, and a 1 percent chance of winning nothing; in the other, you have a 100 percent chance of winning $1 million. Which do you go for? Almost everyone takes the sure thing.
Now suppose instead your choice is slightly more complicated. You can take either an 11 percent chance at $1 million or a 10 percent chance of winning $5 million. Which do you choose? Here, almost everyone goes for the second choice, a 10 percent shot at $5 million.
What would be the rational thing to do? According to the theory of rational choice, you
Another experiment offered undergraduates a choice between two raffle tickets, one with 1 chance in 100 to win a $500 voucher toward a trip to Paris, the other, 1 chance in 100 to win a $500 voucher toward college tuition. Most people, in this case, prefer Paris. No big problem there; if Paris is more appealing than the bursar’s office, so be it. But when the odds increase from 1 in 100 to 99 out of 100, most people’s preferences
To take an entirely different sort of illustration, consider the simple question I posed in the opening chapter: would you drive across town to save $25 on a $100 microwave? Most people would say yes, but hardly anybody would drive across town to save the same $25 on a $1,000 television. From the perspective of an economist, this sort of thinking too is irrational. Whether the drive is worth it should depend on just two things: the value of your time and the cost of the gas, nothing else. Either the value of your time and gas is less than $25, in which case you should make the drive, or your time and gas are worth more than $25, in which case you shouldn’t make the drive — end of story. Since the labor to drive across town is the same in both cases and the monetary amount is the same, there’s no rational reason why the drive would make sense in one instance and not the other.
On the other hand, to anyone who
What leads us to think about money in (less rational) relative terms rather than (more rational) absolute terms?
To start with, humans didn’t evolve to think about numbers, much less money, at all. Neither money nor numerical systems are omnipresent. Some cultures trade only by means of barter, and some have simple counting systems with only a few numerical terms, such as
In some domains, following Weber’s law makes a certain amount of sense: a storehouse of an extra 2 kilos of