Hardcover edition | |
| Author | Daniel Kahneman |
|---|---|
| Language | English language |
| Subject | Psychology |
| Genre | Non-fiction |
| Publisher | Farrar, Straus and Giroux |
Publication date | 2011 |
| Publication place | United States |
| Media type | Print (hardcover,paperback), audio |
| Pages | 498 pages |
| ISBN | 978-0374275631 |
| OCLC | 706020998 |
Thinking, Fast and Slow is a 2011 popular science book by the Israeli-American psychologistDaniel Kahneman. Its main thesis is a differentiation between two modes ofthought: "System 1" is fast,instinctive andemotional; "System 2" is slower, moredeliberative, and morelogical.
The book delineates rational and non-rational motivations or triggers associated with each type of thinking process, and how they complement each other, starting with Kahneman's own research onloss aversion. Fromframing choices to people's tendency to replace a difficult question with one that is easy to answer, the book summarizes several decades of research to suggest that people have too much confidence in human judgement.[1] Kahneman performed his own research, often in collaboration with the psychologistAmos Tversky, which enriched his experience to write the book.[2][3] It covers different phases of his career: his early work concerningcognitive biases, his work onprospect theory andhappiness, and with theIsrael Defense Forces.
Jason Zweig, a columnist atThe Wall Street Journal, helped write and research the book over two years.[4][5] The book was aNew York Times bestseller[6] and was the 2012 winner of the AmericanNational Academies Communication Award for best creative work that helps the public understanding of topics inbehavioral science, engineering and medicine.[7] The integrity of somepriming studies cited in the book has been called into question in the midst of the psychologicalreplication crisis.[8]
In the book's first section, Kahneman describes two different ways the brain forms thoughts:
Kahneman describes a number of experiments which purport to examine the differences between these two thought systems and how they arrive at different results even given the same inputs. Terms and concepts include coherence, attention, laziness, association, jumping to conclusions, WYSIATI (What you see is all there is), and how one forms judgments. The System 1 vs. System 2 debate includes the reasoning or lack thereof for human decision making, with big implications for many areas including law and market research.[9]
The second section offers explanations for why humans struggle to think statistically. It begins by documenting a variety of situations in which we either arrive at binary decisions or fail to associate precisely reasonable probabilities with outcomes. Kahneman explains this phenomenon using the theory ofheuristics. Kahneman and Tversky originally discussed this topic in their 1974 article titled Judgment Under Uncertainty: Heuristics and Biases.[10]
Kahneman uses heuristics to assert that System 1 thinking involves associating new information with existing patterns, or thoughts, rather than creating new patterns for each new experience. For example, a child who has only seen shapes with straight edges might perceive an octagon when first viewing a circle. As a legal metaphor, a judge limited to heuristic thinking would only be able to think of similar historical cases when presented with a new dispute, rather than considering the unique aspects of that case. In addition to offering an explanation for the statistical problem, the theory also offers an explanation for human biases.
The "anchoring effect" names a tendency to be influenced by irrelevant numbers. Shown greater/lesser numbers, experimental subjects gave greater/lesser responses.[2] As an example, most people, when asked whetherMahatma Gandhi was more than 114 years old when he died, will provide a much greater estimate of his age at death than others who were asked whether Gandhi was more or less than 35 years old. Experiments show that people's behavior is influenced, much more than they are aware, by irrelevant information.
The availability heuristic is a mental shortcut that occurs when people make judgments about the probability of events on the basis of how easy it is to think of examples. The availability heuristic operates on the notion that, "if you can think of it, it must be important". The availability of consequences associated with an action is related positively to perceptions of the magnitude of the consequences of that action. In other words, the easier it is to recall the consequences of something, the greater we perceive these consequences to be. Sometimes, this heuristic is beneficial, but the frequencies at which events come to mind are usually not accurate representations of the probabilities of such events in real life.[11][12]
System 1 is prone to substituting a simpler question for a difficult one. In what Kahneman terms their "best-known and most controversial" experiment, "theLinda problem", subjects were told about an imaginary Linda, young, single, outspoken, and intelligent, who, as a student, was very concerned with discrimination and social justice. They asked whether it was more probable that Linda is a bank teller or that she is a bank teller and an active feminist. The overwhelming response was that "feminist bank teller" was more likely than "bank teller", violating the laws ofprobability. (All feminist bank tellers are bank tellers, so the former can't be more likely). In this case System 1 substituted the easier question, "Is Linda a feminist?", neglecting the occupation qualifier. An alternative interpretation is that the subjects added an unstated culturalimplicature to the effect that the other answer implied anexclusive or, that Linda was not a feminist.[2]
Kahneman writes of a "pervasiveoptimistic bias", which "may well be the most significant of the cognitive biases." This bias generates theillusion of control: the illusion that we have substantial control of our lives.
A natural experiment reveals the prevalence of one kind of unwarranted optimism. Theplanning fallacy is the tendency to overestimate benefits and underestimate costs, impelling people to begin risky projects. In 2002, American kitchen remodeling was expected on average to cost $18,658, but actually cost $38,769.[2]
To explainoverconfidence, Kahneman introduces the concept he termsWhat You See Is All There Is (WYSIATI). This theory states that when the mind makes decisions, it deals primarily withKnown Knowns, phenomena it has observed already. It rarely considersKnown Unknowns, phenomena that it knows to be relevant but about which it does not have information. Finally it appears oblivious to the possibility ofUnknown Unknowns, unknown phenomena of unknown relevance.
He explains that humans fail to take into account complexity and that their understanding of the world consists of a small and necessarily un-representative set of observations. Furthermore, the mind generally does not account for the role of chance and therefore falsely assumes that a future event will be similar to a past event.
Framing is the context in which choices are presented. Experiment: subjects were asked whether they would opt for surgery if the "survival" rate is 90 percent, while others were told that the mortality rate is 10 percent. The first framing increased acceptance, even though the situation was no different.[13]
Rather than consider the odds that an incremental investment would produce a positive return, people tend to "throw good money after bad" and continue investing in projects with poor prospects that have already consumed significant resources. In part this is to avoid feelings of regret.[13]
This part (part III, sections 19–24) of the book is dedicated to the undue confidence in what the mind believes it knows. It suggests that people often overestimate how much they understand about the world and underestimate the role of chance in particular. This is related to the excessive certainty of hindsight, when an event seems to be understood after it has occurred or developed. Kahneman's opinions concerning overconfidence are influenced by the authorNassim Nicholas Taleb.[14]
In this section Kahneman returns to economics and expands his seminal work on Prospect Theory. He discusses the tendency for problems to be addressed in isolation and how, when other reference points are considered, the choice of that reference point (called a frame) has a disproportionate effect on the outcome. This section also offers advice on how some of the shortcomings of System 1 thinking can be avoided.
Kahneman developed prospect theory, the basis for his Nobel prize, to account for experimental errors he noticed inDaniel Bernoulli's traditionalutility theory.[15] According to Kahneman, Utility Theory makes logical assumptions of economic rationality that do not represent people's actual choices, and does not take into accountcognitive biases.
One example is that people are loss-averse: they are more likely to act to avert a loss than to achieve a gain. Another example is that the value people place on a change in probability (e.g., of winning something) depends on the reference point: people seem to place greater value on a change from 0% to 10% (going from impossibility to possibility) than from, say, 45% to 55%, and they place the greatest value of all on a change from 90% to 100% (going from possibility to certainty). This occurs despite the fact that by traditional utility theory all three changes give the same increase in utility. Consistent with loss-aversion, the order of the first and third of those is reversed when the event is presented as losing rather than winning something: there, the greatest value is placed on eliminating the probability of a loss to 0.
After the book's publication, theJournal of Economic Literature published a discussion of its parts concerning prospect theory,[16] as well as an analysis of the four fundamental factors on which it is based.[17]
The fifth part of the book describes recent evidence which introduces a distinction between two selves, the 'experiencing self' and 'remembering self'.[18] Kahneman proposed an alternative measure that assessed pleasure or pain sampled from moment to moment, and then summed over time. Kahneman termed this "experienced" well-being and attached it to a separate "self". He distinguished this from the "remembered" well-being that the polls had attempted to measure. He found that these two measures of happiness diverged.[19]
The author's significant discovery was that the remembering self does not care about the duration of a pleasant or unpleasant experience. Instead, it retrospectively rates an experience by the maximum or minimum of the experience, and by the way it ends. The remembering self dominated the patient's ultimate conclusion.
"Odd as it may seem," Kahneman writes, "I am my remembering self, and the experiencing self, who does my living, is like a stranger to me."[3]
Kahneman first began the study of well-being in the 1990s. At the time mosthappiness research relied on polls aboutlife satisfaction. Having previously studied unreliable memories, the author was doubtful that life satisfaction was a good indicator of happiness. He designed a question that emphasized instead the well-being of the experiencing self. The author proposed that "Helen was happy in the month of March" if she spent most of her time engaged in activities that she would rather continue than stop, little time in situations that she wished to escape, and not too much time in a neutral state that wouldn't prefer continuing or stopping the activity either way.
Kahneman suggests that emphasizing a life event such as a marriage or a new car can provide a distorted illusion of its true value. This "focusing illusion" revisits earlier ideas of substituting difficult questions and WYSIATI.
As of 2012 the book had sold over one million copies.[24] On the year of its publication, it was on theNew York Times Bestseller List.[6] The book was reviewed in media including theHuffington Post,[25]The Guardian,[26]The New York Times,[2]The Financial Times,[27]The Independent,[28]Bloomberg[13] andThe New York Review of Books.[29][further explanation needed]
The book was also widely reviewed in academic journals, including theJournal of Economic Literature,[16]American Journal of Education,[30]The American Journal of Psychology,[31]Planning Theory,[32]The American Economist,[33]The Journal of Risk and Insurance,[34]The Michigan Law Review,[35]American Scientist,[36]Contemporary Sociology,[37]Science,[38]Contexts,[39]The Wilson Quarterly,[40]Technical Communication,[41]The University of Toronto Law Journal,[42]A Review of General Semantics[43] andScientific American Mind.[44] The book was also reviewed in the monthly magazineObserver, published by theAssociation for Psychological Science.[45][further explanation needed]
The book has achieved a large following amongbaseball scouts and baseball executives. The ways of thinking described in the book are believed to help scouts, who have to make major judgements off little information and can easily fall into prescriptive yet inaccurate patterns of analysis.[46]
The last chapter ofPaul Bloom'sAgainst Empathy discusses concepts also touched in Daniel Kahneman's book,Thinking, Fast and Slow, that suggest people make a series of rational and irrational decisions.[47][47]: 214 He criticizes the argument that "regardless of reason's virtues, we just aren't any good at it." His point is that people are not as "stupid as scholars think they are."[47]: 216 He explains that people are rational because they make thoughtful decisions in their everyday lives. For example, when someone has to make a big life decision they critically assess the outcomes, consequences, and alternative options.[47]: 230
The authorNicholas Taleb has equated the book's importance to that ofAdam Smith'sThe Wealth of Nations andSigmund Freud'sThe Interpretation of Dreams.[48]
A paper published byFrontiers in Artificial Intelligence discusses a new system: System 3, which represents "socially conditioned behavior influenced by societal norms and self-awareness."[49]
The book has been heavily criticized for relying on shoddy and non-reproducible studies. It was discovered many prominent research findings were difficult or impossible for others to replicate, and thus the original findings were called into question. An analysis[50] of the studies cited in chapter 4, "The Associative Machine", found that theirreplicability index (R-index)[51] is 14, indicating essentially low to no reliability. Kahneman himself responded to the study in blog comments and acknowledged the chapter's shortcomings: "I placed too much faith inunderpowered studies."[52] Others have noted the irony in the fact that Kahneman made a mistake in judgment similar to the ones he studied.[53]
A later analysis[54] made a bolder claim that, despite Kahneman's previous contributions to the field of decision making, most of the book's ideas are based on 'scientific literature with shaky foundations'. A general lack of replication in the empirical studies cited in the book was given as a justification.
{{cite journal}}: CS1 maint: bot: original URL status unknown (link)