SlideShare a Scribd company logo
1 of 89
Jeran Binning DAU San Diego
   http://www.nytimes.com/pages/books/review/index.html

          Copyright 2011 The New York Times Company




                Illustration by David Plunkert
   Expertise can be learnt by prolonged exposure to
    situations that are “sufficiently regular to be
    predictable”, and in which the expert gets quick and
    decisive feedback on whether he did the right or the
    wrong thing.
   Experts can thus train their unconscious “pattern
    recognition” mechanism to produce the right answer
    quickly.
   So this certainly applies to chess, and it certainly does
    not apply to predicting the course of Middle East
    politics.
   Kahneman’s career has had three major phases
   In the first, he and Tversky did a series of
    ingenious experiments that revealed twenty
    or so “cognitive biases” — unconscious errors
    of reasoning that distort our judgment of the
    world.

   Typical of these is the “anchoring effect”:
    our tendency to be influenced by irrelevant
    numbers that we happen to be exposed to.
   Anchoring and adjustment: People who have to
    make judgments under uncertainty use this heuristic
    by starting with a certain reference point (anchor)
    and then adjust it insufficiently to reach a final
    conclusion.

   Example: If you have to judge another persons
    productivity, the anchor for your final (adjusted)
    judgment may be your own level of productivity.
    Depending on your own level of productivity you
    might therefore underestimate or overestimate the
    productivity of this person.
   In the second phase, Kahneman and Tversky
    showed that people making decisions under
    uncertain conditions do not behave in the
    way that economic models have traditionally
    assumed; they do not “maximize utility.”
   The two then developed an alternative
    account of decision making, one more
    faithful to human psychology, which they
    called “prospect theory.” (It was for this
    achievement that Kahneman was awarded
    the Nobel.)
Prospect theory was developed by Daniel Kahneman and Amos Tversky in 1979
   Prospect theory was developed by Daniel Kahneman and Amos Tversky in 1979 as a
    psychologically realistic alternative to expected utility theory. It allows one to describe how people make
    choices in situations where they have to decide between alternatives that involve risk, e.g. in financial
    decisions. Starting from empirical evidence, the theory describes how individuals evaluate potential losses
    and gains. In the original formulation the term prospect referred to a lottery.
   The theory describes such decision processes as consisting of two stages, editing and evaluation. In the
    first, possible outcomes of the decision are ordered following some heuristic. In particular, people decide
    which outcomes they see as basically identical and they set a reference point and consider lower outcomes
    as losses and larger as gains. In the following evaluation phase, people behave as if they would compute a
    value (utility), based on the potential outcomes and their respective probabilities, and then choose the
    alternative having a higher utility.
   The formula that Kahneman and Tversky assume for the evaluation phase is (in its simplest form) given by
    where are the potential outcomes and their respective probabilities. v is a so-called value function that
    assigns a value to an outcome. The value function (sketched in the Figure) which passes through the
    reference point is s-shaped and, as its asymmetry implies, given the same variation in absolute
    value, there is a bigger impact of losses than of gains (loss aversion). In contrast to Expected Utility
    Theory, it measures losses and gains, but not absolute wealth. The function w is called a
    probability weighting function and expresses that people tend to
    overreact to small probability events, but underreact to medium
    and large probabilities
   To see how Prospect Theory (PT) can be applied in an example, consider a decision about buying an
    insurance policy. Let us assume the probability of the insured risk is 1%, the potential loss is $1000 and the
    premium is $15. If we apply PT, we first need to set a reference point. This could be, e.g., the current
    wealth, or the worst case (losing $1000). If we set the frame to the current wealth, the decision would be to
    either pay $15 for sure (which gives the PT-utility of v( − 15)) or a lottery with outcomes $0 (probability
    99%) or $-1000 (probability 1%) which yields the PT-utility of . These expressions can be computed
    numerically. For typical value and weighting functions, the former expression could be larger due to the
    convexity of v in losses, and hence the insurance looks unattractive. If we set the frame to $-1000, both
    alternatives are set in gains. The concavity of the value function in gains can then lead to a preference
    for buying the insurance.
   We see in this example that a strong overweighting of small probabilities can also undo the effect of the
    convexity of v in losses: the potential outcome of losing $1000 is overweighted.
   The interplay of overweighting of small probabilities and concavity-convexity of the value function leads to
    the so-called four-fold pattern of risk attitudes: risk-averse behavior in gains involving moderate
    probabilities and of small probability losses; risk-seeking behavior in losses involving moderate
    probabilities and of small probability gains. This is an explanation for the fact that
    people, e.g., simultaneously buy lottery tickets and insurances, but
    still invest money conservatively.
   An important paper in the development of the behavioral finance and
    economics fields was written by Kahneman and Tversky in 1979. This
    paper, 'Prospect theory: An Analysis of Decision Under Risk', used
    cognitive psychological techniques to explain a number of documented
    divergences of economic decision making from neo-classical theory.
   Over time many other psychological effects have been incorporated into
    behavioral finance, such as overconfidence and the effects of limited
    attention.
   Further milestones in the development of the field include a well
    attended and diverse conference at the University of Chicago,[4] a
    special 1997 edition of the Quarterly Journal of Economics ('In Memory of
    Amos Tversky') devoted to the topic of behavioral economics and the
    award of the Nobel prize to Daniel Kahneman in 2002 "for having
    integrated insights from psychological research into economic science,
    especially concerning human judgment and decision-making under
    uncertainty"
   In the third phase of his career, mainly after
    the death of Tversky, Kahneman has delved
    into “hedonic psychology”: the science of
    happiness, its nature and its causes. His
    findings in this area have proved disquieting
    — and not just because one of the key
    experiments involved a deliberately
    prolonged colonoscopy.
   Kahneman never grapples philosophically with the
    nature of rationality. He does, however, supply a
    fascinating account of what might be taken to be its
    goal: happiness.

   What does it mean to be happy? When Kahneman
    first took up this question, in the mid 1990s, most
    happiness research relied on asking people how
    satisfied they were with their life on the whole.

   But such retrospective assessments depend on
    memory, which is notoriously unreliable.
   What if, instead, a person’s actual experience of pleasure or pain
    could be sampled from moment to moment, and then summed up
    over time? Kahneman calls this “experienced” well-being, as
    opposed to the “remembered” well-being that researchers had
    relied upon.
   And he found that these two measures of happiness diverge in
    surprising ways. What makes the “experiencing self” happy is not
    the same as what makes the “remembering self” happy.
   In particular, the remembering self does not care about duration —
    how long a pleasant or unpleasant experience lasts. Rather, it
    retrospectively rates an experience by the peak level of pain or
    pleasure in the course of the experience, and by the way the
    experience ends.
   Clearly, much remains to be done in hedonic psychology. But
    Kahneman’s conceptual innovations have laid the foundation for many of
    the empirical findings he reports in this book:
   that while French mothers spend less time with their children than
    American mothers, they enjoy it more;
   that headaches are hedonically harder on the poor;
   that women who live alone seem to enjoy the same level of well-being as
    women who live with a mate; and
   that a household income of about $75,000 in high-cost areas of the
    country is sufficient to maximize happiness.
   Policy makers interested in lowering the misery index of society will find
    much to ponder here.
   System 2, in Kahneman’s scheme, is our
    slow, deliberate, analytical and consciously effortful mode of
    reasoning about the world.
   System 1, by contrast, is our fast, automatic, intuitive and largely
    unconscious mode.
   It is System 1 that detects hostility in a voice and effortlessly
    completes the phrase “bread and. . . . ”
   It is System 2 that swings into action when we have to fill out a tax
    form or park a car in a narrow space. (As Kahneman and others
    have found, there is an easy way to tell how engaged a person’s
    System 2 is during a task: just look into his or her eyes and note
    how dilated the pupils are.)
   More generally, System 1 uses association and
    metaphor to produce a quick and dirty draft of
    reality, which System 2 draws on to arrive at explicit
    beliefs and reasoned choices.

   System 1 proposes, System 2 disposes.

   So System 2 would seem to be the boss, right?

   In principle, yes. But System 2, in addition to being
    more deliberate and rational, is also lazy. And it tires
    easily. (The vogue term for this is “ego depletion.”)
   “Although System 2 believes itself to be
    where the action is,” Kahneman writes, “the
    automatic System 1 is the hero of this book.”
    System 2 is especially quiescent, it
    seems, when your mood is a happy one.
    Memory also holds the vast repertory of skills
    we have acquired in a lifetime of
    practice, which automatically produce
    adequate solutions to challenges as they
    arise…. The acquisition of skills requires a
    regular environment, an adequate
    opportunity to practice, and rapid and
    unequivocal feedback about the correctness
    of thoughts and actions.
   When these conditions are fulfilled, skill
    eventually develops, and the intuitive
    judgments and choices that quickly come to
    mind will mostly be accurate. All of this is the
    work of System 1, which means it occurs
    automatically and fast. A marker of skilled
    performance is the ability to deal with vast
    amounts of information swiftly and
    efficiently. Pp 416 conclusions.
   The way to block errors
    that originate in system
    one us a simple in
    principle: recognize
    signs that you are in a
    cognitive minefield, slow
    down and ask for
    reinforcement from
    System 2.
   What can be done about biases? How can we
    improve judgments and decisions, both our
    own and those of the institutions that we
    serve and serve us?
   The short answer is that little can be achieved
    without considerable investment of effort. As
    I know from experience System 1 is not
    readily educable. Except for some effects
    that I attribute mostly to my age, my intuitive
    thinking is just as prone to
    overconfidence, extreme predications, and
    the planning fallacy as it was before I made a
    study of these issues.
   I have improved only in my ability to
    recognize situations in which errors are likely;
    “This number will be an anchor…” “The
    decision could be changes if the question was
    reframed…” And I have made much more
    progress in recognizing the errors of others
    than my own.”
   By the time I got to the end of “Thinking, Fast and
    Slow,” my skeptical frown had long since given way to
    a grin of intellectual satisfaction. Appraising the book
    by the peak-end rule, I overconfidently urge everyone
    to buy and read it. But for those who are merely
    interested in Kahneman’s takeaway on the Malcolm
    Gladwell question it is this:

   If you’ve had 10,000 hours of training in a
    predictable, rapid-feedback environment —
    chess, firefighting, anesthesiology — then blink. In all
    other cases, think.
   Projects fail at a spectacular rate. One reason is that too many
    people are reluctant to speak up about their reservations during
    the all-important planning phase. By making it safe for dissenters
    who are knowledgeable about the undertaking and worried about
    its weaknesses to speak up, you can improve a project’s chances of
    success.

   Research conducted in 1989 by Deborah J. Mitchell, of the
    Wharton School; Jay Russo, of Cornell; and Nancy Pennington, of
    the University of Colorado, found that prospective hindsight—
    imagining that an event has already occurred—increases the
    ability to correctly identify reasons for future outcomes by 30%.
    We have used prospective hindsight to devise a method called a
    premortem, which helps project teams identify risks at the outset.
   Judgment Under Uncertainty:
    Heuristics and Biases. Amos Tversky
    and Daniel Kahneman

   Science, Volume 185, 1974

   Research for DARPA N00014-73C-0438
    monitored by ONR and Research and
    Development Authority of Hebrew
    University, Jerusalem, Israel.
   The same goes for all of us, almost all the time. We think we're smart; we're
    confident we won't be unconsciously swayed by the high list price of a house.
    We're wrong. (Kahneman admits his own inability to counter some of these
    effects.) We're also hopelessly subject to the "focusing illusion", which can be
    conveyed in one sentence: "Nothing in life is as important as you think it is when
    you're thinking about it."
   Whatever we focus on, it bulges in the heat of our attention until we assume its
    role in our life as a whole is greater than it is.
   Another systematic error involves "duration neglect" and the "peak-end rule".
    Looking back on our experience of pain, we prefer a larger, longer amount to a
    shorter, smaller amount, just so long as the closing stages of the greater pain
    were easier to bear than the closing stages of the lesser one.

   Galen Strawson
   guardian.co.uk, Tuesday 13 December 2011 06.56 EST
   System 2 is slothful, and tires easily (a process called "ego depletion") –
    so it usually accepts what System 1 tells it. It's often right to do
    so, because System 1 is for the most part pretty good at what it does; it's
    highly sensitive to subtle environmental cues, signs of danger, and so on.
   It kept our remote ancestors alive. Système 1 a ses raisons que Système 2
    ne connaît point, as Pascal might have said.
   It does, however, pay a high price for speed. It loves to simplify, to assume
    WYSIATI ("what you see is all there is"), even as it gossips and embroiders
    and confabulates.
   It's hopelessly bad at the kind of statistical thinking often required for good
    decisions, it jumps wildly to conclusions and it's subject to a fantastic suite
    of irrational biases and interference effects (the halo effect, the "Florida
    effect", framing effects, anchoring effects, the confirmation bias, outcome
    bias, hindsight bias, availability bias, the focusing illusion, and so on).
   Behavioralists, Kahneman included, have been cataloging people’s
    systematic mistakes and nonlogical patterns for years. A few of
    the examples he cites:
   1. Framing. Test subjects are more likely to opt for surgery if told
    that the “survival” rate is 90 percent, rather than that the mortality
    rate is 10 percent.
   2. The sunk-cost fallacy. People seek to avoid feelings of regret;
    thus, they invest more money and time in a project with dubious
    results rather than give it up and admit they were wrong.
    3. Loss aversion. In experiments, most subjects would prefer to
    receive a sure $46 than have a 50 percent chance of making $100.
   Kahneman is perhaps least persuasive in his treatment of the
    business world. Noting that even top performers in business—also
    sports—tend eventually to revert to the mean, he attributes
    success largely to luck.
   This confuses events that may not be predictable with those that
    are determined by chance.
   A high-achieving retail store, to cite one of his examples, is not
    lucky—it is well-situated. And if its sales later decline, that is not
    necessarily a sign that its prior success was random.
   Business has a self-correcting cycle that fosters mean reversion.
    Success attracts competitors.
   Mr. Kahneman stresses that he is just as susceptible as the rest of us to
    the cognitive illusions he has discovered. He tries to recognize situations
    when mistakes are especially likely to occur—such as when he is starting
    a big project or making a forecast—and then act to rethink his System 1
    inclinations.
   The tendency to underestimate the costs of future projects, he notes, is
    susceptible to taking an "outside view": looking at your own project as an
    outsider would.
   To avoid overconfidence, Mr. Kahneman recommends an exercise called
    the "premortem," developed by the psychologist Gary Klein: Before
    finalizing a decision, imagine that, a year after it has been made, it has
    turned out horribly, then write a history of how it went wrong and why.
   By CHRISTOPHER F. CHABRIS         a psychology professor at Union College and a co-author of "The Invisible Gorilla: And Other Ways Our Intuitions Deceive Us.

   http://online.wsj.com/home-page
   A transitional moment linking the positive and the negative aspects of thinking
    fast illustrates why the author’s personality – and thus the book – is so engaging.
    Kahneman regards even the experts as prone to the mistakes of System 1 listed
    above, and cheerfully admits that he is no exception. But he wants to know
    whether this view can be reconciled with cases such as that of the firefighting
    captain. So he engages one of his vehement critics on this issue and they debate
    their way to a joint paper.
   Their answer is that expertise can be learnt by prolonged exposure to situations
    that are “sufficiently regular to be predictable”, and in which the expert gets
    quick and decisive feedback on whether he did the right or the wrong thing.
   Experts can thus train their unconscious “pattern recognition” mechanism to
    produce the right answer quickly.
   So this certainly applies to chess, and it certainly does not apply to predicting the
    course of Middle East politics.
   One of Kahneman and Tversky’s most famous ideas is
    what they call prospect theory: our inclination to fear
    possible losses more than we value possible gains.
   Would you take a bet on a one-time coin flip that paid
    $200 if you won but cost you $150 if you lost?
   Most people wouldn’t, though it’s tilted in your favor.
   Pro golfers tend to make a higher proportion of their putts
    when they’re trying to avoid a bogey (which would result
    in losing a stroke) than when they have a chance for a
    birdie (for the possible gain of a stroke);
    here, too, avoiding a failure is more crucial than achieving
    a triumph.
   If the brain is a “a machine for jumping to
    conclusions,” as Kahneman writes, it’s System 1
    that yells “Geronimo!” Autopilot thinking
    explains the popular opinion that tornadoes kill
    more people than asthma, although in fact
    asthma kills 20 times as many:

   We fixate on scenes from TV showing homes
    turned to matchsticks and overestimate the
    representativeness of such scenes.
   The first bit of mental machinery, which Kahneman blandly calls System One, works 24/7 to keep
    us out of trouble, while alerting us to fleeting opportunities. Appropriate for a species that is both
    predator and prey, System One lives in a world of snap judgments, extensible metaphors, ill-
    informed biases, and loosely constructed rules of thumb. We sometimes call this decision making
    apparatus intuition. Man’s intuition is sophisticated enough that it has helped us thrive across a
    variety of ever changing environments.
   Despite its utility, System One is often wrong, especially if numbers are involved. For a trivial
    example, answer quickly: If the sum of the cost of a ball and bat is $1.10 and the bat cost a dollar
    more than the ball, what does the ball cost?
   Your System One answer (most likely wrong) is good enough to avoid mistaking a hungry lion for
    a tasty chicken. But it’s not good enough to build an airplane or design an effective income tax
    code. (The answer is a nickel, not a dime.).
   System Two is associated with enumeration, computation, objective analysis, and complex chains
    of logic. It is our rational brain. Kahneman’s work shows that even scientists like himself use
    System Two very sparingly, calling on it only when System One asks for help. In addition, in order
    to function at the highest levels, System Two requires training, discipline, concentration, skeptical
    and impartial evaluation of purported facts, and the ruthless elimination of contradictions.
It afflicts us all. Because confidence in our own
judgments is part of being human.
By DANIEL KAHNEMAN
Published: October 19, 2011
   Decades later, I can see many of the central themes of my thinking
    about judgment in that old experience. One of these themes is that
    people who face a difficult question often answer an easier one instead,
    without realizing it. We were required to predict a soldier’s performance
    in officer training and in combat, but we did so by evaluating his behavior
    over one hour in an artificial situation. This was a perfect instance of a
    general rule that I call WYSIATI, “What you see is all there is.” We had
    made up a story from the little we knew but had no way to allow for what
    we did not know about the individual’s future, which was almost
    everything that would actually matter. When you know as little as we did,
    you should not make extreme predictions like “He will be a star.” The
    stars we saw on the obstacle field were most likely accidental flickers, in
    which a coincidence of random events — like who was near the wall —
    largely determined who became a leader. Other events — some of them
    also random — would determine later success in training and combat.
   We often interact with professionals who exercise their
    judgment with evident confidence, sometimes priding
    themselves on the power of their intuition. In a world rife
    with illusions of validity and skill, can we trust them? How
    do we distinguish the justified confidence of experts from
    the sincere overconfidence of professionals who do not
    know they are out of their depth? We can believe an expert
    who admits uncertainty but cannot take expressions of
    high confidence at face value. As I first learned on the
    obstacle field, people come up with coherent stories and
    confident predictions even when they know little or
    nothing. Overconfidence arises because people are often
    blind to their own blindness.
   Some cognitive biases, of course, are flagrantly
    exhibited even in the most natural of settings.
    Take what Kahneman calls the “planning
    fallacy”: our tendency to overestimate benefits
    and underestimate costs, and hence foolishly to
    take on risky projects.

   In 2002, Americans remodeling their
    kitchens, for example, expected the job to cost
    $18,658 on average, but they ended up paying
    $38,769.
   The planning fallacy is “only one of the manifestations of a
    pervasive optimistic bias,” Kahneman writes, which “may
    well be the most significant of the cognitive biases.”
    Now, in one sense, a bias toward optimism is obviously
    bad, since it generates false beliefs — like the belief that
    we are in control, and not the playthings of luck.
   But without this “illusion of control,” would we even be
    able to get out of bed in the morning?
   Optimists are more psychologically resilient, have
    stronger immune systems, and live longer on average than
    their more reality-based counterparts.
   Biases in the evaluation of compound events
    are particularly significant in the context of
    planning. The successful completion of an
    undertaking, such as the development of a
    new product, typically has a conjunctive
    character: for the undertaking to
    succeed, each of a series of events must
    occur. Even when each of these events is very
    likely, the overall probability of success can
    be quite low if the number of events is large.
   In a series of experiments, Ellen Langer (1975) demonstrated first the
    prevalence of the illusion of control and second, that people were more
    likely to behave as if they could exercise control in a chance situation
    where “skill cues” were present.
   By skill cues, Langer meant properties of the situation more normally
    associated with the exercise of skill, in particular the exercise of
    choice, competition, familiarity with the stimulus and involvement in
    decisions.
   One simple form of this fallacy is found in casinos: when rolling dice in
    craps, it has been shown that people tend to throw harder for high
    numbers and softer for low numbers.
   Under some circumstances, experimental subjects have been induced to
    believe that they could affect the outcome of a purely random coin toss.
    Subjects who guessed a series of coin tosses more successfully began to
    believe that they were actually better guessers, and believed that their
    guessing performance would be less accurate if they were distracted.
   Drawing on a vast body of research, Lears ranges
    through the entire sweep of American history as he
    uncovers the hidden influence of risk
    taking, conjuring, soothsaying, and sheer dumb luck on
    our culture, politics, social lives, and economy.




T.J. Jackson Lears “Something for Nothing” (2003)
   Moreover, as Kahneman notes, exaggerated
    optimism serves to protect both individuals
    and organizations from the paralyzing effects
    of another bias, “loss aversion”: our tendency
    to fear losses more than we value gains.

   It was exaggerated optimism that John
    Maynard Keynes had in mind when he talked
    of the “animal spirits” that drive capitalism.
   "losses loom larger than corresponding gains”
 "In prospect theory, loss aversion refers to the tendency for people to strongly
  prefer avoiding losses than acquiring gains. Some studies suggest that losses are
  as much as twice as psychologically powerful as gains. Loss aversion was first
  convincingly demonstrated by Amos Tversky and Daniel Kahneman.”

 "The principle of loss aversion was first introduced by Kahneman and Tversky
  (1979)"
   Tversky and Kahneman (1991) "The central assumption of the theory is that
    losses and disadvantages have greater impact on preferences than gains and
    advantages.”
   "Numerous studies have shown that people feel losses more deeply than gains of
    the same value (Kahneman and Tversky 1979, Tversky and Kahneman
    1991)."
Goldberg and von Nitzsch (1999) pages 97-98
   "Both the status quo bias and the endowment effect are part of a more general
    issue known as loss aversion." (Montier 2007, p. 32)
   the inclination to see events that have occurred as more predictable than they in fact were before
    they took place. Hindsight bias has been demonstrated experimentally in a variety of
    settings, including politics, games and medicine.
   In psychological experiments of hindsight bias, subjects also tend to remember their predictions of
    future events as having been stronger than they actually were, in those cases where those predictions
    turn out correct.
   Prophecy that is recorded after the fact is an example of hindsight bias, given its own rubric, as
    vaticinium ex eventu. foretelling after the event
   One explanation of the bias is the availability heuristic: the event that did occur is more salient in
    one's mind than the possible outcomes that did not.
   It has been shown that examining possible alternatives may reduce the effects of this bias.
   The framing of alternatives also affects decisions.
   For example, when people (including doctors) who are considering a risky
    medical procedure are told that 90 percent survive five years, they are far
    more likely to accept the procedure than when they are told that 10 percent
    do not survive five years.

   Because framing affects people's behavior, providing more information
    cannot remedy matters, unless the information is presented in a fully
    neutral fashion.


   In some cases, additional information only increases people's anxiety and
    confusion, thereby reducing their welfare.
   A heuristic (hyu̇-ris-tik) is a method to help solve a
    problem, commonly informal. It is particularly used for
    a method that often rapidly leads to a solution that is
    usually reasonably close to the best possible answer.
   Heuristics are "rules of thumb” educated
    guesses, intuitive judgments or simply common sense.
   In more precise terms, heuristics stand for strategies
    using readily accessible, though loosely
    applicable, information to control problem-solving in
    human beings and machines.
“There is always a well-known solution to every human
    problem – neat, plausible, and wrong.”

        H. L. Mencken, Prejudices: Second Series, 1920




Tailors' Rule of Thumb. This is the fictional rule described by Jonathan Swift in his satirical novel Gulliver's Travels:

          Then they measured my right Thumb, and desired
          no more; for by a mathematical Computation, that
          twice round the Thumb is once around the
          Wrist, and so on to the Neck and Waist, and by the
          help of my old Shirt, which I displayed on the
          Ground before them for a Pattern, they fitted me
          exactly."
Rules of Thumb



   People rely on a limited number of heuristic
    (essentially rules of thumbs) principles
    which reduce the complex tasks of
    assessing probabilities and predicting
    values to simpler judgmental operations.
   In general, these heuristics are quite
    useful, but sometimes they lead to severe
    and systematic errors.
   Daniel Kahneman
   Financial - Rule of 72 A rule of thumb for exponential growth at a constant rate. An
    approximation of the "doubling time" formula used in population growth, which says divide 70 by
    the percent growth rate (the actual number is 69.3147181 from the natural logarithm of 2, if the
    percent growth is much much less than 1%). In terms of money, it is frequently easier to use 72
    (rather than 70) because it works better in the 4%-10% range where interest rates often lie.
    Therefore, divide 72 by the percent interest rate to determine the approximate amount of time to
    double your money in an investment. For example, at 8% interest, your money will double in
    approximately 9 years (72/8 = 9).
   Tailors' Rule of Thumb A simple approximation that was used by tailors to determine the
    wrist, neck, and waist circumferences of a person through one single measurement of the
    circumference of that person's thumb. The rule states, typically, that twice the circumference of a
    person's thumb is the circumference of their wrist, twice the circumference of the wrist is the
    circumference of the neck, and twice around the neck is the person's waist. For example, if the
    circumference of the thumb is 4 inches, then the wrist circumference is 8 inches, the neck is 16
    and the waist is 32. An interesting consequence of this is that — for those to whom the rule applies
    — this simple method can be used to determine if trousers will fit: the trousers are wrapped
    around the neck, and if the two ends barely touch, then they will fit. Any overlap or lack thereof
    corresponds to the trousers being too loose or tight, respectively.
   Marine Navigation A ship's captain should navigate to keep the ship more than a thumb's width
    from the shore, as shown on the nautical chart being used. Thus, with a coarse scale chart, that
    provides few details of near shore hazards such as rocks, a thumb's width would represent a great
    distance, and the ship would be steered far from shore; whereas on a fine scale chart, in which
    more detail is provided, a ship could be brought closer to shore.[1]

   Statistics The Statistical Rule of Thumb says that for most large data sets, 68% of data points will
    occur within one standard deviation from the mean, and 95% will occur within two standard
    deviations. Chebyshev's inequality is a more general rule along these same lines and applies to all
    data sets.
   In an effort that spanned several years, we attempted
    to answer one basic question:

   Under what conditions are the intuitions of
    professionals worthy of trust?

   We do not claim that the conclusions we reached are
    surprising (many were anticipated by
    Shanteau, 1992, Hogarth, 2001, and
    Myers, 2002, among others), but we believe that they
    add up to a coherent view of expert intuition, which is
    more than we expected to achieve when we began.
   The NDM approach, which focuses on the successes of expert intuition, grew out
    of early research on master chess players conducted by deGroot (1946/1978) and
    later by Chase and Simon (1973). DeGroot showed that chess grand masters were
    generally able to identify the most promising moves rapidly, while mediocre
    chess players often did not even consider the best moves.
   The chess grand masters mainly differed from weaker players in their unusual
    ability to appreciate the dynamics of complex positions and quickly judge a line
    of play as promising or fruitless.
   Chase and Simon (1973) described the performance of chess experts as a form of
    perceptual skill in which complex patterns are recognized. They estimated that
    chess masters acquire a repertoire of 50,000 to 100,000 immediately
    recognizable patterns, and that this repertoire enables them to identify a good
    move without having to calculate all possible contingencies.
   Strong players need a decade of serious play to assemble this large collection of
    basic patterns, but of course they achieve impressive levels of skill even earlier.
   On the basis of this work, Simon defined intuition as the recognition of patterns
    stored in memory.
   Kahneman read Meehl’s book in 1955 while serving in the Psychological Research
    Unit of the Israel Defense Forces, and the book helped him make sense of his
    own encounters with the difficulties of clinical judgment.
   One of Kahneman’s duties was to assess candidates for officer training, using
    field tests and other observations as well as a personal interview.
   Kahneman (2003) described the powerful sense of getting to know each
    candidate and the accompanying conviction that he could foretell how well the
    candidate would do in further training and eventually in combat.
   The subjective conviction of understanding each case in isolation was not
    diminished by the statistical feedback from officer training school, which
    indicated that the validity of the assessments was negligible.
   Kahneman coined the term illusion of validity for the unjustified sense of
    confidence that often comes with clinical judgment. His early experience with the
    fallibility of intuitive impressions could hardly be more different from Klein’s
    formative encounter with the successful decision making of fire-fighting ground
    commanders.
   Our starting point is that intuitive judgments
    can arise from genuine skill—the focus of the
    Naturalistic Decision Making (NDM)
    approach—but that they can also arise from
    inappropriate application of the heuristic
    processes on which students of the Heuristics
    Based tradition have focused.
   Skilled judges are often unaware of the cues
    that guide them, and individuals whose
    intuitions are not skilled are even less likely to
    know where their judgments come from.
   True experts, it is said, know when they don’t
    know.

   However, non-experts (whether or not they
    think they are) certainly do not know when
    they don’t know.

   Subjective confidence is therefore an
    unreliable indication of the validity of
    intuitive judgments and decisions.
   The determination of whether intuitive
    judgments can be trusted requires an
    examination of the environment in which the
    judgment is made and of the opportunity
    that the judge has had to learn the
    regularities of that environment.
   We describe task environments as “high-validity” if there
    are stable relationships between objectively identifiable
    cues and subsequent events or between cues and the
    outcomes of possible actions.
   Medicine and firefighting are practiced in environments of
    fairly high validity.
   In contrast, outcomes are effectively unpredictable in
    zero-validity environments.
   To a good approximation, predictions of the future value
    of individual stocks and long-term forecasts of political
    events are made in a zero-validity environment.
   Validity and uncertainty are not
    incompatible. Some environments are both
    highly valid and substantially uncertain.

   Poker and warfare are examples. The best
    moves in such situations reliably increase the
    potential for success.
   An environment of high validity is a necessary
    condition for the development of skilled intuitions.

   Other necessary conditions include adequate
    opportunities for learning the environment
    (prolonged practice and feedback that is both rapid
    and unequivocal).

   If an environment provides valid cues and good
    feedback, skill and expert intuition will eventually
    develop in individuals of sufficient talent.
   Although true skill cannot develop in irregular or
    unpredictable environments, individuals will
    sometimes make judgments and decisions that
    are successful by chance.
   These “lucky” individuals will be susceptible to
    an illusion of skill and to overconfidence (Arkes,
    2001).
   The financial industry is a rich source of
    examples.
   The situation that we have labeled fractionation of skill is
    another source of overconfidence. Professionals who have
    expertise in some tasks are sometimes called upon to
    make judgments in areas in which they have no real skill.
   (For example, financial analysts may be skilled at
    evaluating the likely commercial success of a firm, but this
    skill does not extend to the judgment of whether the stock
    of that firm is underpriced.)
   It is difficult both for the professionals and for those who
    observe them to determine the boundaries of their true
    expertise.
   We agree that the weak regularities available in
    low-validity situations can sometimes support
    the development of algorithms that do better
    than chance. These algorithms only achieve
    limited accuracy, but they outperform humans
    because of their advantage of consistency.
   However, the introduction of algorithms to
    replace human judgment is likely to evoke
    substantial resistance and sometimes has
    undesirable side effects.
The Drunkard’s Walk

   – Functional magnetic resonance imaging, for
     example, shows that risk and reward are assessed by
     parts of the dopaminergic system.



   – A brain-reward circuit important for motivational and
     emotional processes


• The images show, too, that the amygdala, which is
  also linked to our emotional state, especially
  fear, is activated when we make decisions
  couched in uncertainty.
The Drunkard’s Walk

• Fortune is fair in potentialities, she is not fair in outcomes.
•   pp 13




• When we look at extraordinary accomplishments in sports--
  -or elsewhere—we should keep in mind that
  extraordinary events can happen without extraordinary
  causes.

• Random events often look like nonrandom
  events, and in interpreting human affairs we must take
  care not to confuse the two.  pp20
Selection Bias
• This bias makes us miscompute the odds and wrongly ascribe skills. If you
  funded 1,000,000 unemployed people endowed with no more than the
  ability to say "buy" or "sell", odds are that you will break-even in the
  aggregate, minus transaction costs, but a few will hit the jackpot, simply
  because the base cohort is very large. It will be almost impossible not to
  have small Warren Buffets by luck alone. After the fact they will be very
  visible and will derive precise and well-sounding explanations about why
  they made it. It is difficult to argue with them; "nothing succeeds like
  success". All these retrospective explanations are pervasive, but there are
  scientific methods to correct for the bias. This has not filtered through to
                            researchers have
  the business world or the news media;
   evidence that professional fund managers are
   just no better than random and cost money to society (the
   total revenues from these transaction costs is in the hundreds of billion of
   dollars) but the public will remain convinced that "some" of these
   investors have skills.
False Discovery Rate
•   July 13, 2008
•   STRATEGIES
•   The Prescient Are Few
•   By MARK HLBERT
•   HOW many mutual fund managers can consistently pick stocks that outperform the broad stock market averages — as opposed to just being lucky now and then?
•   Countless studies have addressed this question, and have concluded that very few managers have the ability to beat the market over the long term. Nevertheless, researchers
    have been unable to agree on how small that minority really is, and on whether it makes sense for investors to try to beat the market by buying shares of actively managed
    mutual funds.
•   A new study builds on this research by applying a sensitive statistical test borrowed from outside the investment world. It comes to a rather sad conclusion: There was once a
    small number of fund managers with genuine market-beating abilities, as judged by having past performance so good that their records could not be attributed to luck alone.
    But virtually none remain today. Index funds are the only rational alternative for almost all mutual fund investors, according to the study’s findings.
•   The study, “False Discoveries in Mutual Fund Performance: Measuring Luck in Estimating Alphas,” has been circulating for over a year in academic circles. Its authors are
    Laurent Barras, a visiting researcher at Imperial College’s Tanaka Business School in London; Olivier Scaillet, a professor of financial econometrics at the University of Geneva
    and the Swiss Finance Institute; and Russ Wermers, a finance professor at the University of Maryland.
•   The statistical test featured in the study is known as the “False Discovery Rate,” and is used in fields as diverse as computational biology and astronomy. In effect, the method
    is designed to simultaneously avoid false positives and false negatives — in other words, conclusions that something is statistically significant when it is entirely random, and
    the reverse.
•   Both of those problems have plagued previous studies of mutual funds, Professor Wermers said. The researchers applied the method to a database of actively managed
    domestic equity mutual funds from the beginning of 1975 through 2006. To ensure that their results were not biased by excluding funds that have gone out of business over
    the years, they included both active and defunct funds. They excluded any fund with less than five years of performance history. All told, the database contained almost 2,100
    funds.
•   The researchers found a marked decline over the last two decades in the number of fund managers able to pass the False Discovery Rate test. If they had focused only on
    managers running funds in 1990 and their records through that year, for example, the researchers would have concluded that 14.4 percent of managers had genuine stock-
    picking ability. But when analyzing their entire fund sample, with records through 2006, this proportion was just 0.6 percent — statistically indistinguishable from
    zero, according to the researchers.
•   This doesn’t mean that no mutual funds have beaten the market in recent years, Professor Wermers said. Some have done so repeatedly over periods as short as a year or
    two. But, he added, “the number of funds that have beaten the market over their entire histories is so small that the False Discovery Rate test can’t eliminate the
    possibility that the few that did were merely false positives” — just lucky, in other words.
•   Professor Wermers says he was surprised by how rare stock-picking skill has become. He had “generally been positive about the existence of fund manager ability,” he said, but
    these new results have been a “real shocker.”
•   WHY the decline? Professor Wermers says he and his co-authors suspect various causes. One is high fees and expenses. The researchers’ tests found that, on a pre-expense
    basis, 9.6 percent of mutual fund managers in 2006 showed genuine market-beating ability — far higher than the 0.6 percent after expenses were taken into account.

    This suggests that one in 10 managers may still have market-
    beating ability. It’s just that they can’t come out ahead after all
    their funds’ fees and expenses are paid.
•   Another possible factor is that many skilled managers have gone to the hedge fund world. Yet a third potential reason is that the market has become more efficient, so it’s
    harder to identify undervalued or overvalued stocks. Whatever the causes, the investment implications of the study are the same: buy and hold an index fund benchmarked
    to the broad stock market.
•   Professor Wermers says his advice has evolved significantly as a result of this study. Until now, he says, he wouldn’t have tried to discourage a sophisticated investor from
    trying to pick a mutual fund that would outperform the market. Now, he says, “it seems almost hopeless.”

•   Mark Hulbert is editor of The Hulbert Financial Digest, a service of MarketWatch. E-mail: strategy@nytimes.com.
•
   Bent Flyvbjerg
     Nils Bruzelius
     Werner Rothengatter
   To begin, the book identifies a common feature of the
    conventional megaprojects development, that
    is, despite the overwhelming costs overrun, below-
    projection revenue, and strikingly poor performance
    records in terms of economy, environment and public
    support, megaprojects grow continuously in number
    and scale around the world, forming the so-called
    megaprojects paradox.
   Understanding of this problem and its consequences
    are explored in the first six chapters, which document
    the cost overruns, the demand over forecasts, and
    viability inflation of major megaprojects.
   In 9 out of 10 transport infrastructure projects, costs are underestimated, resulting in cost overrun;

   For rail, actual costs are, on the average, 45% higher than estimated costs (standard deviation, S.D. = 38);

   For fixed links (tunnels and bridges), actual costs are, on the average, 34% higher than estimated costs (S.D. = 62);

   For roads, actual costs are, on the average, 20% higher than estimated costs (S.D. = 30);

   For all project types, actual costs are, on the average, 28% higher than estimated costs (S.D. = 39);

   Cost underestimation and overrun exist across 20 nations and five continents; it appears to be a global phenomenon;

   Cost underestimation and overrun appear to be more pronounced in developing nations than in North America and
    Europe (data for rail only);

   Cost underestimation and overrun have not decreased over the past 70 years. No learning seems to take place;

   Cost underestimation and overrun cannot be explained by error and seem to be best explained by strategic
    misrepresentation, namely, lying, with a view to getting projects started [5].
   The authors cite as a core factor, underlying all of these
    dreadful situations with megaprojects, the lack of
    transparency in decision making and the weak
    involvement of the civil society, or what they call a
    "democracy deficit."

   They make a big point of the failure to pay attention to
    risks and the lack of accountability in the project decision-
    making processes as a main source of difficulties.

   The authors' concept of risk is almost entirely limited to
    financial risk.
   Heuristic algorithms are often employed because they may be seen to "work" without
    having been mathematically proven to meet a given set of requirements.
   Great care must be given when employing a heuristic algorithm. One common pitfall
    in implementing a heuristic method to meet a requirement comes when the engineer
    or designer fails to realize that the current data set doesn't necessarily represent
    future system states.
   While the existing data can be pored over and an algorithm can be devised to
    successfully handle the current data, it is imperative to ensure that the
    heuristic method employed is capable of handling future data
    sets.
   This means that the engineer or designer must fully understand the rules that
    generate the data and develop the algorithm to meet those requirements and not
    just address the current data sets.
   A simple example of how heuristics can fail is to answer the question "What is the
    next number in this sequence: 1, 2, 4?". One heuristic algorithm might say that the
    next number is 8 because the numbers are doubling - leading to a sequence like
    1, 2, 4, 8, 16, 32... Another, equally valid, heuristic would say that the next number is 7
    because each number is being raised by one higher interval than the one before -
    leading to a series that looks like 1, 2, 4, 7, 11, 16...
   Statistical analysis must be conducted when employing heuristics to ensure that
    enough data points are utilized to make incorrect outcomes statistically
    insignificant
   A major problem with expert estimates is overconfidence.
    To overcome this, Hubbard advocates using calibrated
    probability assessments to quantify analysts’ abilities to
    make estimates. Calibration assessments involve getting
    analysts to answer trivia questions and eliciting confidence
    intervals for each answer. The confidence intervals are
    then checked against the proportion of correct
    answers. Essentially, this assesses experts’ abilities to
    estimates by tracking how often they are right. It has been
    found that people can improve their ability to make
    subjective estimates through calibration training – i.e.
    repeated calibration testing followed by feedback.
   The Dunning–Kruger effect is a cognitive bias in which unskilled people make
    poor decisions and reach erroneous conclusions, but their incompetence
    denies them the metacognitive ability to recognize their mistakes.[1]
   The unskilled therefore suffer from illusory superiority, rating their ability as
    above average, much higher than it actually is, while the highly skilled
    underrate their own abilities, suffering from illusory inferiority.
   Actual competence may weaken self-confidence, as competent individuals
    may falsely assume that others have an equivalent understanding.
   As Kruger and Dunning conclude, "the miscalibration of the incompetent
    stems from an error about the self, whereas the miscalibration of the highly
    competent stems from an error about others" (p. 1127).[2] The effect is about
    paradoxical defects in cognitive ability, both in oneself and as one compares
    oneself to others.
Jeran Binning
   Daniel Kahneman,1 Alan B. Krueger,1,2* David Schkade,3 Norbert Schwarz,4 Arthur A. Stone5


   The belief that high income is associated with good mood is widespread but
    mostly illusory.
   People with above-average income are relatively satisfied with their lives but are
    barely happier than others in moment-to-moment experience, tend to be more
    tense, and do not spend more time in particularly enjoyable activities.
   Moreover, the effect of income on life satisfaction seems to be transient. We
    argue that people exaggerate the contribution of income to happiness because
    they focus, in part, on conventional achievements when evaluating their life or
    the lives of others.


   1 Princeton University, Princeton, NJ 08544, USA.
    2 National Bureau of Economic Research, Cambridge, MA 02138, USA.
    3 Rady School of Management, University of California, San Diego, San Diego, CA   92093, USA.
    4 Department of Psychology, University of Michigan, Ann Arbor, MI 48106, USA.
    5 Stony Brook University, Stony Brook, NY, 11794, USA.
   Proceedings of the National Academy of Sciences
     Preferences and choices can affect long-term
      happiness.
     German Study
      ▪   Marry someone who is not neurotic
      ▪   Focus more on friends, less on material goods
      ▪   Get involved with making the world a better place
      ▪   Have a job but also enough time for leisure
      ▪   Stay physically active
      ▪   For men, don’t be underweight. For women, don’t be obese.
   Don't worry, be happy" may be more than just a wishful mantra. A new
    study finds that people's happiness levels can change substantially over
    their lifetimes, suggesting that happiness isn't predetermined by genes
    or personality.
   Psychologists have long argued that people have a "set point" for
    happiness. Regardless of what life brings, the set-point theory
    goes, happiness levels tend to be stable. A big life event could create a
    boost of joy or a crush of sorrow, but within a few years, people return to
    a predetermined level of life satisfaction, according to the theory.
   The new study, which used a nationally representative sample of almost
    150,000 German adults, finds the opposite. People's long-term life
    satisfaction can change, the researchers report today (Oct. 4) in the
    online early edition of the Proceedings of the National Academy of
    Sciences. In fact, a substantial number of people followed over 25 years
    saw their happiness levels shift by one-third or more.
   The study also echoed previous happiness research in finding that
    money doesn't buy happiness.
   "People with a lot of money are more satisfied with their lives...
    but mainly due to the more interesting and challenging jobs they
    have," study author Gert Wagner, a researcher at the Max Planck
    Institute for Human Development in Germany, told LiveScience.
    "Money is simply a byproduct of good and satisfying jobs. If you
    want to be satisfied with your life, you must spend time with your
    friends and your family."
   Wagner said that previous work suggests findings on happiness
    from one developed country, like Germany, should also hold true
    for another, such as the United States. In fact, a study in May
    found that in the United States. happiness tends to increase with
    age.
Happiness Study
   I'm happier than you
   The researchers used data from a study of German adults spanning from 1984 to 2008. Each year, the participants answered
    questions on their life satisfaction, life goals and other measures like how much they exercise and socialize.

   By averaging life-satisfaction responses to even out any short-term effects, the researchers plotted out the respondents' happiness
    by percentiles. Someone in the 99th percentile, for example, would be happier than 99 percent of the study participants.

   People shifted in the rankings — and thus in their levels of happiness — quite a bit. Just over 38 percent changed their position in the
    distribution by 25 percentiles or more during the study period. About 25 percent changed by 33.3 percentiles or more, and 11.8
    percent changed by 50 percentiles.

   Feel-good factors
   So what contributed to long-term happiness? The researchers found several correlations between life choices and life satisfaction:

   Marry well: The personality traits of partners influenced people's happiness. Neuroticism, or a tendency toward anxiety, emotional
    instability and depression, was most influential. People who married or partnered with neurotic people were less likely to be happy
    than people who married non-neurotic types.

   Focus on the family: People who assigned relatively high value to altruistic and family goals compared with career goals were
    happier. Women were also happier when their male partners ranked family goals high.

   Go to church: People who went to church more often were happier, though the study can't determine whether the happiness is
    related to religious views or to the social circle religious organizations offer.

   Work, but not too much (or too little): People's happiness matched how well they felt their work hours matched their desired work
    hours. In other words, people who worked more or fewer hours than they preferred were less happy. Working less or being
    unemployed was worse than working too much, presumably because underemployment is a financial blow, the researchers wrote.

   Get social, and get moving: Social interaction and exercise were both associated with happiness. Working out made people happier
    regardless of body weight. The only correlation between body weight and happiness was that underweight men and obese women
    were more likely to be unhappy.
Happiness Study
   Mysteries of happiness
   "In its extreme form, set-point theory was never credible," Daniel Kahneman, an emeritus professor of psychology at
    Princeton University and the winner of the 2002 Nobel Prize in Economic Sciences, told LiveScience. "If it was taken to
    mean that the only factor that determines happiness or life satisfaction is genetic, so that people always come back to
    exactly to the same point, this was utterly incredible."
   The current study is a useful demonstration that life changes can influence people's life satisfaction, said Kahneman, who
    was not involved in the research. However, the correlations between certain goals and traits and happiness doesn't
    necessarily answer the nature-versus-nurture question.
   "They're suggesting that the goals are chosen. But the goals may be part of personality," and thus partially genetic, he
    said. "The fact that goals matter, like altruism and materialism, that really doesn't help us distinguish between personality
    and circumstances."
   More studies are needed that track large populations of people after influential changes, like the enactment of new laws,
    said Andrew Oswald, a professor of behavioral science at the University of Warwick who studies happiness but was not
    involved in the current study. By comparing people who lived under, say, a new state tax law that affected income to those
    who lived in a nearby state without the law, researchers could begin to look at happiness in a more experimental way, he
    said.
   "The key thing is that life events good and bad do shape happiness over long periods," Oswald said. "We are, in part, the
    product of our experiences. It's not all born into us."
   7 Ways the Mind and Body Change with Age
   Happiness is Being Old, Male and Republican
   10 Ways to Keep Your Mind Sharp
Thinking fast and_slow

More Related Content

What's hot

Thinking Fast & Slow presentation
Thinking Fast & Slow presentationThinking Fast & Slow presentation
Thinking Fast & Slow presentationLaure Parsons
 
Thinking fast and slow
Thinking fast and slow Thinking fast and slow
Thinking fast and slow Yogesh SN
 
Rawnet Lightning talk - 'Thinking, Fast and Slow' review
Rawnet Lightning talk - 'Thinking, Fast and Slow' reviewRawnet Lightning talk - 'Thinking, Fast and Slow' review
Rawnet Lightning talk - 'Thinking, Fast and Slow' reviewRawnet
 
You're not so smart - Cognitive Biases
You're not so smart - Cognitive BiasesYou're not so smart - Cognitive Biases
You're not so smart - Cognitive BiasesOdair Faléco
 
Cognitive Biases and Bayesian reasoning
Cognitive Biases and Bayesian reasoningCognitive Biases and Bayesian reasoning
Cognitive Biases and Bayesian reasoningMauro DG
 
Critical thinking powerpoint
Critical thinking powerpointCritical thinking powerpoint
Critical thinking powerpointBarbara Bali
 
Critical thinking
Critical thinkingCritical thinking
Critical thinkingLance Jones
 
critical thinking ppt.pdf
critical thinking ppt.pdfcritical thinking ppt.pdf
critical thinking ppt.pdfMamRocilValdez
 
Influencing skills
Influencing skillsInfluencing skills
Influencing skillsMohit Singla
 
Decision Making Process
Decision Making ProcessDecision Making Process
Decision Making ProcessGhufran Ataie
 
Decision making process
Decision making processDecision making process
Decision making processVikash K Giri
 
Introduction to Critical Thinking Skills
Introduction to Critical Thinking SkillsIntroduction to Critical Thinking Skills
Introduction to Critical Thinking SkillsHanis Razak
 
Fundamentals of Decision Making
Fundamentals of Decision Making Fundamentals of Decision Making
Fundamentals of Decision Making Andrew Hirst
 
Decision making
Decision making Decision making
Decision making fn231
 

What's hot (20)

Thinking Fast and Slow
Thinking Fast and SlowThinking Fast and Slow
Thinking Fast and Slow
 
Thinking Fast & Slow presentation
Thinking Fast & Slow presentationThinking Fast & Slow presentation
Thinking Fast & Slow presentation
 
Thinking fast and slow
Thinking fast and slow Thinking fast and slow
Thinking fast and slow
 
Rawnet Lightning talk - 'Thinking, Fast and Slow' review
Rawnet Lightning talk - 'Thinking, Fast and Slow' reviewRawnet Lightning talk - 'Thinking, Fast and Slow' review
Rawnet Lightning talk - 'Thinking, Fast and Slow' review
 
Cognitive biases
Cognitive biasesCognitive biases
Cognitive biases
 
You're not so smart - Cognitive Biases
You're not so smart - Cognitive BiasesYou're not so smart - Cognitive Biases
You're not so smart - Cognitive Biases
 
Cognitive Biases and Bayesian reasoning
Cognitive Biases and Bayesian reasoningCognitive Biases and Bayesian reasoning
Cognitive Biases and Bayesian reasoning
 
Cognitive biases
Cognitive biasesCognitive biases
Cognitive biases
 
Critical thinking powerpoint
Critical thinking powerpointCritical thinking powerpoint
Critical thinking powerpoint
 
Critical thinking
Critical thinkingCritical thinking
Critical thinking
 
critical thinking ppt.pdf
critical thinking ppt.pdfcritical thinking ppt.pdf
critical thinking ppt.pdf
 
Influencing skills
Influencing skillsInfluencing skills
Influencing skills
 
Decision Making Process
Decision Making ProcessDecision Making Process
Decision Making Process
 
Decision making process
Decision making processDecision making process
Decision making process
 
DECISION MAKING
DECISION MAKINGDECISION MAKING
DECISION MAKING
 
Introduction to Critical Thinking Skills
Introduction to Critical Thinking SkillsIntroduction to Critical Thinking Skills
Introduction to Critical Thinking Skills
 
Critical thinking
Critical thinkingCritical thinking
Critical thinking
 
Fundamentals of Decision Making
Fundamentals of Decision Making Fundamentals of Decision Making
Fundamentals of Decision Making
 
Decision Making
Decision MakingDecision Making
Decision Making
 
Decision making
Decision making Decision making
Decision making
 

Viewers also liked

Cognitive biases - a visual study guide
Cognitive biases - a visual study guideCognitive biases - a visual study guide
Cognitive biases - a visual study guideEric Fernandez
 
Insight Paths, a book summary in 24 slides
Insight Paths, a book summary in 24 slidesInsight Paths, a book summary in 24 slides
Insight Paths, a book summary in 24 slidesEmily Hean
 
Comparing Two Types of Decision-making: When experts are better -- or not
Comparing Two Types of Decision-making: When experts are better -- or notComparing Two Types of Decision-making: When experts are better -- or not
Comparing Two Types of Decision-making: When experts are better -- or notKerry Cunningham
 
Why People Make Bad Decisions: The Role of Cognitive Biases
Why People Make Bad Decisions: The Role of Cognitive BiasesWhy People Make Bad Decisions: The Role of Cognitive Biases
Why People Make Bad Decisions: The Role of Cognitive BiasesSigma Consulting Resources, LLC
 
Behavioral economics
Behavioral economicsBehavioral economics
Behavioral economicsJeff Gilman
 
Medical decision making edited
Medical decision making editedMedical decision making edited
Medical decision making editedJeff Young
 
The Quick & The Dead
The Quick & The DeadThe Quick & The Dead
The Quick & The Deadcconvery
 
Cognitive Biases and Pareidolia/Apophenia
Cognitive Biases and Pareidolia/ApopheniaCognitive Biases and Pareidolia/Apophenia
Cognitive Biases and Pareidolia/ApopheniaArthur Doler
 
Личная (не)эффективность в контексте компании или "тупняк" как чума 21-го века
Личная (не)эффективность в контексте компании или "тупняк" как чума 21-го векаЛичная (не)эффективность в контексте компании или "тупняк" как чума 21-го века
Личная (не)эффективность в контексте компании или "тупняк" как чума 21-го векаIT-Доминанта
 
Beyond the Next Quarter: Strategic Thinking in a Fast-Paced Innovative Enviro...
Beyond the Next Quarter: Strategic Thinking in a Fast-Paced Innovative Enviro...Beyond the Next Quarter: Strategic Thinking in a Fast-Paced Innovative Enviro...
Beyond the Next Quarter: Strategic Thinking in a Fast-Paced Innovative Enviro...denisepmorgan
 
The Power of Intuition in Business
The Power of Intuition in BusinessThe Power of Intuition in Business
The Power of Intuition in BusinessDr. Vickram Aadityaa
 

Viewers also liked (12)

Cognitive biases - a visual study guide
Cognitive biases - a visual study guideCognitive biases - a visual study guide
Cognitive biases - a visual study guide
 
Insight Paths, a book summary in 24 slides
Insight Paths, a book summary in 24 slidesInsight Paths, a book summary in 24 slides
Insight Paths, a book summary in 24 slides
 
Comparing Two Types of Decision-making: When experts are better -- or not
Comparing Two Types of Decision-making: When experts are better -- or notComparing Two Types of Decision-making: When experts are better -- or not
Comparing Two Types of Decision-making: When experts are better -- or not
 
Why People Make Bad Decisions: The Role of Cognitive Biases
Why People Make Bad Decisions: The Role of Cognitive BiasesWhy People Make Bad Decisions: The Role of Cognitive Biases
Why People Make Bad Decisions: The Role of Cognitive Biases
 
Behavioral economics
Behavioral economicsBehavioral economics
Behavioral economics
 
Thinking
ThinkingThinking
Thinking
 
Medical decision making edited
Medical decision making editedMedical decision making edited
Medical decision making edited
 
The Quick & The Dead
The Quick & The DeadThe Quick & The Dead
The Quick & The Dead
 
Cognitive Biases and Pareidolia/Apophenia
Cognitive Biases and Pareidolia/ApopheniaCognitive Biases and Pareidolia/Apophenia
Cognitive Biases and Pareidolia/Apophenia
 
Личная (не)эффективность в контексте компании или "тупняк" как чума 21-го века
Личная (не)эффективность в контексте компании или "тупняк" как чума 21-го векаЛичная (не)эффективность в контексте компании или "тупняк" как чума 21-го века
Личная (не)эффективность в контексте компании или "тупняк" как чума 21-го века
 
Beyond the Next Quarter: Strategic Thinking in a Fast-Paced Innovative Enviro...
Beyond the Next Quarter: Strategic Thinking in a Fast-Paced Innovative Enviro...Beyond the Next Quarter: Strategic Thinking in a Fast-Paced Innovative Enviro...
Beyond the Next Quarter: Strategic Thinking in a Fast-Paced Innovative Enviro...
 
The Power of Intuition in Business
The Power of Intuition in BusinessThe Power of Intuition in Business
The Power of Intuition in Business
 

Similar to Thinking fast and_slow

Why Behavioral Finance is Helpful for Investors to Decision Making Process?
Why Behavioral Finance is Helpful for Investors to Decision Making Process?Why Behavioral Finance is Helpful for Investors to Decision Making Process?
Why Behavioral Finance is Helpful for Investors to Decision Making Process?QUESTJOURNAL
 
UNDERSTANDING DECISION/ GAME THEORY FOR BETTER RISK ASSESSMENT.
UNDERSTANDING DECISION/ GAME THEORY FOR BETTER RISK ASSESSMENT.UNDERSTANDING DECISION/ GAME THEORY FOR BETTER RISK ASSESSMENT.
UNDERSTANDING DECISION/ GAME THEORY FOR BETTER RISK ASSESSMENT.Kaustav Lahiri
 
Prospect Theory_John McGinn
Prospect Theory_John McGinnProspect Theory_John McGinn
Prospect Theory_John McGinnJohn McGinn
 
Introduction to Behavioural Finance
Introduction to Behavioural FinanceIntroduction to Behavioural Finance
Introduction to Behavioural Financestockedin
 
Mo Tanweer on Behavioural economics (2010)
Mo Tanweer on Behavioural economics (2010)Mo Tanweer on Behavioural economics (2010)
Mo Tanweer on Behavioural economics (2010)tutor2u
 
A k a d e m i k A r a ç t i r m a l a r D e r g i s i 2 0 1 0 .docx
A k a d e m i k A r a ç t i r m a l a r D e r g i s i 2 0 1 0 .docxA k a d e m i k A r a ç t i r m a l a r D e r g i s i 2 0 1 0 .docx
A k a d e m i k A r a ç t i r m a l a r D e r g i s i 2 0 1 0 .docxrobert345678
 
Disasters and Humans (DEMS3706 SU2020, Dr. Eric Kennedy)APDEMS370
Disasters and Humans (DEMS3706 SU2020, Dr. Eric Kennedy)APDEMS370Disasters and Humans (DEMS3706 SU2020, Dr. Eric Kennedy)APDEMS370
Disasters and Humans (DEMS3706 SU2020, Dr. Eric Kennedy)APDEMS370AlyciaGold776
 
Mathematical Reasoning (unit-5) UGC NET Paper-1 Study Notes (E-books) Down...
Mathematical  Reasoning (unit-5) UGC NET Paper-1  Study Notes (E-books)  Down...Mathematical  Reasoning (unit-5) UGC NET Paper-1  Study Notes (E-books)  Down...
Mathematical Reasoning (unit-5) UGC NET Paper-1 Study Notes (E-books) Down...DIwakar Rajput
 
Book Preview: The Timeless Essence of Financial Science
Book Preview: The Timeless Essence of Financial ScienceBook Preview: The Timeless Essence of Financial Science
Book Preview: The Timeless Essence of Financial ScienceRajesh Mudholkar
 
Derailed Crisis Plans
Derailed Crisis PlansDerailed Crisis Plans
Derailed Crisis Plansmschannon
 
Capitalistic Musings
Capitalistic MusingsCapitalistic Musings
Capitalistic MusingsSam Vaknin
 
behavioral finance:theories, issues and challenges
behavioral finance:theories, issues and challenges behavioral finance:theories, issues and challenges
behavioral finance:theories, issues and challenges Kamaljit Singh
 
Risk Taking To Get Ahead
Risk Taking To Get AheadRisk Taking To Get Ahead
Risk Taking To Get AheadChelse Benham
 

Similar to Thinking fast and_slow (20)

Biases april 2012
Biases april 2012Biases april 2012
Biases april 2012
 
Why Behavioral Finance is Helpful for Investors to Decision Making Process?
Why Behavioral Finance is Helpful for Investors to Decision Making Process?Why Behavioral Finance is Helpful for Investors to Decision Making Process?
Why Behavioral Finance is Helpful for Investors to Decision Making Process?
 
UNDERSTANDING DECISION/ GAME THEORY FOR BETTER RISK ASSESSMENT.
UNDERSTANDING DECISION/ GAME THEORY FOR BETTER RISK ASSESSMENT.UNDERSTANDING DECISION/ GAME THEORY FOR BETTER RISK ASSESSMENT.
UNDERSTANDING DECISION/ GAME THEORY FOR BETTER RISK ASSESSMENT.
 
The Framing Effect
The Framing EffectThe Framing Effect
The Framing Effect
 
Prospect Theory_John McGinn
Prospect Theory_John McGinnProspect Theory_John McGinn
Prospect Theory_John McGinn
 
Introduction to Behavioural Finance
Introduction to Behavioural FinanceIntroduction to Behavioural Finance
Introduction to Behavioural Finance
 
Learning session 2nd
Learning session 2ndLearning session 2nd
Learning session 2nd
 
Mo Tanweer on Behavioural economics (2010)
Mo Tanweer on Behavioural economics (2010)Mo Tanweer on Behavioural economics (2010)
Mo Tanweer on Behavioural economics (2010)
 
A k a d e m i k A r a ç t i r m a l a r D e r g i s i 2 0 1 0 .docx
A k a d e m i k A r a ç t i r m a l a r D e r g i s i 2 0 1 0 .docxA k a d e m i k A r a ç t i r m a l a r D e r g i s i 2 0 1 0 .docx
A k a d e m i k A r a ç t i r m a l a r D e r g i s i 2 0 1 0 .docx
 
K t
K tK t
K t
 
Butterfly paper
Butterfly paperButterfly paper
Butterfly paper
 
Disasters and Humans (DEMS3706 SU2020, Dr. Eric Kennedy)APDEMS370
Disasters and Humans (DEMS3706 SU2020, Dr. Eric Kennedy)APDEMS370Disasters and Humans (DEMS3706 SU2020, Dr. Eric Kennedy)APDEMS370
Disasters and Humans (DEMS3706 SU2020, Dr. Eric Kennedy)APDEMS370
 
Mathematical Reasoning (unit-5) UGC NET Paper-1 Study Notes (E-books) Down...
Mathematical  Reasoning (unit-5) UGC NET Paper-1  Study Notes (E-books)  Down...Mathematical  Reasoning (unit-5) UGC NET Paper-1  Study Notes (E-books)  Down...
Mathematical Reasoning (unit-5) UGC NET Paper-1 Study Notes (E-books) Down...
 
Book Preview: The Timeless Essence of Financial Science
Book Preview: The Timeless Essence of Financial ScienceBook Preview: The Timeless Essence of Financial Science
Book Preview: The Timeless Essence of Financial Science
 
Derailed Crisis Plans
Derailed Crisis PlansDerailed Crisis Plans
Derailed Crisis Plans
 
Capitalistic Musings
Capitalistic MusingsCapitalistic Musings
Capitalistic Musings
 
Maths probability
Maths probabilityMaths probability
Maths probability
 
behavioral finance:theories, issues and challenges
behavioral finance:theories, issues and challenges behavioral finance:theories, issues and challenges
behavioral finance:theories, issues and challenges
 
PSYX320_finalproposal
PSYX320_finalproposalPSYX320_finalproposal
PSYX320_finalproposal
 
Risk Taking To Get Ahead
Risk Taking To Get AheadRisk Taking To Get Ahead
Risk Taking To Get Ahead
 

More from Jeran Binning

A critique of doug hubbards the failure of risk management
A critique of doug hubbards the failure of risk managementA critique of doug hubbards the failure of risk management
A critique of doug hubbards the failure of risk managementJeran Binning
 
Space acquisition environment 01 sep 2016_jeran_binning_v2.0
Space acquisition environment 01 sep 2016_jeran_binning_v2.0Space acquisition environment 01 sep 2016_jeran_binning_v2.0
Space acquisition environment 01 sep 2016_jeran_binning_v2.0Jeran Binning
 
Presentation micro expressions
Presentation micro expressionsPresentation micro expressions
Presentation micro expressionsJeran Binning
 
The biology of bubble and crash
The biology of bubble and crashThe biology of bubble and crash
The biology of bubble and crashJeran Binning
 
History of Governmet Contracting 4 31 jan 12 2
History of Governmet Contracting 4 31 jan 12 2History of Governmet Contracting 4 31 jan 12 2
History of Governmet Contracting 4 31 jan 12 2Jeran Binning
 
The biology of bubble and crash
The biology of bubble and crashThe biology of bubble and crash
The biology of bubble and crashJeran Binning
 
Supplier management v4.0_11_mar_12
Supplier management v4.0_11_mar_12Supplier management v4.0_11_mar_12
Supplier management v4.0_11_mar_12Jeran Binning
 
Hx of contracting 3 16 feb 12
Hx of contracting 3 16 feb 12Hx of contracting 3 16 feb 12
Hx of contracting 3 16 feb 12Jeran Binning
 

More from Jeran Binning (9)

A critique of doug hubbards the failure of risk management
A critique of doug hubbards the failure of risk managementA critique of doug hubbards the failure of risk management
A critique of doug hubbards the failure of risk management
 
Space acquisition environment 01 sep 2016_jeran_binning_v2.0
Space acquisition environment 01 sep 2016_jeran_binning_v2.0Space acquisition environment 01 sep 2016_jeran_binning_v2.0
Space acquisition environment 01 sep 2016_jeran_binning_v2.0
 
Risk management
Risk management Risk management
Risk management
 
Presentation micro expressions
Presentation micro expressionsPresentation micro expressions
Presentation micro expressions
 
The biology of bubble and crash
The biology of bubble and crashThe biology of bubble and crash
The biology of bubble and crash
 
History of Governmet Contracting 4 31 jan 12 2
History of Governmet Contracting 4 31 jan 12 2History of Governmet Contracting 4 31 jan 12 2
History of Governmet Contracting 4 31 jan 12 2
 
The biology of bubble and crash
The biology of bubble and crashThe biology of bubble and crash
The biology of bubble and crash
 
Supplier management v4.0_11_mar_12
Supplier management v4.0_11_mar_12Supplier management v4.0_11_mar_12
Supplier management v4.0_11_mar_12
 
Hx of contracting 3 16 feb 12
Hx of contracting 3 16 feb 12Hx of contracting 3 16 feb 12
Hx of contracting 3 16 feb 12
 

Recently uploaded

Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptx
Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptxQ4-PPT-Music9_Lesson-1-Romantic-Opera.pptx
Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptxlancelewisportillo
 
Global Lehigh Strategic Initiatives (without descriptions)
Global Lehigh Strategic Initiatives (without descriptions)Global Lehigh Strategic Initiatives (without descriptions)
Global Lehigh Strategic Initiatives (without descriptions)cama23
 
How to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERPHow to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERPCeline George
 
Influencing policy (training slides from Fast Track Impact)
Influencing policy (training slides from Fast Track Impact)Influencing policy (training slides from Fast Track Impact)
Influencing policy (training slides from Fast Track Impact)Mark Reed
 
Earth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice greatEarth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice greatYousafMalik24
 
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITYISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITYKayeClaireEstoconing
 
Inclusivity Essentials_ Creating Accessible Websites for Nonprofits .pdf
Inclusivity Essentials_ Creating Accessible Websites for Nonprofits .pdfInclusivity Essentials_ Creating Accessible Websites for Nonprofits .pdf
Inclusivity Essentials_ Creating Accessible Websites for Nonprofits .pdfTechSoup
 
Karra SKD Conference Presentation Revised.pptx
Karra SKD Conference Presentation Revised.pptxKarra SKD Conference Presentation Revised.pptx
Karra SKD Conference Presentation Revised.pptxAshokKarra1
 
Keynote by Prof. Wurzer at Nordex about IP-design
Keynote by Prof. Wurzer at Nordex about IP-designKeynote by Prof. Wurzer at Nordex about IP-design
Keynote by Prof. Wurzer at Nordex about IP-designMIPLM
 
What is Model Inheritance in Odoo 17 ERP
What is Model Inheritance in Odoo 17 ERPWhat is Model Inheritance in Odoo 17 ERP
What is Model Inheritance in Odoo 17 ERPCeline George
 
ANG SEKTOR NG agrikultura.pptx QUARTER 4
ANG SEKTOR NG agrikultura.pptx QUARTER 4ANG SEKTOR NG agrikultura.pptx QUARTER 4
ANG SEKTOR NG agrikultura.pptx QUARTER 4MiaBumagat1
 
Difference Between Search & Browse Methods in Odoo 17
Difference Between Search & Browse Methods in Odoo 17Difference Between Search & Browse Methods in Odoo 17
Difference Between Search & Browse Methods in Odoo 17Celine George
 
Food processing presentation for bsc agriculture hons
Food processing presentation for bsc agriculture honsFood processing presentation for bsc agriculture hons
Food processing presentation for bsc agriculture honsManeerUddin
 
INTRODUCTION TO CATHOLIC CHRISTOLOGY.pptx
INTRODUCTION TO CATHOLIC CHRISTOLOGY.pptxINTRODUCTION TO CATHOLIC CHRISTOLOGY.pptx
INTRODUCTION TO CATHOLIC CHRISTOLOGY.pptxHumphrey A Beña
 
Transaction Management in Database Management System
Transaction Management in Database Management SystemTransaction Management in Database Management System
Transaction Management in Database Management SystemChristalin Nelson
 
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptxMULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptxAnupkumar Sharma
 
Integumentary System SMP B. Pharm Sem I.ppt
Integumentary System SMP B. Pharm Sem I.pptIntegumentary System SMP B. Pharm Sem I.ppt
Integumentary System SMP B. Pharm Sem I.pptshraddhaparab530
 
Active Learning Strategies (in short ALS).pdf
Active Learning Strategies (in short ALS).pdfActive Learning Strategies (in short ALS).pdf
Active Learning Strategies (in short ALS).pdfPatidar M
 

Recently uploaded (20)

Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptx
Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptxQ4-PPT-Music9_Lesson-1-Romantic-Opera.pptx
Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptx
 
Global Lehigh Strategic Initiatives (without descriptions)
Global Lehigh Strategic Initiatives (without descriptions)Global Lehigh Strategic Initiatives (without descriptions)
Global Lehigh Strategic Initiatives (without descriptions)
 
How to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERPHow to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERP
 
Influencing policy (training slides from Fast Track Impact)
Influencing policy (training slides from Fast Track Impact)Influencing policy (training slides from Fast Track Impact)
Influencing policy (training slides from Fast Track Impact)
 
Earth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice greatEarth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice great
 
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITYISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
 
Inclusivity Essentials_ Creating Accessible Websites for Nonprofits .pdf
Inclusivity Essentials_ Creating Accessible Websites for Nonprofits .pdfInclusivity Essentials_ Creating Accessible Websites for Nonprofits .pdf
Inclusivity Essentials_ Creating Accessible Websites for Nonprofits .pdf
 
Karra SKD Conference Presentation Revised.pptx
Karra SKD Conference Presentation Revised.pptxKarra SKD Conference Presentation Revised.pptx
Karra SKD Conference Presentation Revised.pptx
 
Keynote by Prof. Wurzer at Nordex about IP-design
Keynote by Prof. Wurzer at Nordex about IP-designKeynote by Prof. Wurzer at Nordex about IP-design
Keynote by Prof. Wurzer at Nordex about IP-design
 
What is Model Inheritance in Odoo 17 ERP
What is Model Inheritance in Odoo 17 ERPWhat is Model Inheritance in Odoo 17 ERP
What is Model Inheritance in Odoo 17 ERP
 
ANG SEKTOR NG agrikultura.pptx QUARTER 4
ANG SEKTOR NG agrikultura.pptx QUARTER 4ANG SEKTOR NG agrikultura.pptx QUARTER 4
ANG SEKTOR NG agrikultura.pptx QUARTER 4
 
Difference Between Search & Browse Methods in Odoo 17
Difference Between Search & Browse Methods in Odoo 17Difference Between Search & Browse Methods in Odoo 17
Difference Between Search & Browse Methods in Odoo 17
 
Food processing presentation for bsc agriculture hons
Food processing presentation for bsc agriculture honsFood processing presentation for bsc agriculture hons
Food processing presentation for bsc agriculture hons
 
LEFT_ON_C'N_ PRELIMS_EL_DORADO_2024.pptx
LEFT_ON_C'N_ PRELIMS_EL_DORADO_2024.pptxLEFT_ON_C'N_ PRELIMS_EL_DORADO_2024.pptx
LEFT_ON_C'N_ PRELIMS_EL_DORADO_2024.pptx
 
INTRODUCTION TO CATHOLIC CHRISTOLOGY.pptx
INTRODUCTION TO CATHOLIC CHRISTOLOGY.pptxINTRODUCTION TO CATHOLIC CHRISTOLOGY.pptx
INTRODUCTION TO CATHOLIC CHRISTOLOGY.pptx
 
Transaction Management in Database Management System
Transaction Management in Database Management SystemTransaction Management in Database Management System
Transaction Management in Database Management System
 
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptxMULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
 
Raw materials used in Herbal Cosmetics.pptx
Raw materials used in Herbal Cosmetics.pptxRaw materials used in Herbal Cosmetics.pptx
Raw materials used in Herbal Cosmetics.pptx
 
Integumentary System SMP B. Pharm Sem I.ppt
Integumentary System SMP B. Pharm Sem I.pptIntegumentary System SMP B. Pharm Sem I.ppt
Integumentary System SMP B. Pharm Sem I.ppt
 
Active Learning Strategies (in short ALS).pdf
Active Learning Strategies (in short ALS).pdfActive Learning Strategies (in short ALS).pdf
Active Learning Strategies (in short ALS).pdf
 

Thinking fast and_slow

  • 1. Jeran Binning DAU San Diego
  • 2. http://www.nytimes.com/pages/books/review/index.html Copyright 2011 The New York Times Company Illustration by David Plunkert
  • 3. Expertise can be learnt by prolonged exposure to situations that are “sufficiently regular to be predictable”, and in which the expert gets quick and decisive feedback on whether he did the right or the wrong thing.  Experts can thus train their unconscious “pattern recognition” mechanism to produce the right answer quickly.  So this certainly applies to chess, and it certainly does not apply to predicting the course of Middle East politics.
  • 4. Kahneman’s career has had three major phases
  • 5. In the first, he and Tversky did a series of ingenious experiments that revealed twenty or so “cognitive biases” — unconscious errors of reasoning that distort our judgment of the world.  Typical of these is the “anchoring effect”: our tendency to be influenced by irrelevant numbers that we happen to be exposed to.
  • 6. Anchoring and adjustment: People who have to make judgments under uncertainty use this heuristic by starting with a certain reference point (anchor) and then adjust it insufficiently to reach a final conclusion.  Example: If you have to judge another persons productivity, the anchor for your final (adjusted) judgment may be your own level of productivity. Depending on your own level of productivity you might therefore underestimate or overestimate the productivity of this person.
  • 7. In the second phase, Kahneman and Tversky showed that people making decisions under uncertain conditions do not behave in the way that economic models have traditionally assumed; they do not “maximize utility.”
  • 8. The two then developed an alternative account of decision making, one more faithful to human psychology, which they called “prospect theory.” (It was for this achievement that Kahneman was awarded the Nobel.)
  • 9. Prospect theory was developed by Daniel Kahneman and Amos Tversky in 1979
  • 10. Prospect theory was developed by Daniel Kahneman and Amos Tversky in 1979 as a psychologically realistic alternative to expected utility theory. It allows one to describe how people make choices in situations where they have to decide between alternatives that involve risk, e.g. in financial decisions. Starting from empirical evidence, the theory describes how individuals evaluate potential losses and gains. In the original formulation the term prospect referred to a lottery.  The theory describes such decision processes as consisting of two stages, editing and evaluation. In the first, possible outcomes of the decision are ordered following some heuristic. In particular, people decide which outcomes they see as basically identical and they set a reference point and consider lower outcomes as losses and larger as gains. In the following evaluation phase, people behave as if they would compute a value (utility), based on the potential outcomes and their respective probabilities, and then choose the alternative having a higher utility.  The formula that Kahneman and Tversky assume for the evaluation phase is (in its simplest form) given by where are the potential outcomes and their respective probabilities. v is a so-called value function that assigns a value to an outcome. The value function (sketched in the Figure) which passes through the reference point is s-shaped and, as its asymmetry implies, given the same variation in absolute value, there is a bigger impact of losses than of gains (loss aversion). In contrast to Expected Utility Theory, it measures losses and gains, but not absolute wealth. The function w is called a probability weighting function and expresses that people tend to overreact to small probability events, but underreact to medium and large probabilities  To see how Prospect Theory (PT) can be applied in an example, consider a decision about buying an insurance policy. Let us assume the probability of the insured risk is 1%, the potential loss is $1000 and the premium is $15. If we apply PT, we first need to set a reference point. This could be, e.g., the current wealth, or the worst case (losing $1000). If we set the frame to the current wealth, the decision would be to either pay $15 for sure (which gives the PT-utility of v( − 15)) or a lottery with outcomes $0 (probability 99%) or $-1000 (probability 1%) which yields the PT-utility of . These expressions can be computed numerically. For typical value and weighting functions, the former expression could be larger due to the convexity of v in losses, and hence the insurance looks unattractive. If we set the frame to $-1000, both alternatives are set in gains. The concavity of the value function in gains can then lead to a preference for buying the insurance.  We see in this example that a strong overweighting of small probabilities can also undo the effect of the convexity of v in losses: the potential outcome of losing $1000 is overweighted.  The interplay of overweighting of small probabilities and concavity-convexity of the value function leads to the so-called four-fold pattern of risk attitudes: risk-averse behavior in gains involving moderate probabilities and of small probability losses; risk-seeking behavior in losses involving moderate probabilities and of small probability gains. This is an explanation for the fact that people, e.g., simultaneously buy lottery tickets and insurances, but still invest money conservatively.
  • 11. An important paper in the development of the behavioral finance and economics fields was written by Kahneman and Tversky in 1979. This paper, 'Prospect theory: An Analysis of Decision Under Risk', used cognitive psychological techniques to explain a number of documented divergences of economic decision making from neo-classical theory.  Over time many other psychological effects have been incorporated into behavioral finance, such as overconfidence and the effects of limited attention.  Further milestones in the development of the field include a well attended and diverse conference at the University of Chicago,[4] a special 1997 edition of the Quarterly Journal of Economics ('In Memory of Amos Tversky') devoted to the topic of behavioral economics and the award of the Nobel prize to Daniel Kahneman in 2002 "for having integrated insights from psychological research into economic science, especially concerning human judgment and decision-making under uncertainty"
  • 12. In the third phase of his career, mainly after the death of Tversky, Kahneman has delved into “hedonic psychology”: the science of happiness, its nature and its causes. His findings in this area have proved disquieting — and not just because one of the key experiments involved a deliberately prolonged colonoscopy.
  • 13. Kahneman never grapples philosophically with the nature of rationality. He does, however, supply a fascinating account of what might be taken to be its goal: happiness.  What does it mean to be happy? When Kahneman first took up this question, in the mid 1990s, most happiness research relied on asking people how satisfied they were with their life on the whole.  But such retrospective assessments depend on memory, which is notoriously unreliable.
  • 14. What if, instead, a person’s actual experience of pleasure or pain could be sampled from moment to moment, and then summed up over time? Kahneman calls this “experienced” well-being, as opposed to the “remembered” well-being that researchers had relied upon.  And he found that these two measures of happiness diverge in surprising ways. What makes the “experiencing self” happy is not the same as what makes the “remembering self” happy.  In particular, the remembering self does not care about duration — how long a pleasant or unpleasant experience lasts. Rather, it retrospectively rates an experience by the peak level of pain or pleasure in the course of the experience, and by the way the experience ends.
  • 15. Clearly, much remains to be done in hedonic psychology. But Kahneman’s conceptual innovations have laid the foundation for many of the empirical findings he reports in this book:  that while French mothers spend less time with their children than American mothers, they enjoy it more;  that headaches are hedonically harder on the poor;  that women who live alone seem to enjoy the same level of well-being as women who live with a mate; and  that a household income of about $75,000 in high-cost areas of the country is sufficient to maximize happiness.  Policy makers interested in lowering the misery index of society will find much to ponder here.
  • 16. System 2, in Kahneman’s scheme, is our slow, deliberate, analytical and consciously effortful mode of reasoning about the world.  System 1, by contrast, is our fast, automatic, intuitive and largely unconscious mode.  It is System 1 that detects hostility in a voice and effortlessly completes the phrase “bread and. . . . ”  It is System 2 that swings into action when we have to fill out a tax form or park a car in a narrow space. (As Kahneman and others have found, there is an easy way to tell how engaged a person’s System 2 is during a task: just look into his or her eyes and note how dilated the pupils are.)
  • 17. More generally, System 1 uses association and metaphor to produce a quick and dirty draft of reality, which System 2 draws on to arrive at explicit beliefs and reasoned choices.  System 1 proposes, System 2 disposes.  So System 2 would seem to be the boss, right?  In principle, yes. But System 2, in addition to being more deliberate and rational, is also lazy. And it tires easily. (The vogue term for this is “ego depletion.”)
  • 18. “Although System 2 believes itself to be where the action is,” Kahneman writes, “the automatic System 1 is the hero of this book.” System 2 is especially quiescent, it seems, when your mood is a happy one.
  • 19.
  • 20. Memory also holds the vast repertory of skills we have acquired in a lifetime of practice, which automatically produce adequate solutions to challenges as they arise…. The acquisition of skills requires a regular environment, an adequate opportunity to practice, and rapid and unequivocal feedback about the correctness of thoughts and actions.
  • 21. When these conditions are fulfilled, skill eventually develops, and the intuitive judgments and choices that quickly come to mind will mostly be accurate. All of this is the work of System 1, which means it occurs automatically and fast. A marker of skilled performance is the ability to deal with vast amounts of information swiftly and efficiently. Pp 416 conclusions.
  • 22. The way to block errors that originate in system one us a simple in principle: recognize signs that you are in a cognitive minefield, slow down and ask for reinforcement from System 2.
  • 23. What can be done about biases? How can we improve judgments and decisions, both our own and those of the institutions that we serve and serve us?
  • 24. The short answer is that little can be achieved without considerable investment of effort. As I know from experience System 1 is not readily educable. Except for some effects that I attribute mostly to my age, my intuitive thinking is just as prone to overconfidence, extreme predications, and the planning fallacy as it was before I made a study of these issues.
  • 25. I have improved only in my ability to recognize situations in which errors are likely; “This number will be an anchor…” “The decision could be changes if the question was reframed…” And I have made much more progress in recognizing the errors of others than my own.”
  • 26. By the time I got to the end of “Thinking, Fast and Slow,” my skeptical frown had long since given way to a grin of intellectual satisfaction. Appraising the book by the peak-end rule, I overconfidently urge everyone to buy and read it. But for those who are merely interested in Kahneman’s takeaway on the Malcolm Gladwell question it is this:  If you’ve had 10,000 hours of training in a predictable, rapid-feedback environment — chess, firefighting, anesthesiology — then blink. In all other cases, think.
  • 27. Projects fail at a spectacular rate. One reason is that too many people are reluctant to speak up about their reservations during the all-important planning phase. By making it safe for dissenters who are knowledgeable about the undertaking and worried about its weaknesses to speak up, you can improve a project’s chances of success.  Research conducted in 1989 by Deborah J. Mitchell, of the Wharton School; Jay Russo, of Cornell; and Nancy Pennington, of the University of Colorado, found that prospective hindsight— imagining that an event has already occurred—increases the ability to correctly identify reasons for future outcomes by 30%. We have used prospective hindsight to devise a method called a premortem, which helps project teams identify risks at the outset.
  • 28. Judgment Under Uncertainty: Heuristics and Biases. Amos Tversky and Daniel Kahneman  Science, Volume 185, 1974  Research for DARPA N00014-73C-0438 monitored by ONR and Research and Development Authority of Hebrew University, Jerusalem, Israel.
  • 29. The same goes for all of us, almost all the time. We think we're smart; we're confident we won't be unconsciously swayed by the high list price of a house. We're wrong. (Kahneman admits his own inability to counter some of these effects.) We're also hopelessly subject to the "focusing illusion", which can be conveyed in one sentence: "Nothing in life is as important as you think it is when you're thinking about it."  Whatever we focus on, it bulges in the heat of our attention until we assume its role in our life as a whole is greater than it is.  Another systematic error involves "duration neglect" and the "peak-end rule". Looking back on our experience of pain, we prefer a larger, longer amount to a shorter, smaller amount, just so long as the closing stages of the greater pain were easier to bear than the closing stages of the lesser one.  Galen Strawson  guardian.co.uk, Tuesday 13 December 2011 06.56 EST
  • 30. System 2 is slothful, and tires easily (a process called "ego depletion") – so it usually accepts what System 1 tells it. It's often right to do so, because System 1 is for the most part pretty good at what it does; it's highly sensitive to subtle environmental cues, signs of danger, and so on.  It kept our remote ancestors alive. Système 1 a ses raisons que Système 2 ne connaît point, as Pascal might have said.  It does, however, pay a high price for speed. It loves to simplify, to assume WYSIATI ("what you see is all there is"), even as it gossips and embroiders and confabulates.  It's hopelessly bad at the kind of statistical thinking often required for good decisions, it jumps wildly to conclusions and it's subject to a fantastic suite of irrational biases and interference effects (the halo effect, the "Florida effect", framing effects, anchoring effects, the confirmation bias, outcome bias, hindsight bias, availability bias, the focusing illusion, and so on).
  • 31. Behavioralists, Kahneman included, have been cataloging people’s systematic mistakes and nonlogical patterns for years. A few of the examples he cites:  1. Framing. Test subjects are more likely to opt for surgery if told that the “survival” rate is 90 percent, rather than that the mortality rate is 10 percent.  2. The sunk-cost fallacy. People seek to avoid feelings of regret; thus, they invest more money and time in a project with dubious results rather than give it up and admit they were wrong.  3. Loss aversion. In experiments, most subjects would prefer to receive a sure $46 than have a 50 percent chance of making $100.
  • 32. Kahneman is perhaps least persuasive in his treatment of the business world. Noting that even top performers in business—also sports—tend eventually to revert to the mean, he attributes success largely to luck.  This confuses events that may not be predictable with those that are determined by chance.  A high-achieving retail store, to cite one of his examples, is not lucky—it is well-situated. And if its sales later decline, that is not necessarily a sign that its prior success was random.  Business has a self-correcting cycle that fosters mean reversion. Success attracts competitors.
  • 33. Mr. Kahneman stresses that he is just as susceptible as the rest of us to the cognitive illusions he has discovered. He tries to recognize situations when mistakes are especially likely to occur—such as when he is starting a big project or making a forecast—and then act to rethink his System 1 inclinations.  The tendency to underestimate the costs of future projects, he notes, is susceptible to taking an "outside view": looking at your own project as an outsider would.  To avoid overconfidence, Mr. Kahneman recommends an exercise called the "premortem," developed by the psychologist Gary Klein: Before finalizing a decision, imagine that, a year after it has been made, it has turned out horribly, then write a history of how it went wrong and why.  By CHRISTOPHER F. CHABRIS a psychology professor at Union College and a co-author of "The Invisible Gorilla: And Other Ways Our Intuitions Deceive Us.  http://online.wsj.com/home-page
  • 34. A transitional moment linking the positive and the negative aspects of thinking fast illustrates why the author’s personality – and thus the book – is so engaging. Kahneman regards even the experts as prone to the mistakes of System 1 listed above, and cheerfully admits that he is no exception. But he wants to know whether this view can be reconciled with cases such as that of the firefighting captain. So he engages one of his vehement critics on this issue and they debate their way to a joint paper.  Their answer is that expertise can be learnt by prolonged exposure to situations that are “sufficiently regular to be predictable”, and in which the expert gets quick and decisive feedback on whether he did the right or the wrong thing.  Experts can thus train their unconscious “pattern recognition” mechanism to produce the right answer quickly.  So this certainly applies to chess, and it certainly does not apply to predicting the course of Middle East politics.
  • 35. One of Kahneman and Tversky’s most famous ideas is what they call prospect theory: our inclination to fear possible losses more than we value possible gains.  Would you take a bet on a one-time coin flip that paid $200 if you won but cost you $150 if you lost?  Most people wouldn’t, though it’s tilted in your favor.  Pro golfers tend to make a higher proportion of their putts when they’re trying to avoid a bogey (which would result in losing a stroke) than when they have a chance for a birdie (for the possible gain of a stroke); here, too, avoiding a failure is more crucial than achieving a triumph.
  • 36. If the brain is a “a machine for jumping to conclusions,” as Kahneman writes, it’s System 1 that yells “Geronimo!” Autopilot thinking explains the popular opinion that tornadoes kill more people than asthma, although in fact asthma kills 20 times as many:  We fixate on scenes from TV showing homes turned to matchsticks and overestimate the representativeness of such scenes.
  • 37. The first bit of mental machinery, which Kahneman blandly calls System One, works 24/7 to keep us out of trouble, while alerting us to fleeting opportunities. Appropriate for a species that is both predator and prey, System One lives in a world of snap judgments, extensible metaphors, ill- informed biases, and loosely constructed rules of thumb. We sometimes call this decision making apparatus intuition. Man’s intuition is sophisticated enough that it has helped us thrive across a variety of ever changing environments.  Despite its utility, System One is often wrong, especially if numbers are involved. For a trivial example, answer quickly: If the sum of the cost of a ball and bat is $1.10 and the bat cost a dollar more than the ball, what does the ball cost?  Your System One answer (most likely wrong) is good enough to avoid mistaking a hungry lion for a tasty chicken. But it’s not good enough to build an airplane or design an effective income tax code. (The answer is a nickel, not a dime.).  System Two is associated with enumeration, computation, objective analysis, and complex chains of logic. It is our rational brain. Kahneman’s work shows that even scientists like himself use System Two very sparingly, calling on it only when System One asks for help. In addition, in order to function at the highest levels, System Two requires training, discipline, concentration, skeptical and impartial evaluation of purported facts, and the ruthless elimination of contradictions.
  • 38. It afflicts us all. Because confidence in our own judgments is part of being human. By DANIEL KAHNEMAN Published: October 19, 2011
  • 39. Decades later, I can see many of the central themes of my thinking about judgment in that old experience. One of these themes is that people who face a difficult question often answer an easier one instead, without realizing it. We were required to predict a soldier’s performance in officer training and in combat, but we did so by evaluating his behavior over one hour in an artificial situation. This was a perfect instance of a general rule that I call WYSIATI, “What you see is all there is.” We had made up a story from the little we knew but had no way to allow for what we did not know about the individual’s future, which was almost everything that would actually matter. When you know as little as we did, you should not make extreme predictions like “He will be a star.” The stars we saw on the obstacle field were most likely accidental flickers, in which a coincidence of random events — like who was near the wall — largely determined who became a leader. Other events — some of them also random — would determine later success in training and combat.
  • 40. We often interact with professionals who exercise their judgment with evident confidence, sometimes priding themselves on the power of their intuition. In a world rife with illusions of validity and skill, can we trust them? How do we distinguish the justified confidence of experts from the sincere overconfidence of professionals who do not know they are out of their depth? We can believe an expert who admits uncertainty but cannot take expressions of high confidence at face value. As I first learned on the obstacle field, people come up with coherent stories and confident predictions even when they know little or nothing. Overconfidence arises because people are often blind to their own blindness.
  • 41. Some cognitive biases, of course, are flagrantly exhibited even in the most natural of settings. Take what Kahneman calls the “planning fallacy”: our tendency to overestimate benefits and underestimate costs, and hence foolishly to take on risky projects.  In 2002, Americans remodeling their kitchens, for example, expected the job to cost $18,658 on average, but they ended up paying $38,769.
  • 42. The planning fallacy is “only one of the manifestations of a pervasive optimistic bias,” Kahneman writes, which “may well be the most significant of the cognitive biases.” Now, in one sense, a bias toward optimism is obviously bad, since it generates false beliefs — like the belief that we are in control, and not the playthings of luck.  But without this “illusion of control,” would we even be able to get out of bed in the morning?  Optimists are more psychologically resilient, have stronger immune systems, and live longer on average than their more reality-based counterparts.
  • 43. Biases in the evaluation of compound events are particularly significant in the context of planning. The successful completion of an undertaking, such as the development of a new product, typically has a conjunctive character: for the undertaking to succeed, each of a series of events must occur. Even when each of these events is very likely, the overall probability of success can be quite low if the number of events is large.
  • 44. In a series of experiments, Ellen Langer (1975) demonstrated first the prevalence of the illusion of control and second, that people were more likely to behave as if they could exercise control in a chance situation where “skill cues” were present.  By skill cues, Langer meant properties of the situation more normally associated with the exercise of skill, in particular the exercise of choice, competition, familiarity with the stimulus and involvement in decisions.  One simple form of this fallacy is found in casinos: when rolling dice in craps, it has been shown that people tend to throw harder for high numbers and softer for low numbers.  Under some circumstances, experimental subjects have been induced to believe that they could affect the outcome of a purely random coin toss. Subjects who guessed a series of coin tosses more successfully began to believe that they were actually better guessers, and believed that their guessing performance would be less accurate if they were distracted.
  • 45. Drawing on a vast body of research, Lears ranges through the entire sweep of American history as he uncovers the hidden influence of risk taking, conjuring, soothsaying, and sheer dumb luck on our culture, politics, social lives, and economy. T.J. Jackson Lears “Something for Nothing” (2003)
  • 46. Moreover, as Kahneman notes, exaggerated optimism serves to protect both individuals and organizations from the paralyzing effects of another bias, “loss aversion”: our tendency to fear losses more than we value gains.  It was exaggerated optimism that John Maynard Keynes had in mind when he talked of the “animal spirits” that drive capitalism.
  • 47. "losses loom larger than corresponding gains”  "In prospect theory, loss aversion refers to the tendency for people to strongly prefer avoiding losses than acquiring gains. Some studies suggest that losses are as much as twice as psychologically powerful as gains. Loss aversion was first convincingly demonstrated by Amos Tversky and Daniel Kahneman.”
  "The principle of loss aversion was first introduced by Kahneman and Tversky (1979)"  Tversky and Kahneman (1991) "The central assumption of the theory is that losses and disadvantages have greater impact on preferences than gains and advantages.”  "Numerous studies have shown that people feel losses more deeply than gains of the same value (Kahneman and Tversky 1979, Tversky and Kahneman 1991)."
Goldberg and von Nitzsch (1999) pages 97-98  "Both the status quo bias and the endowment effect are part of a more general issue known as loss aversion." (Montier 2007, p. 32)
  • 48. the inclination to see events that have occurred as more predictable than they in fact were before they took place. Hindsight bias has been demonstrated experimentally in a variety of settings, including politics, games and medicine.  In psychological experiments of hindsight bias, subjects also tend to remember their predictions of future events as having been stronger than they actually were, in those cases where those predictions turn out correct.  Prophecy that is recorded after the fact is an example of hindsight bias, given its own rubric, as vaticinium ex eventu. foretelling after the event  One explanation of the bias is the availability heuristic: the event that did occur is more salient in one's mind than the possible outcomes that did not.  It has been shown that examining possible alternatives may reduce the effects of this bias.
  • 49. The framing of alternatives also affects decisions.  For example, when people (including doctors) who are considering a risky medical procedure are told that 90 percent survive five years, they are far more likely to accept the procedure than when they are told that 10 percent do not survive five years.  Because framing affects people's behavior, providing more information cannot remedy matters, unless the information is presented in a fully neutral fashion.  In some cases, additional information only increases people's anxiety and confusion, thereby reducing their welfare.
  • 50. A heuristic (hyu̇-ris-tik) is a method to help solve a problem, commonly informal. It is particularly used for a method that often rapidly leads to a solution that is usually reasonably close to the best possible answer.  Heuristics are "rules of thumb” educated guesses, intuitive judgments or simply common sense.  In more precise terms, heuristics stand for strategies using readily accessible, though loosely applicable, information to control problem-solving in human beings and machines.
  • 51. “There is always a well-known solution to every human problem – neat, plausible, and wrong.”  H. L. Mencken, Prejudices: Second Series, 1920 Tailors' Rule of Thumb. This is the fictional rule described by Jonathan Swift in his satirical novel Gulliver's Travels: Then they measured my right Thumb, and desired no more; for by a mathematical Computation, that twice round the Thumb is once around the Wrist, and so on to the Neck and Waist, and by the help of my old Shirt, which I displayed on the Ground before them for a Pattern, they fitted me exactly."
  • 52. Rules of Thumb  People rely on a limited number of heuristic (essentially rules of thumbs) principles which reduce the complex tasks of assessing probabilities and predicting values to simpler judgmental operations.  In general, these heuristics are quite useful, but sometimes they lead to severe and systematic errors.  Daniel Kahneman
  • 53. Financial - Rule of 72 A rule of thumb for exponential growth at a constant rate. An approximation of the "doubling time" formula used in population growth, which says divide 70 by the percent growth rate (the actual number is 69.3147181 from the natural logarithm of 2, if the percent growth is much much less than 1%). In terms of money, it is frequently easier to use 72 (rather than 70) because it works better in the 4%-10% range where interest rates often lie. Therefore, divide 72 by the percent interest rate to determine the approximate amount of time to double your money in an investment. For example, at 8% interest, your money will double in approximately 9 years (72/8 = 9).  Tailors' Rule of Thumb A simple approximation that was used by tailors to determine the wrist, neck, and waist circumferences of a person through one single measurement of the circumference of that person's thumb. The rule states, typically, that twice the circumference of a person's thumb is the circumference of their wrist, twice the circumference of the wrist is the circumference of the neck, and twice around the neck is the person's waist. For example, if the circumference of the thumb is 4 inches, then the wrist circumference is 8 inches, the neck is 16 and the waist is 32. An interesting consequence of this is that — for those to whom the rule applies — this simple method can be used to determine if trousers will fit: the trousers are wrapped around the neck, and if the two ends barely touch, then they will fit. Any overlap or lack thereof corresponds to the trousers being too loose or tight, respectively.  Marine Navigation A ship's captain should navigate to keep the ship more than a thumb's width from the shore, as shown on the nautical chart being used. Thus, with a coarse scale chart, that provides few details of near shore hazards such as rocks, a thumb's width would represent a great distance, and the ship would be steered far from shore; whereas on a fine scale chart, in which more detail is provided, a ship could be brought closer to shore.[1]  Statistics The Statistical Rule of Thumb says that for most large data sets, 68% of data points will occur within one standard deviation from the mean, and 95% will occur within two standard deviations. Chebyshev's inequality is a more general rule along these same lines and applies to all data sets.
  • 54. In an effort that spanned several years, we attempted to answer one basic question:  Under what conditions are the intuitions of professionals worthy of trust?  We do not claim that the conclusions we reached are surprising (many were anticipated by Shanteau, 1992, Hogarth, 2001, and Myers, 2002, among others), but we believe that they add up to a coherent view of expert intuition, which is more than we expected to achieve when we began.
  • 55. The NDM approach, which focuses on the successes of expert intuition, grew out of early research on master chess players conducted by deGroot (1946/1978) and later by Chase and Simon (1973). DeGroot showed that chess grand masters were generally able to identify the most promising moves rapidly, while mediocre chess players often did not even consider the best moves.  The chess grand masters mainly differed from weaker players in their unusual ability to appreciate the dynamics of complex positions and quickly judge a line of play as promising or fruitless.  Chase and Simon (1973) described the performance of chess experts as a form of perceptual skill in which complex patterns are recognized. They estimated that chess masters acquire a repertoire of 50,000 to 100,000 immediately recognizable patterns, and that this repertoire enables them to identify a good move without having to calculate all possible contingencies.  Strong players need a decade of serious play to assemble this large collection of basic patterns, but of course they achieve impressive levels of skill even earlier.  On the basis of this work, Simon defined intuition as the recognition of patterns stored in memory.
  • 56. Kahneman read Meehl’s book in 1955 while serving in the Psychological Research Unit of the Israel Defense Forces, and the book helped him make sense of his own encounters with the difficulties of clinical judgment.  One of Kahneman’s duties was to assess candidates for officer training, using field tests and other observations as well as a personal interview.  Kahneman (2003) described the powerful sense of getting to know each candidate and the accompanying conviction that he could foretell how well the candidate would do in further training and eventually in combat.  The subjective conviction of understanding each case in isolation was not diminished by the statistical feedback from officer training school, which indicated that the validity of the assessments was negligible.  Kahneman coined the term illusion of validity for the unjustified sense of confidence that often comes with clinical judgment. His early experience with the fallibility of intuitive impressions could hardly be more different from Klein’s formative encounter with the successful decision making of fire-fighting ground commanders.
  • 57. Our starting point is that intuitive judgments can arise from genuine skill—the focus of the Naturalistic Decision Making (NDM) approach—but that they can also arise from inappropriate application of the heuristic processes on which students of the Heuristics Based tradition have focused.
  • 58. Skilled judges are often unaware of the cues that guide them, and individuals whose intuitions are not skilled are even less likely to know where their judgments come from.
  • 59. True experts, it is said, know when they don’t know.  However, non-experts (whether or not they think they are) certainly do not know when they don’t know.  Subjective confidence is therefore an unreliable indication of the validity of intuitive judgments and decisions.
  • 60. The determination of whether intuitive judgments can be trusted requires an examination of the environment in which the judgment is made and of the opportunity that the judge has had to learn the regularities of that environment.
  • 61. We describe task environments as “high-validity” if there are stable relationships between objectively identifiable cues and subsequent events or between cues and the outcomes of possible actions.  Medicine and firefighting are practiced in environments of fairly high validity.  In contrast, outcomes are effectively unpredictable in zero-validity environments.  To a good approximation, predictions of the future value of individual stocks and long-term forecasts of political events are made in a zero-validity environment.
  • 62. Validity and uncertainty are not incompatible. Some environments are both highly valid and substantially uncertain.  Poker and warfare are examples. The best moves in such situations reliably increase the potential for success.
  • 63. An environment of high validity is a necessary condition for the development of skilled intuitions.  Other necessary conditions include adequate opportunities for learning the environment (prolonged practice and feedback that is both rapid and unequivocal).  If an environment provides valid cues and good feedback, skill and expert intuition will eventually develop in individuals of sufficient talent.
  • 64. Although true skill cannot develop in irregular or unpredictable environments, individuals will sometimes make judgments and decisions that are successful by chance.  These “lucky” individuals will be susceptible to an illusion of skill and to overconfidence (Arkes, 2001).  The financial industry is a rich source of examples.
  • 65. The situation that we have labeled fractionation of skill is another source of overconfidence. Professionals who have expertise in some tasks are sometimes called upon to make judgments in areas in which they have no real skill.  (For example, financial analysts may be skilled at evaluating the likely commercial success of a firm, but this skill does not extend to the judgment of whether the stock of that firm is underpriced.)  It is difficult both for the professionals and for those who observe them to determine the boundaries of their true expertise.
  • 66. We agree that the weak regularities available in low-validity situations can sometimes support the development of algorithms that do better than chance. These algorithms only achieve limited accuracy, but they outperform humans because of their advantage of consistency.  However, the introduction of algorithms to replace human judgment is likely to evoke substantial resistance and sometimes has undesirable side effects.
  • 67. The Drunkard’s Walk – Functional magnetic resonance imaging, for example, shows that risk and reward are assessed by parts of the dopaminergic system. – A brain-reward circuit important for motivational and emotional processes • The images show, too, that the amygdala, which is also linked to our emotional state, especially fear, is activated when we make decisions couched in uncertainty.
  • 68. The Drunkard’s Walk • Fortune is fair in potentialities, she is not fair in outcomes. • pp 13 • When we look at extraordinary accomplishments in sports-- -or elsewhere—we should keep in mind that extraordinary events can happen without extraordinary causes. • Random events often look like nonrandom events, and in interpreting human affairs we must take care not to confuse the two. pp20
  • 69. Selection Bias • This bias makes us miscompute the odds and wrongly ascribe skills. If you funded 1,000,000 unemployed people endowed with no more than the ability to say "buy" or "sell", odds are that you will break-even in the aggregate, minus transaction costs, but a few will hit the jackpot, simply because the base cohort is very large. It will be almost impossible not to have small Warren Buffets by luck alone. After the fact they will be very visible and will derive precise and well-sounding explanations about why they made it. It is difficult to argue with them; "nothing succeeds like success". All these retrospective explanations are pervasive, but there are scientific methods to correct for the bias. This has not filtered through to researchers have the business world or the news media; evidence that professional fund managers are just no better than random and cost money to society (the total revenues from these transaction costs is in the hundreds of billion of dollars) but the public will remain convinced that "some" of these investors have skills.
  • 70. False Discovery Rate • July 13, 2008 • STRATEGIES • The Prescient Are Few • By MARK HLBERT • HOW many mutual fund managers can consistently pick stocks that outperform the broad stock market averages — as opposed to just being lucky now and then? • Countless studies have addressed this question, and have concluded that very few managers have the ability to beat the market over the long term. Nevertheless, researchers have been unable to agree on how small that minority really is, and on whether it makes sense for investors to try to beat the market by buying shares of actively managed mutual funds. • A new study builds on this research by applying a sensitive statistical test borrowed from outside the investment world. It comes to a rather sad conclusion: There was once a small number of fund managers with genuine market-beating abilities, as judged by having past performance so good that their records could not be attributed to luck alone. But virtually none remain today. Index funds are the only rational alternative for almost all mutual fund investors, according to the study’s findings. • The study, “False Discoveries in Mutual Fund Performance: Measuring Luck in Estimating Alphas,” has been circulating for over a year in academic circles. Its authors are Laurent Barras, a visiting researcher at Imperial College’s Tanaka Business School in London; Olivier Scaillet, a professor of financial econometrics at the University of Geneva and the Swiss Finance Institute; and Russ Wermers, a finance professor at the University of Maryland. • The statistical test featured in the study is known as the “False Discovery Rate,” and is used in fields as diverse as computational biology and astronomy. In effect, the method is designed to simultaneously avoid false positives and false negatives — in other words, conclusions that something is statistically significant when it is entirely random, and the reverse. • Both of those problems have plagued previous studies of mutual funds, Professor Wermers said. The researchers applied the method to a database of actively managed domestic equity mutual funds from the beginning of 1975 through 2006. To ensure that their results were not biased by excluding funds that have gone out of business over the years, they included both active and defunct funds. They excluded any fund with less than five years of performance history. All told, the database contained almost 2,100 funds. • The researchers found a marked decline over the last two decades in the number of fund managers able to pass the False Discovery Rate test. If they had focused only on managers running funds in 1990 and their records through that year, for example, the researchers would have concluded that 14.4 percent of managers had genuine stock- picking ability. But when analyzing their entire fund sample, with records through 2006, this proportion was just 0.6 percent — statistically indistinguishable from zero, according to the researchers. • This doesn’t mean that no mutual funds have beaten the market in recent years, Professor Wermers said. Some have done so repeatedly over periods as short as a year or two. But, he added, “the number of funds that have beaten the market over their entire histories is so small that the False Discovery Rate test can’t eliminate the possibility that the few that did were merely false positives” — just lucky, in other words. • Professor Wermers says he was surprised by how rare stock-picking skill has become. He had “generally been positive about the existence of fund manager ability,” he said, but these new results have been a “real shocker.” • WHY the decline? Professor Wermers says he and his co-authors suspect various causes. One is high fees and expenses. The researchers’ tests found that, on a pre-expense basis, 9.6 percent of mutual fund managers in 2006 showed genuine market-beating ability — far higher than the 0.6 percent after expenses were taken into account. This suggests that one in 10 managers may still have market- beating ability. It’s just that they can’t come out ahead after all their funds’ fees and expenses are paid. • Another possible factor is that many skilled managers have gone to the hedge fund world. Yet a third potential reason is that the market has become more efficient, so it’s harder to identify undervalued or overvalued stocks. Whatever the causes, the investment implications of the study are the same: buy and hold an index fund benchmarked to the broad stock market. • Professor Wermers says his advice has evolved significantly as a result of this study. Until now, he says, he wouldn’t have tried to discourage a sophisticated investor from trying to pick a mutual fund that would outperform the market. Now, he says, “it seems almost hopeless.” • Mark Hulbert is editor of The Hulbert Financial Digest, a service of MarketWatch. E-mail: strategy@nytimes.com. •
  • 71. Bent Flyvbjerg  Nils Bruzelius  Werner Rothengatter
  • 72. To begin, the book identifies a common feature of the conventional megaprojects development, that is, despite the overwhelming costs overrun, below- projection revenue, and strikingly poor performance records in terms of economy, environment and public support, megaprojects grow continuously in number and scale around the world, forming the so-called megaprojects paradox.  Understanding of this problem and its consequences are explored in the first six chapters, which document the cost overruns, the demand over forecasts, and viability inflation of major megaprojects.
  • 73. In 9 out of 10 transport infrastructure projects, costs are underestimated, resulting in cost overrun;  For rail, actual costs are, on the average, 45% higher than estimated costs (standard deviation, S.D. = 38);  For fixed links (tunnels and bridges), actual costs are, on the average, 34% higher than estimated costs (S.D. = 62);  For roads, actual costs are, on the average, 20% higher than estimated costs (S.D. = 30);  For all project types, actual costs are, on the average, 28% higher than estimated costs (S.D. = 39);  Cost underestimation and overrun exist across 20 nations and five continents; it appears to be a global phenomenon;  Cost underestimation and overrun appear to be more pronounced in developing nations than in North America and Europe (data for rail only);  Cost underestimation and overrun have not decreased over the past 70 years. No learning seems to take place;  Cost underestimation and overrun cannot be explained by error and seem to be best explained by strategic misrepresentation, namely, lying, with a view to getting projects started [5].
  • 74. The authors cite as a core factor, underlying all of these dreadful situations with megaprojects, the lack of transparency in decision making and the weak involvement of the civil society, or what they call a "democracy deficit."  They make a big point of the failure to pay attention to risks and the lack of accountability in the project decision- making processes as a main source of difficulties.  The authors' concept of risk is almost entirely limited to financial risk.
  • 75. Heuristic algorithms are often employed because they may be seen to "work" without having been mathematically proven to meet a given set of requirements.  Great care must be given when employing a heuristic algorithm. One common pitfall in implementing a heuristic method to meet a requirement comes when the engineer or designer fails to realize that the current data set doesn't necessarily represent future system states.  While the existing data can be pored over and an algorithm can be devised to successfully handle the current data, it is imperative to ensure that the heuristic method employed is capable of handling future data sets.  This means that the engineer or designer must fully understand the rules that generate the data and develop the algorithm to meet those requirements and not just address the current data sets.  A simple example of how heuristics can fail is to answer the question "What is the next number in this sequence: 1, 2, 4?". One heuristic algorithm might say that the next number is 8 because the numbers are doubling - leading to a sequence like 1, 2, 4, 8, 16, 32... Another, equally valid, heuristic would say that the next number is 7 because each number is being raised by one higher interval than the one before - leading to a series that looks like 1, 2, 4, 7, 11, 16...  Statistical analysis must be conducted when employing heuristics to ensure that enough data points are utilized to make incorrect outcomes statistically insignificant
  • 76.
  • 77. A major problem with expert estimates is overconfidence. To overcome this, Hubbard advocates using calibrated probability assessments to quantify analysts’ abilities to make estimates. Calibration assessments involve getting analysts to answer trivia questions and eliciting confidence intervals for each answer. The confidence intervals are then checked against the proportion of correct answers. Essentially, this assesses experts’ abilities to estimates by tracking how often they are right. It has been found that people can improve their ability to make subjective estimates through calibration training – i.e. repeated calibration testing followed by feedback.
  • 78. The Dunning–Kruger effect is a cognitive bias in which unskilled people make poor decisions and reach erroneous conclusions, but their incompetence denies them the metacognitive ability to recognize their mistakes.[1]  The unskilled therefore suffer from illusory superiority, rating their ability as above average, much higher than it actually is, while the highly skilled underrate their own abilities, suffering from illusory inferiority.  Actual competence may weaken self-confidence, as competent individuals may falsely assume that others have an equivalent understanding.  As Kruger and Dunning conclude, "the miscalibration of the incompetent stems from an error about the self, whereas the miscalibration of the highly competent stems from an error about others" (p. 1127).[2] The effect is about paradoxical defects in cognitive ability, both in oneself and as one compares oneself to others.
  • 79.
  • 80.
  • 81.
  • 83. Daniel Kahneman,1 Alan B. Krueger,1,2* David Schkade,3 Norbert Schwarz,4 Arthur A. Stone5  The belief that high income is associated with good mood is widespread but mostly illusory.  People with above-average income are relatively satisfied with their lives but are barely happier than others in moment-to-moment experience, tend to be more tense, and do not spend more time in particularly enjoyable activities.  Moreover, the effect of income on life satisfaction seems to be transient. We argue that people exaggerate the contribution of income to happiness because they focus, in part, on conventional achievements when evaluating their life or the lives of others.  1 Princeton University, Princeton, NJ 08544, USA. 2 National Bureau of Economic Research, Cambridge, MA 02138, USA. 3 Rady School of Management, University of California, San Diego, San Diego, CA 92093, USA. 4 Department of Psychology, University of Michigan, Ann Arbor, MI 48106, USA. 5 Stony Brook University, Stony Brook, NY, 11794, USA.
  • 84. Proceedings of the National Academy of Sciences  Preferences and choices can affect long-term happiness.  German Study ▪ Marry someone who is not neurotic ▪ Focus more on friends, less on material goods ▪ Get involved with making the world a better place ▪ Have a job but also enough time for leisure ▪ Stay physically active ▪ For men, don’t be underweight. For women, don’t be obese.
  • 85. Don't worry, be happy" may be more than just a wishful mantra. A new study finds that people's happiness levels can change substantially over their lifetimes, suggesting that happiness isn't predetermined by genes or personality.  Psychologists have long argued that people have a "set point" for happiness. Regardless of what life brings, the set-point theory goes, happiness levels tend to be stable. A big life event could create a boost of joy or a crush of sorrow, but within a few years, people return to a predetermined level of life satisfaction, according to the theory.  The new study, which used a nationally representative sample of almost 150,000 German adults, finds the opposite. People's long-term life satisfaction can change, the researchers report today (Oct. 4) in the online early edition of the Proceedings of the National Academy of Sciences. In fact, a substantial number of people followed over 25 years saw their happiness levels shift by one-third or more.
  • 86. The study also echoed previous happiness research in finding that money doesn't buy happiness.  "People with a lot of money are more satisfied with their lives... but mainly due to the more interesting and challenging jobs they have," study author Gert Wagner, a researcher at the Max Planck Institute for Human Development in Germany, told LiveScience. "Money is simply a byproduct of good and satisfying jobs. If you want to be satisfied with your life, you must spend time with your friends and your family."  Wagner said that previous work suggests findings on happiness from one developed country, like Germany, should also hold true for another, such as the United States. In fact, a study in May found that in the United States. happiness tends to increase with age.
  • 87. Happiness Study  I'm happier than you  The researchers used data from a study of German adults spanning from 1984 to 2008. Each year, the participants answered questions on their life satisfaction, life goals and other measures like how much they exercise and socialize.  By averaging life-satisfaction responses to even out any short-term effects, the researchers plotted out the respondents' happiness by percentiles. Someone in the 99th percentile, for example, would be happier than 99 percent of the study participants.  People shifted in the rankings — and thus in their levels of happiness — quite a bit. Just over 38 percent changed their position in the distribution by 25 percentiles or more during the study period. About 25 percent changed by 33.3 percentiles or more, and 11.8 percent changed by 50 percentiles.  Feel-good factors  So what contributed to long-term happiness? The researchers found several correlations between life choices and life satisfaction:  Marry well: The personality traits of partners influenced people's happiness. Neuroticism, or a tendency toward anxiety, emotional instability and depression, was most influential. People who married or partnered with neurotic people were less likely to be happy than people who married non-neurotic types.  Focus on the family: People who assigned relatively high value to altruistic and family goals compared with career goals were happier. Women were also happier when their male partners ranked family goals high.  Go to church: People who went to church more often were happier, though the study can't determine whether the happiness is related to religious views or to the social circle religious organizations offer.  Work, but not too much (or too little): People's happiness matched how well they felt their work hours matched their desired work hours. In other words, people who worked more or fewer hours than they preferred were less happy. Working less or being unemployed was worse than working too much, presumably because underemployment is a financial blow, the researchers wrote.  Get social, and get moving: Social interaction and exercise were both associated with happiness. Working out made people happier regardless of body weight. The only correlation between body weight and happiness was that underweight men and obese women were more likely to be unhappy.
  • 88. Happiness Study  Mysteries of happiness  "In its extreme form, set-point theory was never credible," Daniel Kahneman, an emeritus professor of psychology at Princeton University and the winner of the 2002 Nobel Prize in Economic Sciences, told LiveScience. "If it was taken to mean that the only factor that determines happiness or life satisfaction is genetic, so that people always come back to exactly to the same point, this was utterly incredible."  The current study is a useful demonstration that life changes can influence people's life satisfaction, said Kahneman, who was not involved in the research. However, the correlations between certain goals and traits and happiness doesn't necessarily answer the nature-versus-nurture question.  "They're suggesting that the goals are chosen. But the goals may be part of personality," and thus partially genetic, he said. "The fact that goals matter, like altruism and materialism, that really doesn't help us distinguish between personality and circumstances."  More studies are needed that track large populations of people after influential changes, like the enactment of new laws, said Andrew Oswald, a professor of behavioral science at the University of Warwick who studies happiness but was not involved in the current study. By comparing people who lived under, say, a new state tax law that affected income to those who lived in a nearby state without the law, researchers could begin to look at happiness in a more experimental way, he said.  "The key thing is that life events good and bad do shape happiness over long periods," Oswald said. "We are, in part, the product of our experiences. It's not all born into us."  7 Ways the Mind and Body Change with Age  Happiness is Being Old, Male and Republican  10 Ways to Keep Your Mind Sharp

Editor's Notes

  1. http://www.nytimes.com/pages/books/review/index.html
  2. NY TimesTwo Brains RunningBy JIM HOLTPublished: November 25, 2011
  3. • Galen Strawson'sSelves: An Essay in Revisionary Metaphysics is published by Oxford University Press.
  4. Galen Strawsonguardian.co.uk, Tuesday 13 December 2011 06.56 EST
  5. http://www.businessweek.com/
  6. Book Review: Thinking, Fast and Slow by Daniel Kahneman http://www.businessweek.com/
  7. —Mr. Chabris is a psychology professor at Union College and a co-author of "The Invisible Gorilla: And Other Ways Our Intuitions Deceive Us. http://online.wsj.com/home-pageBy CHRISTOPHER F. CHABRIS
  8. Thinking, Fast and SlowReview by William EasterlyWhy even experts must rely on intuition and often get it wrongTs&Cs and Copyright Policy for more detail. Email ftsales.support@ft.com to buy additional rights. http://www.ft.com/cms/s/2/15bb6522-04ac-11e1-91d9-00144feabdc0.html#ixzz1mTqFzlCW
  9. bookworld@washpost.comChristopher Shea writes the Ideas Market blog and Week in Ideas column for the Wall Street Journal.http://www.washingtonpost.com/entertainment
  10. Bill Frezza, Contributor