-
Notifications
You must be signed in to change notification settings - Fork 1
/
map_and_territory.tex
7862 lines (5979 loc) · 303 KB
/
map_and_territory.tex
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
\part{Map and Territory}
\chapter{Predictably Wrong}
\mysection{What Do I Mean By ``Rationality''?}
I mean:
\begin{enumerate}
\item {
\textbf{Epistemic rationality}: systematically improving the
accuracy of your beliefs.}
\item {
\textbf{Instrumental rationality}: systematically achieving your
values.}
\end{enumerate}
When you open your eyes and look at the room around you,
you'll locate your laptop in relation to the table, and
you'll locate a bookcase in relation to the wall. If
something goes wrong with your eyes, or your brain, then your mental
model might say there's a bookcase where no bookcase
exists, and when you go over to get a book, you'll be
disappointed.
This is what it's like to have a false belief, a
map of the world that doesn't correspond to the
territory. Epistemic rationality is about building accurate maps
instead. This correspondence between belief and reality is commonly
called ``truth,'' and
I'm happy to call it that.
Instrumental rationality, on the other hand, is about
\textit{steering} reality---sending the future where you want it to go.
It's the art of choosing actions that lead to outcomes
ranked higher in your preferences. I sometimes call this
``winning.''
So rationality is about forming true beliefs and making winning
decisions.
Pursuing ``truth'' here
doesn't mean dismissing uncertain or indirect evidence.
Looking at the room around you and building a mental map of it
isn't different, in principle, from believing that the
Earth has a molten core, or that Julius Caesar was bald. Those
questions, being distant from you in space and time, might seem more
airy and abstract than questions about your bookcase. Yet there are
facts of the matter about the state of the Earth's core
in 2015 CE and about the state of Caesar's head in 50
BCE. These facts may have real effects upon you even if you never find
a way to meet Caesar or the core face-to-face.
And ``winning'' here need not
come at the expense of others. The project of life can be about
collaboration or self-sacrifice, rather than about competition.
``Your values'' here means
\textit{anything you care about}, including other people. It
isn't restricted to \textit{selfish} values or
\textit{unshared} values.
When people say ``X is
rational!'' it's usually just a more
strident way of saying ``I think X is
true'' or ``I think X is
good.'' So why have an additional word for
``rational'' as well as
``true'' and
``good''?
An analogous argument can be given against using
``true.'' There is no need to say
``it is true that snow is white''
when you could just say ``snow is
white.'' What makes the idea of truth useful is that
it allows us to talk about the general features of map-territory
correspondence. ``True models usually produce better
experimental predictions than false models'' is a
useful generalization, and it's not one you can make
without using a concept like
``true'' or
``accurate.''
Similarly, ``Rational agents make decisions that
maximize the probabilistic expectation of a coherent utility
function'' is the kind of thought that depends on a
concept of (instrumental) rationality, whereas
``It's rational to eat
vegetables'' can probably be replaced with
``It's useful to eat
vegetables'' or
``It's in your interest to eat
vegetables.'' We need a concept like
``rational'' in order to note
general facts about those ways of thinking that systematically produce
truth or value---and the systematic ways in which we fall short of
those standards.
Sometimes experimental psychologists uncover human reasoning that
seems very strange. For example, someone rates the probability
``Bill plays jazz'' as \textit{less}
than the probability ``Bill is an accountant who plays
jazz.'' This seems like an odd judgment, since any
particular jazz-playing accountant is obviously a jazz player. But to
what higher vantage point do we appeal in saying that the judgment is
\textit{wrong}?
Experimental psychologists use two gold standards:
\textit{probability theory}, and \textit{decision theory}.
Probability theory is the set of laws underlying rational belief.
The mathematics of probability describes equally and without
distinction (a) figuring out where your bookcase is, (b) figuring out
the temperature of the Earth's core, and (c) estimating
how many hairs were on Julius Caesar's head.
It's all the same problem of how to process the
evidence and observations to revise
(``update'') one's
beliefs. Similarly, decision theory is the set of laws underlying
rational action, and is equally applicable regardless of what
one's goals and available options are.
Let ``$P(\text{such-and-such})$'' stand
for ``the probability that such-and-such
happens,'' and $P(A,B)$ for ``the
probability that both $A$ and $B$ happen.'' Since it is a
universal law of probability theory that $P(A) \geq P(A,B)$, the
judgment that $P$(Bill plays jazz) is less than $P$(Bill plays jazz, Bill
is an accountant) is labeled incorrect.
To keep it technical, you would say that this probability judgment
is \textit{non-Bayesian}. Beliefs and actions that are rational in this
mathematically well-defined sense are called
``Bayesian.''
Note that the modern concept of rationality is not about reasoning
in words. I gave the example of opening your eyes, looking around you,
and building a mental model of a room containing a bookcase against the
wall. The modern concept of rationality is general enough to include
your eyes and your brain's visual areas as
things-that-map. It includes your wordless intuitions as well. The math
doesn't care whether we use the same English-language
word, ``rational,'' to refer to
Spock and to refer to Bayesianism. The math models good ways of
achieving goals or mapping the world, regardless of whether those ways
fit our preconceptions and stereotypes about what
``rationality'' is supposed to be.
This does not quite exhaust the problem of what is meant in
practice by ``rationality,'' for two
major reasons:
First, the Bayesian formalisms in their full form are
computationally intractable on most real-world problems. No one can
\textit{actually} calculate and obey the math, any more than you can
predict the stock market by calculating the movements of quarks.
This is why there is a whole site called ``Less
Wrong,'' rather than a single page that simply states
the formal axioms and calls it a day. There's a whole
further art to finding the truth and accomplishing value \textit{from
inside a human mind}: we have to learn our own flaws, overcome our
biases, prevent ourselves from self-deceiving, get ourselves into good
emotional shape to confront the truth and do what needs doing, et
cetera, et cetera.
{
Second, sometimes the meaning of the math itself is called into
question. The exact rules of probability theory are called into
question by, e.g., anthropic problems in which the number of observers
is uncertain. The exact rules of decision theory are called into
question by, e.g., Newcomblike problems in which other agents may
predict your decision before it happens.\footnote{\textbf{Editor's Note:} For a good introduction
to Newcomb's Problem, see Holt.\footnotemark \ More
generally, you can find definitions and explanations for many of the
terms in this book at the website
wiki.lesswrong.com/wiki/RAZ\_Glossary.\comment{1}}\footnotetext{Jim Holt, ``Thinking Inside the
Boxes,'' \textit{Slate} (2002),
\url{http://www.slate.com/articles/arts/egghead/2002/02/thinkinginside\%5C\_the\%5C\_boxes.single.html}.\comment{2}}}
In cases like these, it is futile to try to settle the problem by
coming up with some new definition of the word
``rational'' and saying,
``Therefore my preferred answer, \textit{by
definition,} is what is meant by the word
`rational.'\,'' This
simply raises the question of why anyone should pay attention to your
definition. I'm not interested in probability theory
because it is the holy word handed down from Laplace.
I'm interested in Bayesian-style belief-updating (with
Occam priors) because I expect that this style of thinking gets us
systematically closer to, you know, \textit{accuracy}, the map that
reflects the territory.
And then there are questions of how to think that seem not quite
answered by either probability theory or decision theory---like the
question of how to feel about the truth once you have it. Here, again,
trying to define ``rationality'' a
particular way doesn't support an answer, but merely
presumes one.
I am not here to argue the meaning of a word, not even if that
word is ``rationality.'' The point
of attaching sequences of letters to particular concepts is to let two
people \textit{communicate}{}---to help transport thoughts from one
mind to another. You cannot change reality, or prove the thought, by
manipulating which meanings go with which words.
{
So if you understand what concept I am \textit{generally getting
at} with this word ``rationality,''
and with the sub-terms ``epistemic
rationality'' and ``instrumental
rationality,'' we \textit{have communicated}: we have
accomplished everything there is to accomplish by talking about how to
define ``rationality.''
What's left to discuss is not \textit{what meaning} to
attach to the syllables
``ra-tion-al-i-ty'';
what's left to discuss is \textit{what is a good way to
think}.}
If you say, ``It's
(epistemically) rational for me to believe $X$, but the truth is
$Y$,'' then you are probably using the word
``rational'' to mean something other
than what I have in mind. (E.g.,
``rationality'' should be
\textit{consistent under
reflection}{}---``rationally''
looking at the evidence, and
``rationally'' considering how your
mind processes the evidence, shouldn't lead to two
different conclusions.)
Similarly, if you find yourself saying, ``The
(instrumentally) rational thing for me to do is $X$, but the right thing
for me to do is $Y$,'' then you are almost certainly
using some other meaning for the word
``rational'' or the word
``right.'' I use the term
``rationality''
\textit{normatively}, to pick out desirable patterns of thought.
In this case---or in any other case where people disagree about
word meanings---you should substitute more specific language in place
of ``rational'':
``The self-benefiting thing to do is to run away, but
I hope I would at least try to drag the child off the railroad
tracks,'' or ``Causal decision
theory as usually formulated says you should two-box on
Newcomb's Problem, but I'd rather have
a million dollars.''
In fact, I recommend reading back through this essay, replacing
every instance of ``rational'' with
``foozal,'' and seeing if that
changes the connotations of what I'm saying any. If so,
I say: strive not for rationality, but for foozality.
The word ``rational'' has
potential pitfalls, but there are plenty of \textit{non}{}-borderline
cases where ``rational'' works fine
to communicate what I'm getting at. Likewise
``irrational.'' In these cases
I'm not afraid to use it.
Yet one should be careful not to \textit{overuse} that word. One
receives no points merely for pronouncing it loudly. If you speak
overmuch of the Way, you will not attain it.
\myendsectiontext
\bigskip
\mysection{Feeling Rational}
A popular belief about
``rationality'' is that rationality
opposes all emotion---that all our sadness and all our joy are
automatically anti-logical by virtue of being \textit{feelings.} Yet
strangely enough, I can't find any theorem of
probability theory which proves that I should appear ice-cold and
expressionless.
So is rationality orthogonal to feeling? No; our emotions arise
from our models of reality. If I believe that my dead brother has been
discovered alive, I will be happy; if I wake up and realize it was a
dream, I will be sad. P.~C.~Hodgell said: ``That which
can be destroyed by the truth should be.'' My
dreaming self's happiness was opposed by truth. My
sadness on waking is rational; there is no truth which destroys it.
Rationality begins by asking how-the-world-is, but spreads virally
to any other thought which depends on how we think the world is. Your
beliefs about ``how-the-world-is''
can concern anything you think is out there in reality, anything that
either does or does not exist, any member of the class
``things that can make other things
happen.'' If you believe that there is a goblin in
your closet that ties your shoes' laces together, then
this is a belief about how-the-world-is. Your shoes are real---you can
pick them up. If there's something out there that can
reach out and tie your shoelaces together, it must be real too, part of
the vast web of causes and effects we call the
``universe.''
\textit{Feeling angry at} the goblin who tied your shoelaces
involves a state of mind that is not \textit{just} about
how-the-world-is. Suppose that, as a Buddhist or a lobotomy patient or
just a very phlegmatic person, finding your shoelaces tied together
didn't make you angry. This wouldn't
affect what you expected to see in the world---you'd
still expect to open up your closet and find your shoelaces tied
together. Your anger or calm shouldn't affect your best
guess here, because what happens in your closet does not depend on your
emotional state of mind; though it may take some effort to think that
clearly.
But the angry feeling is tangled up with a state of mind that
\textit{is} about how-the-world-is; you become angry \textit{because}
you think the goblin tied your shoelaces. The criterion of rationality
spreads virally, from the initial question of whether or not a goblin
tied your shoelaces, to the resulting anger.
Becoming more rational---arriving at better estimates of
how-the-world-is---can diminish feelings \textit{or intensify them}.
Sometimes we run away from strong feelings by denying the facts, by
flinching away from the view of the world that gave rise to the
powerful emotion. If so, then as you study the skills of rationality
and train yourself not to deny facts, your feelings will become
stronger.
In my early days I was never quite certain whether it was
\textit{all right} to feel things strongly---whether it was allowed,
whether it was proper. I do not think this confusion arose only from my
youthful misunderstanding of rationality. I have observed similar
troubles in people who do not even aspire to be rationalists; when they
are happy, they wonder if they are really allowed to be happy, and when
they are sad, they are never quite sure whether to run away from the
emotion or not. Since the days of Socrates at least, and probably long
before, the way to appear cultured and sophisticated has been to never
let anyone see you care strongly about anything. It's
\textit{embarrassing} to feel---it's just not done in
polite society. You should see the strange looks I get when people
realize how much I care about rationality. It's not the
unusual subject, I think, but that they're not used to
seeing sane adults who visibly care about \textit{anything.}
But I know, now, that there's nothing wrong with
feeling strongly. Ever since I adopted the rule of
``That which can be destroyed by the truth should
be,'' I've also come to realize
``That which the truth nourishes should
thrive.'' When something good happens, I am happy,
and there is no confusion in my mind about whether it is rational for
me to be happy. When something terrible happens, I do not flee my
sadness by searching for fake consolations and false silver linings. I
visualize the past and future of humankind, the tens of billions of
deaths over our history, the misery and fear, the search for answers,
the trembling hands reaching upward out of so much blood, what we could
become someday when we make the stars our cities, all that darkness and
all that light---I know that I can never truly understand it, and I
haven't the words to say. Despite all my philosophy I
am still embarrassed to confess strong emotions, and
you're probably uncomfortable hearing them. But I know,
now, that it is rational to feel.
\myendsectiontext
\mysection{Why Truth? And \ldots}
\label{why_truth}
Some of the comments on \textit{Overcoming Bias} have touched on
the question of why we ought to seek truth. (Thankfully not many have
questioned what truth is.) Our shaping motivation for configuring our
thoughts to rationality, which determines whether a given configuration
is ``good'' or
``bad,'' comes from whyever we
wanted to find truth in the first place.
It is written:\footnote{Eliezer Yudkowsky, {\em Rationality From AI to Zombies}, pg \pageref{twelve_virtues}} ``The first virtue is
curiosity.'' Curiosity is one reason to seek truth,
and it may not be the only one, but it has a special and admirable
purity. If your motive is curiosity, you will assign priority to
questions according to how the questions, themselves, tickle your
personal aesthetic sense. A trickier challenge, with a greater
probability of failure, may be worth more effort than a simpler one,
just because it is more fun.
As I noted, people often think of rationality and emotion as
adversaries. Since curiosity is an emotion, I suspect that some people
will object to treating curiosity as a part of rationality. For my
part, I label an emotion as ``not
rational'' if it rests on mistaken beliefs, or
rather, on mistake-producing epistemic conduct: ``If
the iron approaches your face, and you believe it is hot, and it is
cool, the Way opposes your fear. If the iron approaches your face, and
you believe it is cool, and it is hot, the Way opposes your
calm.'' Conversely, then, an emotion that is evoked
by correct beliefs or epistemically rational thinking is a
``rational emotion''; and this has
the advantage of letting us regard calm as an emotional state, rather
than a privileged default.
When people think of
``emotion'' and
``rationality'' as opposed, I
suspect that they are really thinking of System 1 and System 2---fast
perceptual judgments versus slow deliberative judgments. Deliberative
judgments aren't always true, and perceptual judgments
aren't always false; so it is very important to
distinguish that dichotomy from
``rationality.'' Both systems can
serve the goal of truth, or defeat it, depending on how they are used.
Besides sheer emotional curiosity, what other motives are there
for desiring truth? Well, you might want to accomplish some specific
real-world goal, like building an airplane, and therefore you need to
know some specific truth about aerodynamics. Or more mundanely, you
want chocolate milk, and therefore you want to know whether the local
grocery has chocolate milk, so you can choose whether to walk there or
somewhere else. If this is the reason you want truth, then the priority
you assign to your questions will reflect the expected utility of their
information---how much the possible answers influence your choices, how
much your choices matter, and how much you expect to find an answer
that changes your choice from its default.
To seek truth merely for its instrumental value may seem
impure---should we not desire the truth for its own sake?---but such
investigations are extremely important because they create an outside
criterion of verification: if your airplane drops out of the sky, or if
you get to the store and find no chocolate milk, it's a
hint that you did something wrong. You get back feedback on which modes
of thinking work, and which don't. Pure curiosity is a
wonderful thing, but it may not linger too long on verifying its
answers, once the attractive mystery is gone. Curiosity, as a human
emotion, has been around since long before the ancient Greeks. But what
set humanity firmly on the path of Science was noticing that certain
modes of thinking uncovered beliefs that let us \textit{manipulate the
world.} As far as sheer curiosity goes, spinning campfire tales of gods
and heroes satisfied that desire just as well, and no one realized that
anything was wrong with that.
Are there motives for seeking truth besides curiosity and
pragmatism? The third reason that I can think of is morality: You
believe that to seek the truth is noble and important and worthwhile.
Though such an ideal also attaches an intrinsic value to truth,
it's a very different state of mind from curiosity.
Being curious about what's behind the curtain
doesn't feel the same as believing that you have a
moral duty to look there. In the latter state of mind, you are a lot
more likely to believe that someone \textit{else} should look behind
the curtain, too, or castigate them if they deliberately close their
eyes. For this reason, I would also label as
``morality'' the belief that
truthseeking is pragmatically important \textit{to society}, and
therefore is incumbent as a duty upon all. Your priorities, under this
motivation, will be determined by your ideals about which truths are
most important (not most useful or most intriguing), or about when,
under what circumstances, the duty to seek truth is at its strongest.
I tend to be suspicious of morality as a motivation for
rationality, \textit{not} because I reject the moral ideal, but because
it invites certain kinds of trouble. It is too easy to acquire, as
learned moral duties, modes of thinking that are dreadful missteps in
the dance. Consider Mr. Spock of \textit{Star Trek}, a naive archetype
of rationality. Spock's emotional state is always set
to ``calm,'' even when wildly
inappropriate. He often gives many significant digits for probabilities
that are grossly uncalibrated. (E.g., ``Captain, if
you steer the Enterprise directly into that black hole, our probability
of surviving is only 2.234\%.'' Yet nine times out of
ten the Enterprise is not destroyed. What kind of tragic fool gives
four significant digits for a figure that is off by two orders of
magnitude?) Yet this popular image is how many people conceive of the
duty to be ``rational''---small
wonder that they do not embrace it wholeheartedly. To make rationality
into a moral duty is to give it all the dreadful degrees of freedom of
an arbitrary tribal custom. People arrive at the wrong answer, and then
indignantly protest that they acted with propriety, rather than
learning from their mistake.
And yet if we're going to \textit{improve} our
skills of rationality, go beyond the standards of performance set by
hunter-gatherers, we'll need deliberate beliefs about
how to think with propriety. When we write new mental programs for
ourselves, they start out in System 2, the deliberate system, and are
only slowly---if ever---trained into the neural circuitry that
underlies System 1. So if there are certain kinds of thinking that we
find we want to \textit{avoid}{}---like, say, biases---it will end up
represented, within System 2, as an injunction not to think that way; a
professed duty of avoidance.
If we want the truth, we can most effectively obtain it by
thinking in certain ways, rather than others; these are the techniques
of rationality. And some of the techniques of rationality involve
overcoming a certain class of obstacles, the biases \ldots
\myendsectiontext
\mysection{\ldots What's a Bias, Again?}
A \textit{bias} is a certain kind of obstacle to our goal of
obtaining truth. (Its character as an
``obstacle'' stems from this goal of
truth.) However, there are many obstacles that are not
``biases.''
If we start right out by asking ``What is
bias?,'' it comes at the question in the wrong order.
As the proverb goes, ``There are forty kinds of lunacy
but only one kind of common sense.'' The truth is a
narrow target, a small region of configuration space to hit.
``She loves me, she loves me not''
may be a binary question, but $E = mc^2$ is a tiny dot
in the space of all equations, like a winning lottery ticket in the
space of all lottery tickets. Error is not an exceptional condition; it
is success that is a priori so improbable that it requires an
explanation.
{
We don't start out with a moral duty to
``reduce bias,'' because biases are
bad and evil and Just Not Done. This is the sort of thinking someone
might end up with if they acquired a deontological duty of
``rationality'' by social osmosis,
which leads to people trying to execute techniques without appreciating
the reason for them. (Which is bad and evil and Just Not Done,
according to \textit{Surely You're Joking, Mr.
Feynman}, which I read as a kid.)}
Rather, we want to get to the truth, for whatever reason, and we
find various obstacles getting in the way of our goal. These obstacles
are not wholly dissimilar to each other---for example, there are
obstacles that have to do with not having enough computing power
available, or information being expensive. It so happens that a large
group of obstacles seem to have a certain character in common---to
cluster in a region of obstacle-to-truth space---and this cluster has
been labeled ``biases.''
What is a bias? Can we look at the empirical cluster and find a
compact test for membership? Perhaps we will find that we
can't really give any explanation better than pointing
to a few extensional examples, and hoping the listener understands. If
you are a scientist just beginning to investigate fire, it might be a
lot wiser to point to a campfire and say ``Fire is
that orangey-bright hot stuff over there,'' rather
than saying ``I define fire as an alchemical
transmutation of substances which releases
phlogiston.'' You should not ignore something just
because you can't define it. I can't
quote the equations of General Relativity from memory, but nonetheless
if I walk off a cliff, I'll fall. And we can say the
same of biases---they won't hit any less hard if it
turns out we can't define compactly what a
``bias'' is. So we might point to
conjunction fallacies, to overconfidence, to the availability and
representativeness heuristics, to base rate neglect, and say:
``Stuff like that.''
With all that said, we seem to label as
``biases'' those obstacles to truth
which are produced, not by the cost of information, nor by limited
computing power, but by the shape of our own mental machinery. Perhaps
the machinery is evolutionarily optimized to purposes that actively
oppose epistemic accuracy; for example, the machinery to win arguments
in adaptive political contexts. Or the selection pressure ran skew to
epistemic accuracy; for example, believing what others believe, to get
along socially. Or, in the classic heuristic-and-bias, the machinery
operates by an identifiable algorithm that does some useful work but
also produces systematic errors: the availability heuristic is not
itself a bias, but it gives rise to identifiable, compactly describable
biases. Our brains are doing something wrong, and after a lot of
experimentation and/or heavy thinking, someone identifies the problem
in a fashion that System 2 can comprehend; then we call it a
``bias.'' Even if we can do no
better for knowing, it is still a failure that arises, in an
identifiable fashion, from a particular kind of cognitive
machinery---not from having too little machinery, but from the
machinery's shape.
``Biases'' are distinguished
from errors that arise from cognitive content, such as adopted beliefs,
or adopted moral duties. These we call
``mistakes,'' rather than
``biases,'' and they are much easier
to correct, once we've noticed them for ourselves.
(Though the source of the mistake, or the source of the source of the
mistake, may ultimately be some bias.)
``Biases'' are distinguished
from errors that arise from damage to an individual human brain, or
from absorbed cultural mores; biases arise from machinery that is
humanly universal.
Plato wasn't
``biased'' because he was ignorant
of General Relativity---he had no way to gather that information, his
ignorance did not arise from the shape of his mental machinery. But if
Plato believed that philosophers would make better kings because he
himself was a philosopher---and this belief, in turn, arose because of
a universal adaptive political instinct for self-promotion, and not
because Plato's daddy told him that everyone has a
moral duty to promote their own profession to governorship, or because
Plato sniffed too much glue as a kid---then that was a bias, whether
Plato was ever warned of it or not.
Biases may not be cheap to correct. They may not even be
correctable. But where we look upon our own mental machinery and see a
causal account of an identifiable class of errors; and when the problem
seems to come from the evolved shape of the machinery, rather from
there being too little machinery, or bad specific content; then we call
that a bias.
Personally, I see our quest in terms of acquiring personal skills
of rationality, in improving truthfinding technique. The challenge is
to attain the positive goal of truth, not to avoid the negative goal of
failure. Failurespace is wide, infinite errors in infinite variety. It
is difficult to describe so huge a space: ``What is
true of one apple may not be true of another apple; thus more can be
said about a single apple than about all the apples in the
world.'' Success-space is narrower, and therefore
more can be said about it.
While I am not averse (as you can see) to discussing definitions,
we should remember that is not our primary goal. We are here to pursue
the great human quest for truth: for we have desperate need of the
knowledge, and besides, we're curious. To this end let
us strive to overcome whatever obstacles lie in our way, whether we
call them ``biases'' or not.
\myendsectiontext
\mysection{Availability}
The \textit{availability heuristic} is judging the frequency or
probability of an event by the ease with which examples of the event
come to mind.
A famous 1978 study by Lichtenstein, Slovic, Fischhoff, Layman,
and Combs, ``Judged Frequency of Lethal
Events,'' studied errors in quantifying the severity
of risks, or judging which of two dangers occurred more
frequently.\footnote{Sarah Lichtenstein et al., ``Judged Frequency
of Lethal Events,'' \textit{Journal of Experimental
Psychology: Human Learning and Memory} 4, no. 6 (1978): 551--578,
doi:10.1037/0278-7393.4.6.551.\comment{1}} Subjects thought that accidents caused
about as many deaths as disease; thought that homicide was a more
frequent cause of death than suicide. Actually, diseases cause about
sixteen times as many deaths as accidents, and suicide is twice as
frequent as homicide.
An obvious hypothesis to account for these skewed beliefs is that
murders are more likely to be talked about than suicides---thus,
someone is more likely to recall hearing about a murder than hearing
about a suicide. Accidents are more dramatic than diseases---perhaps
this makes people more likely to remember, or more likely to recall, an
accident. In 1979, a followup study by Combs and Slovic showed that the
skewed probability judgments correlated strongly (0.85 and 0.89) with
skewed reporting frequencies in two newspapers.\footnote{Barbara Combs and Paul Slovic, ``Newspaper
Coverage of Causes of Death,'' \textit{Journalism \&
Mass Communication Quarterly} 56, no. 4 (1979): 837--849,
doi:10.1177/107769907905600420.\comment{2}} This
doesn't disentangle whether murders are more available
to memory because they are more reported-on, or whether newspapers
report more on murders because murders are more vivid (hence also more
remembered). But either way, an availability bias is at work. Selective
reporting is one major source of availability biases. In the ancestral
environment, much of what you knew, you experienced yourself; or you
heard it directly from a fellow tribe-member who had seen it. There was
usually at most one layer of selective reporting between you, and the
event itself. With today's Internet, you may see
reports that have passed through the hands of six bloggers on the way
to you---six successive filters. Compared to our ancestors, we live in
a larger world, in which far more happens, and far less of it reaches
us---a much stronger selection effect, which can create much larger
availability biases.
In real life, you're unlikely to ever meet Bill
Gates. But thanks to selective reporting by the media, you may be
tempted to compare your life success to his---and suffer hedonic
penalties accordingly. The objective frequency of Bill Gates is
0.00000000015, but you hear about him much more often. Conversely, 19\%
of the planet lives on less than \$1/day, and I doubt that one fifth of
the blog posts you read are written by them.
{
Using availability seems to give rise to an absurdity bias; events
that have never happened are not recalled, and hence deemed to have
probability zero. When no flooding has recently occurred (and yet the
probabilities are still fairly calculable), people refuse to buy flood
insurance even when it is heavily subsidized and priced far below an
actuarially fair value. Kunreuther et al.~suggest underreaction to
threats of flooding may arise from ``the inability of
individuals to conceptualize floods that have never occurred\,\ldots Men
on flood plains appear to be very much prisoners of their experience\,\ldots
Recently experienced floods appear to set an upward bound to the
size of loss with which managers believe they ought to be
concerned.''\footnote{Howard Kunreuther, Robin Hogarth, and Jacqueline Meszaros,
``Insurer Ambiguity and Market
Failure,'' \textit{Journal of Risk and Uncertainty} 7
(1 1993): 71--87, doi:10.1007/BF01065315.\comment{3}}}
Burton et al.~report that when dams and levees are built, they
reduce the frequency of floods, and thus apparently create a false
sense of security, leading to reduced precautions.\footnote{Ian Burton, Robert W. Kates, and Gilbert F. White, \textit{The
Environment as Hazard}, 1st ed. (New York: Oxford University Press,
1978).\comment{4}}
While building dams decreases the \textit{frequency} of floods, damage
\textit{per flood} is afterward so much greater that average yearly
damage \textit{increases.} The wise would extrapolate from a memory of
small hazards to the possibility of large hazards. Instead, past
experience of small hazards seems to set a perceived upper bound on
risk. A society well-protected against minor hazards takes no action
against major risks, building on flood plains once the regular minor
floods are eliminated. A society subject to regular minor hazards
treats those minor hazards as an upper bound on the size of the risks,
guarding against regular minor floods but not occasional major floods.
Memory is not always a good guide to probabilities in the past,
let alone in the future.
\myendsectiontext
\bigskip
\mysection{Burdensome Details}
\label{burdensome_details}
\begin{quote}
Merely corroborative detail, intended to give artistic
verisimilitude to an otherwise bald and unconvincing narrative\,\ldots
{\raggedleft
{}---Pooh-Bah, in Gilbert and Sullivan's
\textit{The Mikado}\footnote{William S. Gilbert and Arthur Sullivan, \textit{The Mikado},
Opera, 1885.\comment{1}}
\par}
\end{quote}
The conjunction fallacy is when humans rate the probability $P(A,B)$
higher than the probability $P(B)$, even though it is a theorem that
$P(A,B) \leq P(B)$. For example, in one experiment in 1981, 68\% of
the subjects ranked it more likely that ``Reagan will
provide federal support for unwed mothers and cut federal support to
local governments'' than that
``Reagan will provide federal support for unwed
mothers.''
A long series of cleverly designed experiments, which weeded out
alternative hypotheses and nailed down the standard interpretation,
confirmed that conjunction fallacy occurs because we
``substitute judgment of representativeness for
judgment of probability.'' By adding extra details,
you can make an outcome seem \textit{more} characteristic of the
process that generates it. You can make it sound more plausible that
Reagan will support unwed mothers, by \textit{adding} the claim that
Reagan will \textit{also} cut support to local governments. The
implausibility of one claim is compensated by the plausibility of the
other; they ``average out.''
{
Which is to say: Adding detail can make a scenario \textsc{sound more
plausible}, even though the event necessarily \textsc{becomes less probable}.}
If so, then, \textit{hypothetically speaking,} we might find
futurists spinning unconscionably plausible and detailed future
histories, or find people swallowing huge packages of unsupported
claims bundled with a few strong-sounding assertions at the center. If
you are presented with the conjunction fallacy in a naked, direct
comparison, then you may succeed on that particular problem by
consciously correcting yourself. But this is only slapping a band-aid
on the problem, not fixing it in general.
In the 1982 experiment where professional forecasters assigned
systematically higher probabilities to ``Russia
invades Poland, followed by suspension of diplomatic relations between
the USA and the USSR'' than to
``Suspension of diplomatic relations between the USA
and the USSR,'' each experimental group was only
presented with one proposition.\footnote{Tversky and Kahneman, ``Extensional Versus
Intuitive Reasoning.''\comment{2}} What strategy could
these forecasters have followed, as a group, that would have eliminated
the conjunction fallacy, when no individual knew directly about the
comparison? When no individual even knew that the experiment was
\textit{about} the conjunction fallacy? How could they have done better
on their probability judgments?
Patching one gotcha as a special case doesn't fix
the general problem. The gotcha is the symptom, not the disease.
What could the forecasters have done to avoid the conjunction
fallacy, without seeing the direct comparison, or even knowing that
anyone was going to test them on the conjunction fallacy? It seems to
me, that they would need to notice the word
``and.'' They would need to be wary
of it---not just wary, but leap back from it. Even without knowing that
researchers were afterward going to test them on the conjunction
fallacy particularly. They would need to notice the conjunction of
\textit{two entire details}, and be \textit{shocked} by the audacity of
anyone asking them to endorse such an insanely complicated prediction.
And they would need to penalize the probability
\textit{substantially}{}---a factor of four, at least, according to the
experimental details.
It might also have helped the forecasters to think about possible
reasons why the US and Soviet Union would suspend diplomatic relations.
The scenario is not ``The US and Soviet Union suddenly
suspend diplomatic relations for no reason,'' but
``The US and Soviet Union suspend diplomatic relations
for any reason.''
And the subjects who rated ``Reagan will provide
federal support for unwed mothers and cut federal support to local
governments''? Again, they would need to be shocked
by the word ``and.'' Moreover, they
would need to \textit{add} absurdities---where the absurdity is the log
probability, so you can add it---rather than averaging them. They would
need to think, ``Reagan might or might not cut support
to local governments (1 bit), but it seems very unlikely that he will
support unwed mothers (4 bits). \textit{Total} absurdity: 5
bits.'' Or maybe, ``Reagan
won't support unwed mothers. One strike and
it's out. The other proposition just makes it even
worse.''
Similarly, consider the six-sided die with four green faces and
two red faces. The subjects had to bet on the sequence (1) \textsc{rgrrr}, (2)
\textsc{grgrrr}, or (3) \textsc{grrrrr} appearing anywhere in twenty rolls of the
dice.\footnote{Amos Tversky and Daniel Kahneman, ``Judgments
of and by Representativeness,'' in \textit{Judgment
Under Uncertainty: Heuristics and Biases}, ed. Daniel Kahneman, Paul
Slovic, and Amos Tversky (New York: Cambridge University Press, 1982),
84--98.\comment{3}} Sixty-five percent of the subjects chose
\textsc{grgrrr}, which is strictly dominated by \textsc{rgrrr}, since any sequence
containing \textsc{grgrrr} also pays off for \textsc{rgrrr}. How could the subjects have
done better? By noticing the inclusion? Perhaps; but that is only a
band-aid, it does not fix the fundamental problem. By explicitly
calculating the probabilities? That would certainly fix the fundamental
problem, but you can't always calculate an exact
probability.
The subjects lost heuristically by thinking:
``Aha! Sequence 2 has the highest proportion of green
to red! I should bet on Sequence 2!'' To win
heuristically, the subjects would need to think:
``Aha! Sequence 1 is \textit{short}! I should go with
Sequence 1!''
They would need to feel a stronger \textit{emotional impact} from
Occam's Razor---feel \textit{every} added detail as a
burden, even a single extra roll of the dice.
Once upon a time, I was speaking to someone who had been
mesmerized by an incautious futurist (one who adds on lots of details
that sound neat). I was trying to explain why I was not likewise
mesmerized by these amazing, incredible theories. So I explained about
the conjunction fallacy, specifically the ``suspending
relations {\textpm} invading Poland'' experiment. And
he said, ``Okay, but what does this have to do
with---'' And I said, ``It is more
probable that universes replicate \textit{for any reason}, than that
they replicate \textit{via black holes because advanced civilizations
manufacture black holes because universes evolve to make them do
it}.'' And he said,
``Oh.''
Until then, he had not felt these extra details as extra burdens.
Instead they were corroborative detail, lending verisimilitude to the
narrative. Someone presents you with a package of strange ideas,
\textit{one} of which is that universes replicate. Then they present
support \textit{for the assertion that universes replicate.} But this
is not support for the package, though it is all told as one story.
You have to disentangle the details. You have to hold up every one
independently, and ask, ``How do we know \textit{this}
detail?'' Someone sketches out a picture of
humanity's descent into nanotechnological warfare,
where China refuses to abide by an international control agreement,
followed by an arms race\,\ldots Wait a minute---how do you know it will
be China? Is that a crystal ball in your pocket or are you just happy
to be a futurist? Where are all these details coming from? Where did
\textit{that specific} detail come from?
For it is written:\footnote{Eliezer Yudkowsky, {\em Rationality From AI to Zombies}, pg \pageref{seventh_virtue}}
\begin{quote}
\textit{If you can lighten your burden you must do so.}
{
\textit{There is no straw that lacks the power to break your
back.}}
\end{quote}
\myendsectiontext
\bigskip
\mysection{Planning Fallacy}
{
The Denver International Airport opened 16 months late, at a cost
overrun of \$2 billion. (I've also seen \$3.1 billion
asserted.) The Eurofighter Typhoon, a joint defense project of several
European countries, was delivered 54 months late at a cost of \$19
billion instead of \$7 billion. The Sydney Opera House may be the most
legendary construction overrun of all time, originally estimated to be
completed in 1963 for \$7 million, and finally completed in 1973 for
\$102 million.\footnote{Roger Buehler, Dale Griffin, and Michael Ross,
``Inside the Planning Fallacy: The Causes and
Consequences of Optimistic Time Predictions,'' in
Gilovich, Griffin, and Kahneman, \textit{Heuristics and Biases},
250--270.\comment{1}} }