Epistemology and Congnition - Alvin I. Goldman
Epistemology and Congnition - Alvin I. Goldman
Epistemology and Congnition - Alvin I. Goldman
EPISTEMOLOGY
and
COGNITION
Alvin I. Goldman
THEIDEA of this book germinated while I was a fellow at the Center for
Advanced Study in the Behavioral Sciences during 1975-76, with support
from the Guggenheim Foundation. Conceived there was the idea of epis-
temics: an enterprise linking traditional epistemology, first, with cognitive
science and, second, with social scientific and humanistic disciplines that
explore the interpersonal and cultural processes impinging on knowledge
and belief. The intention was to enrich epistemology while preserving its
own identity. This book articulates the first part of epistemics: the relation
between epistemology and cognitive science. A sequel is planned to delin-
eate the second part: social epistemics.
In trying to lay a satisfactory conceptual foundation for epistemics, I was
aided at the Behavioral Science Center by discussions with John Perry,
Keith Donnellan, Michael Bratman, and Holly Smith, and by later debates
with Richard Nisbett during a seminar we taught jointly. Initial program-
matic statements appeared in "Epistemics: The Regulative Theory of Cog-
nition," The Journal of Philosophy, 75 (1978):509-523, and "Varieties of
Cognitive Appraisal," Now, 13 (1979):23-38. Other precursors of material
.
in this book include "What Is Justified Belief?" in G. Pappas, ed., J w t i . a -
tion and Knowledge (Dordrecht: D. Reidel, 1979), and "The Internalist
Conception of Justification," in P. French, T. Uehling, Jr., and H. Wett-
stein, ed~.,Midwest Studies in Philosophy, vol. 5, Studies in Epistemology
(Minneapolis: University of Minnesota Press, 1980).Themes from "The Re-
lation between Epistemology and Psychology," Synthese, 64 (1985):29-68
(written in 1981), form the basis for parts of Chapters 4 and 5. "Epistemol-
ogy and the Theory of Problem Solving," Synthese, 55 (1983):21-48, is the
basis for much of Chapter 6. A version of Chapter 11, "Constraints on Rep-
resentation," appears in Myles Brand and Robert Harnish, eds., The Repre-
sentation of Knowledge and Belief (Tucson: University of Arizona Press,
1986).
viii Preface
Introduction 1
THEORETICAL FOUNDATIONS
The Ekments of E p i s t e m o l o 13
Skepticism 28
Knowledge 42
Justification: A Rule Framework 58
Justification and Reliability 81
Problem Solving, Power, and Speed 122
Truth and Realism 142
The Problem of Content 162
part I1 ASSESSING OUR COGNITIVE RESOURCES
9 Perception 181
10 Memory 199
11 Constraints on Representation 227
12 Internal Codes 252
13 Deductive Reasoning 278
14 Probability Judgments 305
15 Acceptance and Uncertainty 324
16 Belief Updating 344
17 Production Systems and Second-Order Processes 359
Conclusion: Primary Epistemics and Cognitive Science 378
Notes 383
Illustration Credits 421
Author Index 423
Subject Index 429
Introduction
epistemology and the sciences of cognition is treated in this book. The na-
ture and structure of social epistemology will be examined in a future study.
To some readers the interdisciplinary theme may sound banal. Episte-
mology deals with knowledge, which is the property of individual minds. So
of course epistemology must be interested in the knowing mind. Similarly,
most knowledge is a cultural product, channeled through language and so-
cial communication. So how could epistemology fail to be intertwined with
studies of culture and social systems?
Despite these truisms, strong countercurrents in the history of epistemol-
ogy run against the interdisciplinary theme. Here is a sampling of such
countercurrents (not all mutually compatible, being drawn from different
traditions).
(1)As the study of method, epistemology should be autonomous. It should
2 Introduction
be prior to the sciences; so it must not seek help from them. (2) Epistemol-
ogy should only be concerned with the analysis of concepts, specifically
epistemic concepts such as 'knowledge', 'warrant', 'rationality', and the
like. But conceptual or linguistic analysis is the province of philosophy; so
epistemology needs no help from behavioral or social sciences. (3)The true
aim of philosophy is to "show the fly the way out of the fly-bottle,"' that is,
to dissolve puzzles and paradoxes that lead to skepticism. Such dissolution
requires only linguistic analysis, not a model of the mind-brain or empirical
models of intellectual influence. (4) Epistemology is the study of methodol-
ogy, and proper methodology is the province of deductive logic, inductive
logic, probability theory, and statistics. Epistemology reduces to these sub-
jects, all of which are f o m l disciplines. Empirical sciences are not needed.
(5) Epistemology is normative, evaluative, or critical, not descriptive. So
empirical sciences, which are purely descriptive, cannot help epistemology.
The psychology of reasoning, for example, cannot shed light on proper rea-
soning, on logically or scientifically sound reasoning. In view of these me-
taepistemological currents, the proposed interdisciplinary theme is
controversial, not trivial. It needs sustained clarification and defense.
Furthermore, the interdisciplinary theme is only one of my themes. Of
equal or greater importance is the specific epistemological framework to be
proposed, which specifies the particular ways in which disciplinary collabo-
ration should proceed. This framework contrasts with other systems of
epistemology that would equally favor an interdisciplinary orientation. Let
me briefly mention a few components of this framework.
One crucial component is the evaluative mission of epistemology. Along
with the dominant tradition, I regard epistemology as an evaluative, or
normative, field, not a purely descriptive one. This makes it far from obvi-
ous how positive science can have inputs to epistemology. How, exactly, do
facts of cognition or social intercourse bear on epistemic evaluations or
norms?
A few other recent characterizations of epistemology also link it with
psychology. But these characterizations depict the field as a descriptive one.
On a purely descriptive conception it is not surprising that epistemology
should be indebted to psychology-should even reduce to it. Thus, on W: V.
Quine's naturalistic conception, the epistemologist would study how the
human subject responds to certain input; how, in response to various stimu-
lus patterns, the subject delivers a description of the external world and its
history. In studying the relation between this "meager input" and "torren-
tial output," epistemology "simply falls into place as a chapter of psychol-
ogy and hence of natural s ~ i e n c e . Similarly,
"~ Donald Campbell advances a
conception of the field which he calls "evolutionary epistemology." On this
conception epistemology takes cognizance of "man's status as a product of
biological and social ev~lution."~ Campbell explicitly characterizes his
conception as descriptive: descriptive of man as k n ~ w e r . ~
Introduction 3
of interest in its social dimensions. This interest goes back at least to Plato
and the Socratic elenchus. Probably influenced by the sophists of his day,
Plato treated inquiry as an interpersonal, dialectical process, in which two
inquirers try to sustain their opinions against the disputations of their oppo-
nent. This dialectic was regarded by Plato as an essential epistemic device.
In more recent times the epistemology of science has often emphasized
social elements. Pragmatists like C. S. Peirce and John Dewey stressed the
corporate aspect of scientific inquiry, and positivists also stressed the inter-
subjectivity, or publicity, of scientific knowledge. Thomas Kuhn is identi-
fied with a different social perspective on science, with his emphasis on the
paradigms that scientific communities share and transmit.12 ~istoriansand
sociologists of science are generally concerned with the communities, or
schools, that have marked scientific controversies.
A different social slant to epistemology was introduced by Wittgenstein,
who viewed epistemological criteria as portions of language games, under-
stood as species of social practices.'3 This Wittgensteinian slant underlies
Richard Rorty's recent formulation of a social conception of epistemology.
For Rorty, knowledge is "the social justification of belief," and rationality
and epistemic authority are supposed to be explained by reference to "what
society lets us say.""
Precedents for social epistemology are plainly legion, but most of the
current approaches strike me as unsatisfactory in various ways. Historians
and sociologists of science typically devote their energy to descriptions of
scientific practice and development. Little attention is given to the evalua-
tive issues that are characteristic of epistemology. When evaluative ques-
tions are broached, they are seldom tackled with a firm theoretical
foundation. I shall try to remedy some of these deficiencies in my subse-
quent work on social epistemology.
The territorial instinct may incline some philosophers to view the pro-
posed interdisciplinary conception of epistemology as a regrettable surren-
der of philosophical autonomy. But epistemology's lack of autonomy would
only parallel that of other branches of philosophy. Neither moral nor
social-political philosophy can proceed without recourse to other disci-
plines. How can moral philosophy abstract from contingent facts of the
human psyche: the sources of aggression and sympathy, the sense of fairness
and reciprocity? How can it abstract from our feelings about ourselves, our
offspring, and infrahuman creatures? How can social philosophy neglect the
feasibilities and infeasibilities imposed by economic, historical, and political
contingencies? Admittedly, foundational questions of ethical and social the-
ory may be addressable without appeal to particular social facts. But sub-
stantive conclusions of social theory must rest in part on human nature and
economic and political feasibilities. Therefore, substantive ethics and social
theory must be multidisciplinary affairs.
Introduction 9
acts, the ones of central concern are assertions. Most intellectual endeavors
try to arrive at some belief on a designated topic, or to formulate a state-
ment on the problem at hand. Accordingly, the 'product' of the scientist or
scholar is typically a body of assertions-presumably accompanied by a
body of beliefs. So epistemology naturally focuses on either beliefs or asser-
tive claims. In this book I will largely abstract from natural language and
public speech acts, since there will be enough complexities on our hands
without the complexities these topics introduce. I will concentrate on the
mental side of epistemology, without implying that this is its only side.
Belief is not the mentalistic concept of choice for all epistemologists. Be-
lief is normally a categorical, or binary, concept, an all-or-nothing affair.
You either believe something or you don't. But epistemologists commonly
point out that there are degrees of conviction or confidence in a statement.
14 The Elements of Epistemology
occur in consciousness to count as mental. The belief that you reside at 3748
Hillview Road may be held for many years, though you only 'think about'
your address intermittently during this period.
Many issues about beliefs and belief ascriptions are controversial. Most of
the controversy concerns belief contents. Does a mental state really have a
determinate content? If so, what is the source of this content? Is content
determined exclusively by what's 'in the head', or by external factors as
well? If external factors are relevant, can we really say that belief states (in-
cluding their contents) are purely mental, that is, inner states of the individ-
ual considered in abstraction from his or her environment or causal
ancestry?
Another problem concerns the objects (or relata) of beliefs. The term
'propositional attitude' naturally suggests that the objects in question are
propositions. And this is a common way of talking, both among philosophers
and psychologists. But propositions are problematic entities. As classically
interpreted, they are logical or abstract entities, somewhat akin to Platonic
forms. Philosophers widely regard this sort of ontological status with suspi-
cion. There are other theories of propositions, but none is free from criti-
cism. In place of propositions some philosophers posit sentences of an inner
language, a lingua mentis, as the relata of beliefs. But this sort of posit is also
controversial.
These sticky issues about beliefs are relevant to my enterprise. They are
relevant, first, because it is hard to do any mentalistic epistemology without
using something like the belief construct. Second, they are relevant because
of the intended bearing of psychology on mentalistic epistemology. Beliefs
may not be a kind of entity with which psychology is capable of dealing. Or,
being a construct of folk psychology, scientific psychology may choose to
replace the belief construct entirely and work with different theoretical
resources.
While all these issues are relevant to my undertaking, they cannot be
fully resolved here. That would require a full-scale treatment of difficult
topics in the philosophy of mind and the philosophy of language, which go
beyond the scope of this book. Several of these issues will be addressed
squarely in Chapter 8, but even there they will not receive a full-dress dis-
cussion. Nonetheless, some remarks are also advisable at this juncture, to
indicate roughly where I stand on these issues.
For convenience, I will proceed on the assumption that the objects, or
relata, of beliefs are propositions. The utility of proposition-talk is that it
nicely unifies the treatment of mental attitudes, overt speech, and truth-
value ascriptions. Suppose Alex says 'The moon is round', and Kurt says
'Der Mond is rund'. It is natural to say not merely that both asserted a truth,
but that they asserted the same truth, namely, that the moon is round. But
this single truth cannot be their two sentences, since these were different.
16 The Elements of Epistemology
The suggestion is that the two sentences expressed one proposition, which is
the truth in question. Suppose further that Oscar utters no such sentence,
but has a belief he could express as 'The moon is round'. Then it is tempting
to say that he believes the same truth that Alex and Kurt assert. How can
this be explained? Again, proposition-talk can do the job. If Oscar's belief
content is a proposition-the same one expressed by Alex's and Kurt's sen-
tences-then it is easy to identify the single truth with this very proposition.
Thus, by letting propositions be (1)the contents of verbal assertions, (2) the
contents of beliefs, and (3) the bearers of truth, we nicely unify a whole
range of discourse.
Given the utility of proposition-talk, I shall avail myself of this talk, but
only as a f q o n de parler. I regard propositions as a temporary theoretical
posit from which we should ultimately ascend to a better theory. Since such
a theory may be quite complex, and its details probably would not seriously
affect my project, it is an issue to which I will devote little attention.
The bearing of psychology on the belief construct, however, is more cen-
tral to my concerns. So let me briefly anticipate a fuller discussion of this
matter in later chapters. What stance should cognitive science adopt to-
ward beliefs? At least four approaches are possible: reduction, replacement,
neutrality, and refinement.
A reductionist says that beliefs are perfectly fine entities for a serious
cognitive science, only they need to be reduced to, or explained in terms of,
scientifically respectable elements. Since propositions, for example, are du-
biously respectable, beliefs need to be understood as relations to syntactic
entities, such as sentences in a language of thought.
An advocate of replacement holds that beliefs are not scientifically ten-
able posits. As part of a radically false theory, a degenerating research pro-
gram, they will ultimately be discredited and abandoned like phlogiston.
An advocate of neutrality holds that cognitive science can remain neutral
about beliefs. It can proceed perfectly well without them, but this does not
imply their overthrow or illegitimacy. Cognitive science may not be in a
position to ascribe (propositional) content, or to say how such content
should be ascribed. But it may allow that a suitable theory of interpreta-
tion-a good 'psychosemantics'-can ascribe content, or show how content
is standardly ascribed.
The refinement position I have in mind is not an alternative to the pre-
ceding approaches, mainly because it does not address the problem of con-
tent. Content is not the only dimension of interest in the belief construct.
Psychology should also be interested in the range of contentful states. Cog-
nitive science could endorse the notion of content but stress the need for
acknowledging a richer array of distinct content-bearing states. Many dif-
ferent states, it may hold, are all lumped under the folk-psychological label
of 'belief, and these need to be sorted out for the purposes of an adequate
science of the mind.
The Elements of Epistemology 17
must distinguish between a proposition's being true and people hauing eui-
dence in its favor. Surely there are plenty of cases of true propositions for
which nobody now (or perhaps ever) has good evidence. The example of the
1993 toothache sufferers is a case in point.
Since many people, including philosophers, insist on conflating truth
with other notions-mostly related to evidence, justification, or the like-it
is worth dilating on this point at this early juncture. To take a recent exam-
ple, consider Richard Rorty's claim that there is a sense of 'true' in which it
means (roughly) "what you can defend against all comers."' The idea of
"what can be defended against all comers" is some sort of social justification
notion. According to Rorty, there is a sense of 'true' in which it is necessary
and sufficient for a proposition's truth that somebody can defend it, pre-
sumably successfully, against all who argue against it.
There is no such sense of 'true'. To appreciate what's wrong with this def-
inition, suppose you are an unfortunate victim of circumstance and misi-
dentification. A horrible crime has been committed of which you stand
accused. You are totally innocent-such is the truth. But you are a look-
alike of the dastardly criminal, and numerous witnesses come forward to
identify you as the doer of the deed. Sadly, you have no alibis. You were out
for a walk at the time of the act, and nobody can vouch for your where-
abouts. Meanwhile, the real criminal has died in an accident. Given these
facts, the real truth cannot be successfully defended against all comers. You
cannot defend it successfully, for there are too many eyewitnesses to make
your case believable. (Furthermore, we may suppose, you actually did have
a motive for the act, though it did not motivate you enough to do it.) Nor is
there anyone else who could successfully defend your innocence against all
comers. Nonetheless, the truth is: you are innocent. The only correct sense
of 'true' makes truth independent of how well it can be defended. Its de-
fensibility is a separate matter, which may depend on a variety of extrane-
ous circumstances. Any innocent person accused of a crime surely wants the
real truth to emerge; and the real truth is all that is normally meant by
'true'.
The innocent defendant case shows that "what can be defended against
all comers" is not a necessary condition of truth. But neither is it sufficient.
A totalitarian regime may arrange a successful defense of certain false prop-
ositions against all comers (at least all comers in that society); but this can-
not make the false true.
I will examine more sophisticated attempts to explain truth in terms of
evidential, or justificatory, notions in Chapter 7. At this point it is important
to distinguish truth and justification, whether social justification or intra-
personal justification.
It is equally imperative to distinguish truth from knowledge. A proposi-
tion may be true without being known, just as it may be true without being
The Elements of Epistemology 19
believed. Let me pause here to make a few remarks about the concept of
knowledge. There is a loose and a strict sense of 'knowledge7. In a loose
sense a person knows something (a proposition) if he believes it and it is
true. If he believes the truth, he is not ignorant of it. He is cognizant of the
fact, and so, in a loose sense, knows it. Many writers, especially in the be-
havioral and social sciences, use 'knowledge' in an ultra-loose sense, to
mean simply belief, or representation of the world. This is probably a mis-
use, though one so common that perhaps it has achieved legitimacy. In
more proper usage, to which I shall cleave, no proposition can be known
unless it is true.
To repeat: in the loose sense of 'know' someone knows proposition p if
and only if he believes p, and p is true. This suffices to show that knowledge
is not equivalent to truth: truth does not require belief, but knowledge does.
There is also, however, a strict sense of 'know', which has much occupied
epistemologists. In this strict sense knowledge requires more than belief and
truth. It requires satisfaction of some third, and perhaps fourth, condition
beyond belief and truth. Since epistemology has often been regarded as 'the
theory of knowledge', many epistemologists have devoted great energy to
analyzing this strict sense of 'know'. I too shall pay it a fair bit of attention,
especially in Chapter 3. But the analysis of knowledge will not be one of my
sustained points of preoccupation. I will use it partly as a jumping-off point
for an examination of the notion of justification, which will be one central
focus of my discussion. For the present, it suffices to note the complexity of
knowledge and to distinguish it from truth.
A principal reason for resisting a realist construal of truth is an alleged
threat to knowledge. Many philosophers fear that if truth is definitionally
prized off from evidence, or justification, it will be impossible for anyone to
know the truth. Truth(s) will be epistemically inaccessible, or unknowable.
But this dire consequence is not at all indicated, at least not without lengthy
argumentation. The mere fact that extramental reality is what makes a
proposition true (or false), as realists maintain, does not imply that no truths
can be known. The mere fact that truth does not (definitionally) require
knowledge, or justification, does not mean that it precludes knowledge, or
justification. (These comments do not convey the most serious historical or
contemporary arguments against realism. I mention them only in an intro-
ductory fashion, to forestall very simple confusions that may befall those
unfamiliar with epistemology.)
I have already had occasion to introduce a central epistemic concept, the
concept of justification. This has been a doubly important term in the field.
First, it is important in its own right, as a term of epistemic evaluation, or
appraisal, which is widely applied. Second, it is often linked to knowledge,
which, as I have noted, is traditionally central to epistemology. Its connec-
tion to knowledge will be discussed later on, in Chapter 3. But I want to
20 The Elements of Epistemology
turn now to justification in its own right; more precisely, to the larger fam-
ily of terms of epistemic evaluation of which 'justification' is one member.
controversial just which factual states are, or should be, linked with specific
evaluations. This is notoriously problematic in the ethical domain. But few
writers, except perhaps emotivists, doubt that some factual basis (whether
objective or subjective) underlies evaluative judgments, or should underlie
them. A central problem for epistemology, equally, is the factual basis on
which epistemic evaluations are made. Since there are numerous different
terms of epistemic evaluation, it must be expected that different bases will
be paired with different terms. Some terms may be vague, or ambiguous, in
which case a family of different factual bases may be associated with a sin-
gle term.
An illustration from ethical theory should help concretize the idea of fac-
tual grounding for evaluative judgments. 'Right' is a core term of moral ap-
praisal. But what factual circumstances make an action right? A sample
answer is provided by Act Utilitarianism. This theory counts an action A
right if and only if A causes, or would cause, as much net happiness as any
alternative action open to the agent. The rightness of an action is deter-
mined, on this theory, by the relative amounts of happiness that actions in
the agent's alternative set would produce. This is a purely factual matter
(using 'factual' in a broad sense, to include counterfactual states of affairs).
There is a fact of the matter as to what quantities of happiness and unhap-
piness would result from sundry actions. So there is a fact of the matter-
however hard it may be to know this fact-as to whether action A would
produce as much net happiness as other actions in the alternative set. Thus,
Act Utilitarianism provides a factual basis on which rightness (allegedly)
supervenes. It is doubtful, of course, that this theory is correct. But it pro-
vides a clear example of a factual basis.
In the case of epistemic status, we also want to know the factual, or sub-
stantive, conditions on which it supervenes. Consider justifiedness. What
has to be true of a belief for it to qualify as justified? What factual standard
determines justifiedness?
What kinds of answers to this question are legitimate?It is not admissible ,
This contrast can be illustrated with the Act Utilitarian theory of moral
rightness. Such a theory provides substantive conditions of rightness, but
not a verification procedure for determining rightness. For example, the
theory invokes the notion of happiness without offering any observational
reduction of happiness. It presupposes truth-values for various counterfac-
tual statements, without specifying verification procedures for such coun-
terfactuals. These omissions are not objectionable for the purposes in
question. No such verificational, or observational, reductions are being
sought. Similarly, in offering factual standards for epistemic terms, it is ad-
missible to include truth-linked standards, despite the fact that 'true' is not
an observational, or directly verifiable, term. I stress this point because my
own approach will feature truth-linked standards prominently. So it needs
emphasis at the outset that this is quite unobjectionable, given the goals in
question.
(4) Styles of evaluation. Evaluative, or normative, discourse appears in
different styles. Some specimens of such discourse include deontic terms,
like 'obligatory', 'permitted', and 'forbidden', or more familiarly, 'right' and
'wrong'. Often such terms have systems of rules associated with them. Let
us call this kind of normative discourse deontic. There are other forms of
evaluative discourse, however. Some evaluative terms merely appraise cer-
tain qualities as good-making or bad-making features, relative to some suit-
able dimensions. In other words, they pick out certain traits as virtues or
vices. This contrast is found among terms of epistemic appraisal. Epistemo-
logists often try to identify rules of rationality, and they sometimes talk
about what is rationally required. This is deontic language in the epistemic
domain. By contrast, terms like 'careful' and 'original' might be used to
evaluate intellectual performances or traits, but such terms apply no deon-
tic category. They merely identify certain good-making characteristics.
Let me mark the foregoing contrast by saying that evaluative terms can
register either deontic classifications, on the one hand, or virtues and vices,
on the other. Both styles of evaluation will be employed in this book. I will
treat justifiedness (Chapters 4 and 5) as a deontic notion and examine it in a
framework of rules. But other evaluations to be studied will be in the vir-
tue-vice style.
A different division of styles distinguishes regulative and nonregulative
normative schemes. A regulative system of norms formulates rules to be
consciously adopted and followed, for example, precepts or recipes by
which an agent should guide his conduct. A nonregulative system of evalua-
tion, by contrast, formulates principles for appraising a performance or
trait, or assigning a normative status, but without providing instructions for
the agent to follow, or apply. They are only principles for an appraiser to
utilize in judging.' Many epistemologists have sought regulative precepts:
principles to guide a cognizer's reasoning processes.5 But it is also possible
to do epistemic evaluation-even in a rule-based framework-without
26 The Elements of Epistemology
Skepticism
nents who portrayed the attitude of skeptics as a definite view, Sextus said
that it was like a purge that eliminates everything, including itself.
Pyrrhonian skepticism, then, might be characterized as the cultivation of
suspension of judgment, or doubt. This kind of skeptical theme does not
make pronouncements on the possibility of knowledge, certainty, or justifi-
cation. It only urges suspension of belief-though it is accompanied by ar-
guments to support this attitude. This form of skepticism contrasts with
other forms, which are constituted by theses about knowledge, certainty, or
other epistemic attainments. (In other historical periods, the term 'skeptic'
was often applied to anyone who rejected, or failed to endorse, widely ac-
cepted views.)
In contemporary epistemology, as well as in much of its history, skepti-
cism is primarily some sort of negative thesis about epistemic attainments. I
will focus on two of these attainments: knowledge and iustificatwn (or war-
rant). Skepticism has often been directed against certainty, but we can in-
terpret this formulation as a special case of either skepticism about
knowledge or skepticism about justification. Certainty can be viewed as
either the highest grade of knowledge or the highest grade of justification.
So skepticism directed against certainty falls either within the knowledge
theme or the justification theme.
Skepticisms differ not only in theme, but in scope. Some skepticisms are
universal, denying all knowledge or justification. This extreme sort of skep-
ticism is rather rare. Even the ancient skeptics tended to exempt proposi-
tions about appearances from the scope of their doubts. Most skepticisms,
then, are somewhat less comprehensive, restricting themselves to selected
domains such as ethics, mathematics, other minds, the past, the future, the
unobserved, and the external world. More global skepticisms tend to focus
on a selected theme, such as knowledge or justification, and try to show
what makes it impossible, or difficult, to achieve in any domain. More local
skepticisms tend to focus on the peculiarity of the selected subject mat-
ter-for example, the external world, or the future-and try to show what
makes it inaccessible to knowledge or justified belief.
Skepticisms vary not only in theme and scope, but in strength. Roughly, a
skepticism's strength is inversely related to the strength of the epistemic
achievement that it disputes. A common form of skepticism, we have seen,
is the denial of certainty. Certainty can be thought of as either the highest
grade of knowledge or the highest grade of justifiedness. Certainty-denying
skepticism disputes the existence of such high-grade epistemic attainments.
But this is a relatively weak sort of skepticism. Significant forms of knowl-
edge or justification may be weaker than certainty. We might be moder-
ately justified in believing something even if not completely justified. Such
lower grades of epistemic accomplishment are not disputed by this kind of
skepticism. A stronger form of skepticism, however, would deny even lower
30 Skepticism
out the threat of error. So the skeptic's denial of knowledge often takes the
form of insisting that the threat of error is serious.
Three kinds of error possibility are prominent in the skeptical literature:
(1) the fallibility of our cognitive faculties, (2) the relation between the
mind and the objects of cognition, and (3)the logical relation between hy-
pothesis and evidence.
Fallibility of the faculties is the earliest and perhaps most common
ground for skepticism. Foremost is the fallibility of the senses, but the falli-
bility of reuson runs a close second. Deception by the senses was intensively
discussed by Sextus Empiricus, and most later skeptics including Descartes
employed similar ideas. The deceptiveness of reason was commonly demon-
strated by showing how it yields incompatible conclusions. In the seven-
teenth century Pierre Bayle argued that most theories are "big with
contradiction and absurdity," that man's efforts to comprehend the world
rationally always end in perplexities and insoluble diffic~lties.~ Similarly,
Kant undermined the pretensions of Reason with his antinomies, a classical
skeptical device intended to show the equal strength of contrary lines of
reasoning.'
Problems about the relation between mind and cognized objects also cut
across many periods and topics. Some of these dwell on the character of the
objects in question, and some on the nature of the relation. In the former
category there is the problem of mutability or change. Plato, following
Heraclitus, held that physical objects are constantly in flux, and felt that
such flux precludes knowledge. Only unchangeable, immutable objects such
as mathematical entities and Forms can be known. In more recent philoso-
phy just the opposite conclusion has been drawn: only physical objects, not
abstract entities, can be brought within our ken. In both cases, however, the
trouble arises from the possibility of error. Change is a problem because the
real properties of a changing object are difficult to d e t e ~ tAbstractness
.~ is a
problem because it is hard to see how the mind can get reliable information
about an abstract entity, which inhabits a nonspatiotemporal realm. '
liefs. These rivals include Descartes's dreaming and evil genius hypotheses
as well as their modem, brain-in-a-vat counterpart. These rival hypotheses
are all intended to raise doubts about our warrant in believing standard, ex-
ternal world propositions. These ploys might be construed as demonstrating
the possibility of error. But, alternatively, another style of argumentation
can be attributed to this sort of skeptic.
The skeptic might be arguing that a person is justified in believing p only
if there is no other hypothesis q which explains the phenomena as well as p.
The evil genius and brain-in-a-vat hypotheses might be adduced to show
that for any selected external world proposition p, a rival hypothesis can al-
ways be constructed that is equally explanatory. This kind of strategy is
often used in the philosophy of science. If, for any scientific theory T, a
nonequivalent but equally explanatory theory can be constructed, perhaps
we are never justified in believing any scientific theory.''
I suspect that the rival hypothesis stratagem is more commonly moti-
vated, at least in historical writings, by the possibility-of-error worry. Cer-
tainly Descartes, in the first Meditation, seems to have been thinking along
these lines. The other construal of the rival hypothesis stratagem is men-
tioned for the sake of completeness. As the discussion proceeds, however, I
shall pay much greater attention to the error possibility category.
truths are a matter of proofs that have been or can be constructed, then
these truths should be within human ken.
The general pattern here is plain. Metaphysical reductionism is attrac-
tive because the reducing entities are thought to be more epistemically ac-
cessible than the sorts of entities being reduced or eliminated. The
reduction is intended to avoid the skeptical conclusion that we cannot have
knowledge of the specified domain. The epistemic theme, however, is not
always knowledge or justification. It might feature other epistemic notions,
even a quasi-epistemic notion like reference. Hilary Putnam advocates a
metaphysical view he calls "internal realism," which seems principally de-
signed to avoid a skeptical threat of no determinate reference relation.13
The attractiveness of the metaphysical reductionist strategy depends on
two things: first, how plausible the advocated reduction is; and second,
whether the advocated reduction really facilitates knowledge of the target
domain. Most historical reductionisms have foundered on the first point.
The reduction of physical objects to ideas, or sense data, has not fared well;
nor did Kant's efforts to interiorize the domain of mathematics. In general,
reductive metaphysics has usually proved more costly in overall philosophi-
cal perplexity than the epistemological benefits it promised. So cost-benefit
analysis has turned the tide of philosophical opinion against it. Frequently,
moreover, the epistemological fruits have not proved as sweet as billed.
None of the twentieth-century movements in the philosophy of mathemat-
ics, for example, has been the epistemological panacea its proponents
proclaimed.
I would not make the sweeping claim that all metaphysical reductionism
is doomed to failure, that it is never advisable to use such a strategy to cope
with skepticism. But the track record of these strategies is not good, and my
own inclination is to steer in other directions. This is one reason why my
epistemological program does not venture deeply into metaphysics.
Let us turn next to the third category of response: conceptual (or linguis-
tic) analysis. Two different strains of this response must be distinguished: (a)
analysis of the sentences (or statements) in various special domains, and (b)
analysis of key epistemic terms, such as 'knowledge' and 'warrant'.
The first kind of conceptual analysis breeds many of the same kinds of
theory as metaphysical reductionism, only the theories assume a linguistic
form. Thus, linguistic phenomenalism of the twentieth century sought the
same style of refuge from skepticism about the external world as Berkeleyan
idealism and Millian phenomenalism. Similarly, logical behaviorism was a
linguistic reductionism, which (among other things) aimed to salvage the
knowledge of other minds. Another kind of conceptual reduction was
Hume's treatment of cause and effect. By tracing the idea of causation to
constant conjunction, Hume may have hoped to preserve the possibility of
knowledge of causation (though this interpretation does not sit well with
36 Skepticism
Let me further illustrate how conceptual analysis can help answer skep-
ticism, and how different proposed answers guide epistemology down alter-
native paths of inquiry. One approach to justification says (roughly) that a
belief is justified if the believed proposition better explains what it purports
to explain than any rival proposition.15Furthermore, this 'best-explanation'
approach may be presented as a theory of evidence. That is, it may claim
that evidential support for a proposition is to be understood in terms of the
position of that proposition in an explanatory network. Two comments are
in order about this argument. First, if the account is right, it stands a good
chance of resolving certain skeptical challenges, specifically, those that
arise from rival skeptical hypotheses. To make this response stick, though,
the epistemologist must show that ordinary physical object beliefs are bet-
ter explanations of the phenomena than, say, the Cartesian demon hypoth-
esis or the brain-in-a-vat hypothesis. This will necessitate a satisfactory
theory of explanatoriness. (How does simplicity, for example, affect degree
of explanatoriness?How is simplicity measured? And so on.) In addition, the
epistemologist will be called upon to defend the view that a best-explana-
tion account of evidence is the right sort of theory of evidence. My second
comment, then, is that this sort of theory of justification not only bears on
skepticism but also bears on the direction of one's epistemological inquiries.
Obviously, this approach to justification points more in the direction of the
theory of evidence (the fifth response listed above) than in the direction of
psychological scrutiny of cognitive mechanisms.
By contrast, suppose an account of justification asserts that a belief is jus-
tified if it results from reliable cognitive processes. Such an account is rele-
vant to various forms of skepticism, specifically those forms of skepticism
that impugn our cognitive faculties. If this account of justification is correct,
then the skeptic may be right in insisting that our beliefs are justified only if
our cognitive faculties are not excessively error-prone. A full account of
justification would shed light on what is excessive or tolerable in the way of
fallibility. (Total infallibility is presumably a misplaced demand.) Finally, .
there would be the question of whether human cognitive processes are suf-
ficiently reliable. This question, in turn, breaks into two components: (a)
What cognitive processes do human beings in fact have? and (b) What is the
reliability of these processes? At least question (a) falls in the province of
psychology. Hence this account of justification would lead epistemology
down a path that fosters collaboration with cognitive psychology. And it
leads to a version of the fourth response to skepticism (the defense of cogni-
tive powers).
Conceptual analysis is sometimes deployed against skepticism in a much
more direct fashion. Some philosophers try to show that skepticism can be
defused by purely conceptual dissection of the skeptical position. I have in
mind here so-called transcendental arguments, at least semantic versions
38 Skepticism
assured (on this theory) that some satisfactory processes are available. But
how many such processes are there? And which ones are they?
Notice that even in the history of epistemology, the central issues have
not always been couched in terms of skepticism. Consider the rationalist-
empiricist controversy, for example. This was a debate not so much over
whether knowledge can be attained as over how it can be attained: by
principal reliance on the senses, or by greater reliance on reason? Rational-
ists questioned the ability of the senses and induction to deliver the goods;
they advocated the power of innate concepts and principles, intuition and
demonstrative proof. Empiricists challenged the existence of innate ideas
and a priori principles; they doubted the ability of pure ratiocination to
achieve intellectual ends. This debate continues today among neorational-
40 Skepticism
ists and their opponents. Both parties to the dispute can cheerfully admit
that knowledge is attainable (or perhaps that knowledge in a strong sense is
not attainable). The dispute centers on the means by which knowledge--or
true belief-is to be attained.
Now let us assume that global skepticism wins the day. Would that spell
the end of epistemology? Not necessarily. Even if there are not any suffi-
ciently reliable cognitive processes to qualify a person for either knowledge
or justified belief, there might still be differences among processes in de-
grees of reliability. A worthy epistemological task is to identify the cornpar-
atiuely reliable processes, to discriminate the better from the worse. This
would indicate how to maximize one's truth ratio, or maximize the proba-
bility of getting a high truth ratio. Even if epistemic honors like knowledge
or justified belief are beyond our grasp, that does not mean all distinctions
should be abandoned.
The next point to notice is that knowledge and justifiedness, at least on
my accounts, critically invoke one particular standard of appraisal: reliabil-
ity. But that is not the only standard of epistemological interest. Even if no
cognitive processes are sufficiently reliable to confer knowledge or justi-
fiedness, indeed, even if all are equally unreliable so that none stands out
above its peers, distinctions could still be drawn according to other stan-
dards. I have already mentioned the standards of power and speed. Proce-
dures might differ in the range of questions to which they can deliver
correct answers and in the speed of getting these answers. Such differences
are certainly of epistemic interest, even if the reliability of all processes is
too low to qualify for knowledge or justifiedness.
Another dimension of appraisal possibly orthogonal to skepticism is
rationality. This is a widely cited epistemic desideratum, one that may be
feasible for human beings even if both knowledge and justifiedness are not.
It may be possible to have rational beliefs even if knowledge is unattain-
able. This point is often taken for granted in the dominant philosophy of sci-
ence tradition, which commonly pursues questions of scientific rationality
without concern for standard problems of skepticism.
It should be noted, in this connection, that the weapons in the skeptical
arsenal are usually best equipped for demolishing the epistemic qualifica-
tions of full-fledged beliefs. When it is claimed that we cannot know, this is
often on the grounds that we do not have a right to be sure or completely
confident. It is more difficult for the skeptic to provide grounds for saying
that we have no right to even partial beliefs, say, to subjective probabilities
of .60. Theories of rationality often try to lay down conditions for rational
subjective probabilities. And such conditions may be important for episte-
mology even if prospects for knowledge are f o r e c ~ o s e dI. ~shall
~ not myself
offer any theory of rationality. But I shall tentatively tender an account of
Skepticism 41
Knowledge
not know he is in brain state B, although his belief to this effect is a reliable
indicator of the truth.4
It is worth stressing that even a stronger variant of the reliable-indicator
condition would not suffice to capture knowledge. Even if the indicator re-
lation is strengthened from nomological to logical sufficiency, the desired
result is not secured. Even if the having of a given belief 'logically' guaran-
tees the truth of the belief, the belief may not qualify as knowledge.
Suppose young Humperdink considers a proposition having fifteen dis-
juncts, the eighth of which is 'I exist'. Schematically, then, the proposition
is: '1 or 2 or . . . or I exist or . . . or 15'. Now Humperdink is not logically
sophisticated enough to realize that this proposition is entailed by 'I exist',
so he does not believe it for this reason. Nor does he believe it because he
already knows (or believes) one of the other fourteen disjuncts to be true.
He only believes it because he is inclined to believe any complex proposi-
tion with ten or more disjuncts. The very thought of such a complex propo-
sition overtaxes his meager mental powers and induces belief. If these are
the facts, one would hardly say that Humperdink knows this fifteen-disjunct
proposition to be true. Nonetheless, his believing it is logically sufficient for
its truth. This is because his believing it logically guarantees that he exists,
which in turn logically guarantees that the disjunctive proposition is true.
Thus, his belief is a logically reliable indicator of the truth. Yet this does not
raise the belief to the status of knowledge.
Once again, the process that causes the belief is critical to the knowledge
question. Since the causal process in Humperdink's case is just being over-
whelmed by sheer complexity, and since this is an unreliable process, the
resulting belief fails to qualify as knowledge. The fact that the belief hap-
pens to be a reliable indicator of the truth is irrelevant to the knowledge
question.
of uses for which the process is reliable. Global reliability is reliability for
all (or many) uses of the process, not just its use in forming the belief in
question. Local reliability concerns only the reliability of the process in the
context of the belief under assessment. However, this might include its reli-
ability in certain counterfactual situations centered on the target belief.
This brings us to a second distinction among reliable-process approaches:
the cu:twl-counterfcu:tual distinction. Some approaches might invoke the
process's reliability only in actual applications. Other approaches might in-
voke its reliability in counterfactual situations as well.
If counterfactual applications are allcwed, a third distinction comes into
play. On the one hand, the admissible counterfactual situation may be re-
stricted to the one determined by the pure subjunctive: what would happen
if proposition p were false? Alternatively, more counterfactual situations
could be invoked, for example, all situations involving relevant alternatives
to the truth of p.
Combinations of options chosen from this menu determine an array of
reliable-process approaches. Let us see how current versions of reliabilism
fit into the resulting classification. Robert Nozick's version of reliabilism-
what he calls the "tracking" theory--can be seen as a choice of local relia-
bility, in counterfactual applications, as specified by the pure subjunctive.'
In first approximation his account features the following four conditions for
knowledge: (1)p is true, (2) person S believes that p, (3)if p weren't true, S
wouldn't believe that p, and (4) if p were true, S would believe that p. No-
zick adds the proviso that beliefs in the counterfactual situation(s) must re-
sult from the same method M, used in the actual situation.
Condition (3)is obviously the critical one in this account, and introduces
the local subjunctive as the linchpin of the analysis. Unfortunately, the local
subjunctive is too weak. Here are a few examples to show why.
Suppose a parent takes a child's temperature and the thermometer reads
98.6 degrees, leading the parent to believe that the child's temperature is
normal, which is true. Suppose also that the thermometer works properly,
so that if the child's temperature were not 98.6, it would not read 98.6 and
the parent would not believe that the temperature is normal. This satisfies
the first three conditions of Nozick's analysis. Presumably the fourth condi-
tion is also satisfied: in (close) counterfactual situations in which the child's
temperature is normal, the parent would believe that it is normal.
But now suppose there are many thermometers in the parent's medicine
cabinet, and all but the one actually selected are defective. All the others
would read 98.6 even if the child had a fever. Furthermore, the parent can-
not tell which thermometer is which; it was just luck that a good thermome-
ter was selected. Then we would not say that the parent knows that the
child's temperature is normal, even though Nozick's analysis is satisfied.
Another case will solidify the point. Suppose S sees a smiling stranger on
46 Knowledge
the street, and concludes that the man just won a lottery. His only basis for
this belief is that the stranger is smiling. S knows nothing else about him
that forms the basis for this belief. As it happens, this stranger would not be
smiling unless he had just won a lottery; the only thing with the power to
make him smile is a lottery prize. So if he had not won a lottery prize, the
cognizer would not believe that he had. According to Nozick's account, S
knows that this man just won a lottery. But intuitively that is not right.
Prior to Nozick, Fred Dretske offered a very similar analysis, although
the formulation of his analysis is slightly more complex.s However,
Dretske's analysis succumbs to the same sorts of counterexamples as I have
just presented.'
Another version of reliabilism was presented in my paper "Discrimina-
tion and Perceptual ~ n o w l e d ~ e . "That
' ~ paper also invokes counterfactual
situations in accounting for knowledge attributions; but it does not employ
the pure subjunctive. It says (roughly) that a true belief fails to be knowl-
edge if there are any relevant alternative situations in which the proposition
p would be false, but the process used would cause S to believe p anyway. If
there are such relevant alternatives, then the utilized process cannot dis-
criminate the truth of p from them; so S does not know.
This relevant alternatives account works better than the pure subjunctive
one. The pure subjunctive account is just too permissive. Suppose Sam spots
Judy across the street and correctly believes that it is Judy. If it were Judy's
twin sister, Trudy, he would mistake her for Judy. Does Sam know that it is
Judy? As long as there is a serious possibility that the person across the
street might have been Trudy rather than Judy (even if Sam does not realize
this), we would deny that Sam knows. This is handled properly by the rele-
vant alternatives account, since the possibility of its being Trudy would
presumably qualify as a relevant alternative. But the pure subjunctive ac-
count would not necessarily give the correct verdict. We may suppose that
what would be the case if Judy weren't across the street is that nobody
would be across the street; or we may suppose that if Judy weren't there,
her colleague in the mathematics department would be there. (Say they had
flipped a coin to see who would fetch lunch, and Judy lost. If the colleague
had lost, he would be the person walking down the street.) In these coun-
terfactual situations, Sam would not believe that Judy is across the street. So
Sam's belief would survive the pure subjunctive test, and that account
would incorrectly say that Sam knows.
Notice that the no-relevant-alternatives condition can handle Gettier and
post-Gettier cases." Consider Brian Skyrms's case in which a pyromaniac
believes that the next Sure-Fire match he strikes will light.'' It is true that it
will light, but not because of the usual chemical condition of previously en-
countered Sure-Fire matches. This particular match is defective, but it will
light because of a coincidental burst of Q-radiation. Here the pyromaniac
Knowledge 47
does not know that the match will light. Why not? He does not know be-
cause there is a relevant alternative situation in which the pyromaniac's
evidence is exactly the same but the match does not light, that is, a situation
in which the match is defective (as in the actual situation) but there is no
burst of Q-radiation. Other Gettier and post-Gettier cases can be handled
similarly, as I will show in section 3.4.13
In "Discrimination and Perceptual Knowledge" I emphasized what I
here call a local reliability requirement for knowledge. But I did not mean
to endorse local reliability to the exclusion of global reliability. Indeed, in
other papers I have made the global reliability requirement quite explicit.14
Notice that local and global reliability are not mutually exclusive. A theory
can require both, and I mean to require both in a theory of knowing. To
qualify as knowledge, a true belief must result from a generally reliable pro-
cess, not just one that is (counterfactually) reliable for the case in question,
Another variant of reliabilism, by contrast, selects global reliability to the
exclusion of local reliability. This is Colin McGinn's recent adaptation of
the discrimination approach.'5 McGinn's theory also contrasts with the two
kinds of theories discussed thus far in eschewing counterfactuals. In terms of
our classification, it selects global reliability; purely factual, not counter-
factual, situations; and the no-relevant-alternatives option (although this
third option does not invoke counterfactual alternatives).
The discrimination theme I stressed in "Discrimination and Perceptual
Knowledge" calls attention to the fact that one sense of 'know' listed in the
Oxford English Dictionary is "to distinguish (one thing) from (another)."
Examples of such a sense are "I know a hawk from a handsaw" (Hamlet),
"We'll teach him to know Turtles from Jayes" (Merry Wives of Windsor), 'I
don't know him from Adam', and 'He doesn't know right from left', as well
as other current sayings that readily come to mind. The conjecture is that
the propositional sense of 'know' is related to this underlying meaning, in
that knowing that p involves discriminating the truth of p from relevant al-
ternatives. On my development of this theme, the alternatives are counter-
factual alternatives. McGinn develops the discrimination theme slightly
differently. He wants us to consider the truth-discriminating power of the
belief-forming process (or 'method') within some relevant class R of propo-
sitions, where R is not restricted to proposition p and contraries of p. We are
to look at the propensity of the process to form true beliefs across a range
of propositions whose truth-values are taken as fixed in the actual w o r ~ d . ' ~
One of McGinn's examples that motivates his approach, especially his es-
chewal of counterfactuals, is this: You visit a hitherto unexplored country in
which the inhabitants have the custom of simulating being in pain. You do
not know that their pain behavior is mere pretense, and so you form the be-
lief, of each person you meet, that he or she is in pain. In this way you ac-
quire many false beliefs. One person, however, is an exception to the cus-
48 Knowledge
tom of pain pretense: this hapless individual is in constant pain and shows
it. You also believe of this person, N, that he is in pain. In this case your true
belief is not, intuitively, knowledge. But the relevant counterfactual 'If N
were not in pain, you would not believe that N was in pain' is true, since if
N were not in pain, then (unlike the pretenders), he would not behave as if
he was, and so you would not believe that he was. Since this counterfactual
is true, it cannot accommodate the absence of knowledge in this case. So a
noncounterfactual version of reliabilism, McGinn surmises, is needed to
handle it.
Notice, though, that this case is only a problem for the pure subjunctive
variant of the counterfactual approach. The relevant alternatives variant
can handle it. A relevant alternative possibility is one in which N is not in
pain but is a pain simulator, like the other people around him. In this coun-
terfactual situation you would falsely believe that he is in pain, which is
why you do not know in the actual situation.
McGinnYsdismissal of counterfactual conditions is unconvincing for other
reasons. First of all, he does not offer a fully formulated analysis. He does
not tell us exactly what the discriminative range of the belief-forming pro-
cess must be in order to be capable of knowledge production. Moreover,
there seems to be a general pitfall for his strategy. A process might be
wholly reliable in actual situations. This would apparently imply that every
true belief it produces counts as knowledge. But that seems implausible:
there might still be cases in which counterfactual possibilities defeat a
knowledge ascription. My worry on this score cannot be nailed down in the
absence of a fully formulated analysis from McGinn; but the outlines of his
approach incur this probable difficulty."
Although I differ from McGinn on the appropriateness of a counterfac-
tual requirement, especially a local requirement, we are in agreement on
the need for a global reliability requirement. This comes out clearly for
knowledge of necessities, such as the truths of mathematics. Can such
knowledge or nonknowledge be handled by invoking such counterfactuals
as 'if it were not the case that 7 + 5 = 12, S would not believe that 7 + 5 =
12'? One problem is that we have no adequate account of counterfactuals
with impossible antecedents. Another problem is that it is not clear that
such a counterfactual is true in the case of someone who does know that 7 +
5 = 12. So Nozick's third condition might be violated by someone who
knows this proposition. (The no-relevant-alternatives approach would not
be imperiled, though, because there may be no alternative in which 7 5 +
does not equal 12.) Also, while Nozick says that his fourth condition is de-
signed to handle true beliefs in necessities that do not qualify as knowledge,
McGinn shows that this claim does not succeed."
How should we handle true beliefs in necessities that do not qualify as
knowledge? My suspicion is that they are mostly cases in which the belief-
Knowledge 49
Presumably not. For consider a case in which the perceptual input is not
degraded, and the actual degree of match to the template is, let's say, .99
(on a scale of 0 to 1).Then presumably we will want to say that this is ade-
quate for knowledge (if everything else goes well). But if the selected pro-
cess type still includes the minimal value T, the type as a whole may not
have sufficient reliability.
My proposed account would handle this case by noting that the mecha-
nism's property of having T (say, .70) as minimally sufficient is not causally
operative in this case. The property of having this threshold value does not
play a critical causal role in eliciting the belief. The degree of match in this
case is actually .99. So the critical aspect of the mechanism's functioning
that produces the belief is the propensity to produce a belief when the de-
gree of match is .99. If this property, rather than the others, is included in
the selected process type, an appropriate degree of reliability is chosen that
meshes with our knowledge ascription intuition.
Clearly, this proposal needs to be developed and refined, but I will not
try to do that here. I present it only as a promising lead toward a solution of
the generality problem.
the neurosurgeon is wearing a yellow smock, and so on. Now these beliefs
are all, in fact, true. Moreover, they are formed by the usual, quite reliable,
perceptual processes. But are they specimens of knowledge? Intuitively, no.
The reason is that Millicent is not justijied in holding these beliefs; they
contravene her best evidence. It seems, then, that causation by reliable
processes is not sufficient for knowing.
Another example is this: Maurice uses a reliable-but not perfectly reli-
able-heuristic to arrive at a certain belief p. Now the fact that the process
is not perfectly reliable does not by itself preclude knowledge. I do not as-
sume that perfect reliability is required. But there is another heuristic
Maurice knows, which is more reliable than the first (though still not per-
-feet), and Maurice believes it is more reliable. He even suspects, in this case,
that the better heuristic might yield a different result, since it has differed
from the first in similar cases before. But despite these beliefs, Maurice ne-
glects the superior heuristic. Had he used it, it would indeed have led him
to a different conclusion: to believe not-p rather than p. But in this particu-
lar case the first heuristic gets things right: p happens to be true. Does
Maurice know p? Intuitively, no. He does not know, once again, because bis
belief in p is not justified. He should have consulted the superior heuristic.
If knowledge requires justifiedness, as these cases suggest, I cannot com-
plete my account of knowledge without a theory of justifiedness. But does it
follow that the reliable-process theory is on the wrong track entirely? That
justifiedness is a wholly different element, quite out of spirit with the
themes I have stressed until now? That of course is possible. It might bk
supposed that once the justificational element is introduced into knowl-
edge, it will change our picture of knowledge dramatically. But that is not
what I shall argue. On the contrary, my account of justified belief will fea-
ture some of the same ingredients discussed thus far, though couched in a
different format. Further requirements will be added, however, to handle
the foregoing cases (especially the Millicent case).
The reader may wonder whether my account handles the standard exam-
ples in the post-Gettier knowledge trade. What about the original cases in
which the (true) belief in p is (justifiably) inferred from false premises? And
what about the so-called social dimensions, where things known to other
cognizers, but not to S, may keep S from knowing p? These cases can be
handled, I believe, by the local reliability requirement, the no-relevant-al-
ternatives condition.
Suppose S believes that either Jones owns a Ford or Brown is in Barce-
lona, and his justification consists in evidence for Jones's owning a Ford. In
fact, Jones does not own a Ford, but by sheer coincidence Brown is in Bar-
celona. S does not know that either Jones owns a Ford or Brown is in Bar-
celona, although this is true.% Why doesn't he know? s's true belief is not
knowledge because it is defeated by the following relevant alternative,
Knowledge 55
which is compatible with all his evidence: Jones doesn't own a Ford and
Brown isn't in Barcelona. This counterfactual situation is one that S cannot
discriminate (with his evidence) from the actual state of affairs.
Similarly, consider the case in which you apparently see Tom Grabit steal
a book from the library. Unbeknownst to you, Tom's mother has said that
Tom is miles away while his twin brother, John, is in the library. This may
be enough to defeat any claim on your part to know that Tom stole the
book, even if it is true." My explanation of this case is that you cannot dis-
criminate this truth from John's stealing the book. Furthermore, the latter
alternative seems to be relevant, as long as Mrs. Grabit says what she says.
When it transpires, however, that there is no twin John-he's only a fig-
ment of Mrs. Grabit's demented imagination-this ceases to be a relevant
alternative, and you can be credited with knowing. In this fashion, the no-
relevant-alternatives element of my account seems able to handle Gettier
and post-Gettier examples.
There is, however, a related argument for skepticism not handled by this
reply: the problem of closure. It is an initially plausible principle that
knowledge is closed under known logical implication. If I know that p and I
know that p logically implies q, then I must know that q. In the present case
let p be that there is a keyboard in front of me, and let q be that it's not the
case that I am a mere brain in a vat . . . Since I know that p logically implies
q, I only know p if I know q. But do I know q? Do I know that the envat-
ment hypothesis is false? It does not seem as if I discriminate that from its
relevant alternatives. So it seems as if I do not know q. But that seems to
imply that I do not know p.
Dretske and Nozick have argued, plausibly, against the principle of clo-
sure.* To legitimate the rejection of closure, though, one needs an account
of knowledge that sustains this rejection. But, as McGinn points out, the dis-
crimination account does sustain n o n c l o s ~ r eDifferent
.~~ propositions carry
with them different requirements as to the discriminative capacities neces-
sary for knowledge of them. And knowledge of logically weaker proposi-
tions can require greater discriminative powers than knowledge of logically
stronger propositions. It can be easier to know p than q although p implies q
(and not vice versa) because q requires more in the way of discrimination
than p. To know there is a keyboard before me I needn't discriminate this
state of affairs from envatment. But to know I am not envatted, I do need to
discriminate this state of affairs from envatment. So I may not know I am
not envatted. Yet this possibility does not preclude my knowing there is a
keyboard before me.
The kind of knowledge skepticism I have principally examined is direct
skepticism, which claims that it is impossible to know. But there is a differ-
ent sort of question, concerning iterative skepticism (see Chapter 2), that
may perturb many people regarding the reliable-process approach.
Granted that such a theory leaves open the possibility of knowledge, does it
allow us to know that we know? Many discussions in epistemology go astray
in failing to distinguish questions of first-order and higher-order knowledge
(or ju~tification).~'But once the distinction is carefully made, there is no
objection to raising questions and doubts about the prospects for higher-
order knowledge, and the implications different theories have for such
knowledge.
Is the reliable-process theory worse off than other theories vis-a-vis
higher-order knowledge? No. At least it isn't worse off than a good theory
ought to be. We have already seen that it makes knowledge a logical possi-
bility. By the same token, it makes it logically possible to know that one
knows. To know that we (sometimes or often) know, we would have to
know that we (sometimes or often) use reliable processes of belief forma-
tion. But since the analysis makes it (logically) possible for us to know what
processes we use, and makes it (logically) possible for us to know all sorts of
Knowledge 57
truths about the world (which is essential for knowing the reliability of our
processes), the analysis makes it possible for us to have higher-order knowl-
edge. To be sure, the analysis does not entail the actual existence of higher-
order knowledge. It does not even entail that if we know, then we know
that we know. But no analysis of knowledge should entail this. A plausible
theory ought to have the property that knowing that one knows is more
difficult than simply knowing. It is a virtue of this theory that it has this
property.
Granted that my analysis allows the possibility of both first-order and
higher-order knowledge, what does it say, or imply, about whether we do in
fact know (very much of the time)? As indicated earlier, the analysis per se
does not speak to this question. It does imply, though, that we do know only
if we have reliable (and higher-order reliable) cognitive processes. It follows
that to find out whether we know, we need to ascertain the properties of
our native cognitive processes. This is where psychology enters the picture.
Psychology can (in principle) tell us about the nature of our cognitive pro-
cesses. When these processes are spelled out, we can try to determine their
reliability. (It is doubtful, though, that psychology alone can determine
their reliability. For this we need help from a variety of intellectual fields,
probably including logic and statistics.)
Notice that psychology is needed not merely to tell us whether we do
know, but whether it is humanly possible to know. The reliable-process the-
ory of knowing entails the logical possibility of knowledge, but it does not
entail that knowledge is humanly possible. It is humanly possible only if
humans have suitable cognitive equipment. And this is something of which
we can best be apprised only with the help of psychology.
In saying that whether we do or can know depends on psychological
facts, I partially concur with Quine when he says that "sceptical doubts are
scientific doubt^."^' But my agreement is only partial. Some, but not all,
skeptical doubts are scientific doubts. Some routes to skepticism arise from
concern over the propriety of crediting someone with knowledge if certain
logically possible alternatives cannot be excluded. The best way to counter
this skeptical maneuver is through a satisfactory analysis of knowledge, not
through psychology or other branches of science. (Of course, any piece of
linguistic analysis is plausibly viewed as applied psycholinguistics, and
hence as part of science. This is not what Quine had in mind, though, given
his strictures about analysis and analyticity.) But this sort of maneuver, as I
have discussed, can be met by the reliable-process account.
chapter 4
will play an important role. Meanwhile, I can strengthen (PI) by adding the
no-undermining condition, thereby yielding the following principle:
(P3) S's believing p at t is justified if and only if
(a) S's believing p at t is permitted by a right system of J-rules, and
(b) this permission is not undermined by S's cognitive state at t.
Although (P3) is the real framework principle I mean to endorse, in most
of what follows I will be mainly interested in clause (a), in other words, in
principle (PI). For convenience, then, I will often refer to (Pl) as the frame-
work principle, and I will make reference to (P3) only when it is specially
pertinent.
With these general comments in mind, let us return to the list of candi-
date criteria and look briefly at some of their characteristics and problems.
This is only a preliminary examination, to get a feel for possibilities and
difficulties.
Criteria (Cl) and (Cl") invoke truths of logic, or rules of logic, as the
basis for a J-rule system. This reflects a widespread view that principles of
logical inference provide justificational procedures. On this approach, then,
proper rules of justifiedness are to be generated by truths of logic, or by
someone omniscient about logical truths.
The first point to raise about such an approach concerns completeness. At
most, deductive logic can generate a proper subset of desired rules of justi-
fication; not all appropriate rules can be derived from it. Where nondeduc-
tive inferences are appropriate-and such cases are certainly
legion--truths of deductive logic are of little or no help. Are there truths of
inductive logic to fill the breach? That is very dubious. Many epistemolo-
gists and philosophers of science now question whether there is a body of
truths we can call 'inductive logic'.15 Perhaps probability theory can be ad-
duced to provide a relevant class of truths. But, again, the epistemological
pertinence of probability theory is complicated and problematic.16 Pre-
sumably, some rules of justification should pertain to perceptual beliefs, and
some to beliefs involving memory. Neither truths of deductive logic nor of
probability theory can generate J-rules in these domains. Hence, criteria
(Cl) and (Cl') are inadequate on grounds of incompleteness.
There are, moreover, further fundamental weaknesses in (Cl) and (Cl").
They both assume-(Cl) most explicitly-that some J-rules can be derived
from truths of logic. But that is not so, as we shall see in Chapter 5. A typi-
cal truth of deductive logic is that modus ponens is a valid argument form.
But no (correct) rule of belief formation is derivable from this sort of truth.
Validity of argument forms has no immediate implication for belief-forming
practices.
It might seem as if belief-forming rules can be derived from logical
truths. Such rules could simply permit a cognizer to believe any proposition
in the corpus of logical truths. This is perhaps what (ClO)envisages: a logi-
cally omniscient cognizer formulating rules that allow people to believe, at
a minimum, what the cognizer himself believes, namely, all logical truths.
But this would have unacceptable implications. It would imply that when-
ever a person believes a truth of logic, the belief is justified. That is clearly
wrong. It is possible to believe a truth of logic unjustifiably. I will elaborate
on this point in Chapter 5.
Criterion (ClO)articulates a common stance in epistemology. The epis-
temologist is allowed to appeal to truths of logic, and perhaps probability
theory, in formulating methodological rules; but appeal to contingent prop-
ositions (except perhaps first-person experiential statements) is disallowed.
This idea bears some resemblance to Rawls's conception of the Original Po-
68 Justification: A Rule Framework
Proposal (C4) is next on our list. This criterion has the virtue of being ap-
plicable to all doxastic attitudes, not just beliefs. But it appears to be for-
mally inadmissible, since it employs the term 'evidence', an epistemic term.
This point is correct, but probably remediable. (C4) would need to be re-
vised so as to indicate, in purely factual, nonepistemic terms, what the class
of evidence beliefs are. There are also other objections to be raised against
(C4). I shall present these objections early in Chapter 5.
Criterion (C5)belongs to a family of possible criteria that invoke truth. It
is what I call a 'truth-linked' criterion. One possible objection to (C5) is an
objection against truth-linkedness in general. Is it appropriate to invoke
truth in a criterion of J-rule rightness? Won't this raise a serious problem of
determining which system(s) of J-rules satisfies the criterion? Won't it be
either impossible or circular to use such a criterion? Since I shall ultimately
advocate a truth-linked criterion (or family of criteria), this is an important
issue. But I must defer its full treatment until the end of Chapter 5. The ra-
tionale and plausibility of a truth-linked criterion will be spelled out gradu-
ally as I proceed. It should be recalled, however, that a criterion is not
intended as a test or verification procedure for J-rule rightness. So the pos-
sible difficultyof determining which J-rule system(s) satisfies a truth-linked
criterion is not an insuperable objection.
Apart from truth-linkedness, (C5) is not a proposal I deem acceptable.
The substance of ((25)-maximizing the number of true beliefs-just isn't
the correct ingredient for justifiedness. Also, (C5) omits elements that an
acceptable criterion must include. But these will not be explored until
Chapter 5.
In short, I do not endorse any of the proposals in this sample list. Still,
they are useful illustrations, to give a sense of the options available and the
considerations that may be brought to bear in assessing candidate criteria.
ble. I already gave some arguments against this in Chapter 1, but there is
more to say. Which other potential reason-receivers are relevant? 'What
if those persons are too stupid or stubborn to follow one's arguments or
abandon prior prejudices? One's inability to convince them might only
reflect badly on them, not on the believer. If nobody else had been capa-
ble of understanding Einstein's theory of relativity, Einstein would not
have been able to satisfy anyone with his reasons. But it would not follow
that Einstein was not justified in believing his own theory (at least once ap-
propriate observations had been made, such as those concerning the peri-
helion of Mercury). Collectively, these are strong reasons in favor of
the intrapersonal approach, and that is what I shallmainly pursue in this
Granted that I-rules are not literally entailed by truths of logic, can't we
construct J-rules on the basis of formal logic? If a certain argument is valid,
isn't a transition from belief in the argument's premises to belief in its con-
clusion justificationally permissible? So aren't (right) J-rules derivable in a
loose sense from facts of validity? No.
Before defending this answer, I need to expand framework principle (Pl),
to make it more determinate. This should be done, I believe, by reformulat-
ing it as a historical principle:
(Pl') A cognizer's belief in p at time t is justified if and only if
it is the final member of a finite sequence of doxastic states of the
cognizer such that some (single) right J-rule system licenses the
transition of each member of the sequence from some earlier
state(s).
Of course, our real framework principle, (P3), should also be recast in this
manner, yielding (P3'). First let us distinguish two sorts of J-rules: belief-
independent and belief-dependent J-rules. Belief-independent J-rules (pos-
sibly an empty class) permit beliefs to be formed independently of prior be-
lief states (or other doxastic states). They might include perceptual J-rules
that permit beliefs as a function of sensory cognitive states only. (However,
perceptual J-rules might make provisions for prior beliefs, thereby becom-
ing belief-dependent rules.) Belief-dependent J-rules permit beliefs as a
function of prior beliefs, and perhaps other prior doxastic states as well.
They would include rules that permit inferences from old beliefs to new,
and rules that permit retention of old beliefs from one time period to an-
other.' Now according to (PIQ),the justificationally relevant history of a
belief may reach back quite far. If a belief is inferred from previous beliefs,
whose permissibility is in turn traceable to still earlier beliefs, and so on, the
justificational history may be extensive.
Let us now return to the question of deriving J-rules from logic. Before
defending my denial of such derivability (even in a loose sense of 'derive'), a
technical point must be noted. Although I have introduced rightness as a .
predicate of J-rule systems, I also want to discuss, especially here, rightness
of individual rules. The relationship is straightforward: an individual J-rule
is right if and only if it belongs to some right J-rule system.2
I have two arguments against (loose) derivability. The first has been
stressed by Gilbert arm an.^ Suppose p is entailed by q, and S already be-
lieves q. Does it follow that S ought to believe p; or even that he m y be-
lieve p? Not at all. Even if he notices that q entails p, it is not clear that an
appropriate move is to add p to his belief corpus. Perhaps what he ought to
do, upon noting that q entails p, is abandon his belief in q! After all, some-
times we learn things that make it advisable to abandon prior belief^.^ We
might learn that prior beliefs have a certain untenable consequence. For
example, you might discover that your prior beliefs jointly entail that you
84 Justijication and Reliability
don't have a head. Instead of concluding that you don't have a head, you
may be better advised to infer the falsity of some prior belief or beliefs.
We see, then, that the mere validity of an argument-for example, < q,
therefore p>--does not imply that a believer in its premise or premises
should become a believer in its conclusion. This should come as no surprise.
An inconsistent set of premises implies every proposition. But this does not
mean that someone who believes an inconsistent set of propositions should
(or even may) form a belief in any random (further)
This point can be explained as follows (though it need not depend on this
explanation). Rules for forming beliefs should promote the formation of true
beliefs. Drawing conclusions by valid inference patterns promotes this goal
if prior beliefs are true. But prior beliefs are not always true; nor is it likely
that they will always be true even if they are formed in accordance with
right rules. (Even right rules are likely to allow some falsehoods to sneak in.)
So it does not follow that right rules should always license the acceptance of
any logical consequence of prior beliefs.
In view of this consideration, J-rules like the following are incorrect
J-rules:
If <q, therefore p> is a valid argument, then M (Bq / Bp).
If this is the envisioned pattern by which J-rules are derivable from truths of
logic (truths of validity), logic would only generate incorrect J-rules.
The second argument for the nonderivability of (right) J-rules from logic
runs as follows. Again let us examine the candidate rule that permits forma-
tion of belief in any logical consequence of prior beliefs. I now raise a differ-
ent kind of difficulty for the rightness of such a rule.
Suppose Claude believes proposition x, at which he arrived earlier by
transitions permitted by a right J-rule system. In other words, given (Pl*),
Claude's belief in x is justified. Suppose further that x entails y, but this en-
tailment is extremely complex (since, say, both x and y are very complex).
Even a sophisticated logician would have a hard time determining that <x,
therefore y> is a valid argument, and Claude is just a logical neophyte.
Claude does not 'see' that it is valid, prove that it is valid, or learn this on
anyone else's authority. Nonetheless, he proceeds to believe y. He 'makes a
transition' from a state of believing x to a state of believing y (as well). Is his
belief in y justified? Intuitively, no. But if the target J-rule were right, the
belief would be justified. Claude would have arrived at his belief in y via a
sequence of transitions permitted by right rules. Since the belief is not justi-
fied, the rule in question is not a right J-rule.'
The problem with the foregoing J-rule is that it permits any old transition
that accords with valid inference patterns. In particular, it places no restric-
tions on how the transition is made. I t does not require the cognizer to un-
derstand why the inference is valid, nor to see the connection between
Justification and Reliability 85
the validity belief in the rule means that if Claude's belief in y is to be justi-
fied, he must not only believe ex,therefore y> to be valid, but that belief
must be justified.
Does this ploy work? No, it has two problems. First, it is too strong, at
least if it is to serve as the prototype for deductive J-rules. Not all inferential
J-rules can include the provision that the cognizer believe in the inference's
validity. Children and adults untrained in formal logic presumably use in-
ference procedures before they acquire the concept of validity. Lacking the
concept of validity, they have no beliefs about validity. Nonetheless, they
presumably form justified beliefs by justification-transmitting inferences.
Second, the revised rule is still too weak, for the same reasons as its prede-
cessor. It still does not impose any requirements on the causal processes that
produce the successor belief states. In fact, it does not even require the
prior belief states to be causes of the successor state. This allows counterex-
amples to be readily constructed. For example, suppose a person (justifi-
ably) believes x and (justifiably)believes that x entails y, but the latter belief
is merely stored in long-term memory and is not presently activated. Under
these circumstances, he may start to believe y on a mere hunch. His belief
may not 'connect up' with his prior beliefs in x and the validity of <x,
therefore y>. If so, his belief in y is not justified, contrary to the implication
of the currently proposed J-rule.
Would it help to amend the rule to say that prior beliefs are 'activated'?
No, this would still leave it open whether a causal connection is established
between predecessor and successor beliefs; and it would leave open the de-
tails of any such connection. By leaving these matters open, similar coun-
terexamples as before are still constructible.
Since this kind of problem is intrinsic to the state-transition framework,
we really do need the previously proposed revision. To be right, J-rules
should not merely permit state transitions; they should permit selected
cognitive operations or processes-processes that sometimes output beliefs.
Thus, right J-rules should take the form of process permissions, which may
be represented schematically as
M (Proc 1), or
M (Proc 1 & Proc 2 & . . . & Proc N),
where the latter represents a complex process formed from elementary
processes by serial concatenation, parallel composition, or other composi-
tional devices.
The conclusions for which I am currently arguing are bolstered by fur-
ther reflection on a logic-based criterion and a state-transition rule format.
In Chapter 4 I noted the possibility of J-rules that grant unconditional belief
permissions. One such J-rule, mentioned in Chapter 4, is this:
If L is any truth of logic, then M (V / B(L)).
Justifcationand Reliability 87
This sort of ]-rule has been endorsed by many confirmation theorists (such
as Camap), though usually formulated in terms of 'rationality'. For exam-
ple, if one assumes that it is rational to place credence in a proposition to
the degree it is confirmed by one's total evidence, and assumes that a logical
truth is automatically confirmed to degree 1no matter what one's evidence,
then it is always rational to place degree of credence 1in any truth of logic.
Translating this into talk of (all or none) 'belief and 'justifiedness', we get
the principle that one is always justified in believing a logical truth, no
matter what one's prior state.
This J-rule runs into the same trouble as its earlier cousins. Since it places
no constraints on how a belief in a logical truth is produced, it implies, to-
gether with (Pl*), that such a belief is justified even if it is arrived at by
sheer hunch or guesswork. No matter how complex the proposition L, no
matter how naive the cognizer, no matter how weak or undeveloped his
cognitive capacity to interrelate L's constituents so as to determine its logi-
cal necessity, his belief in L is justified. This is clearly unacceptable. As be-
fore, the J-rule runs into trouble because it licenses beliefs irrespective of
the cognitive processes that produce them. Correct J-rules can only permit
belief formation via appropriate cognitive processes; this is what the newly
envisaged rule format proposes. But appropriateness cannot be specified by
pure logic. Logic is silent about cognitive processes, so it cannot yield cor-
rect J-rules.
Again, it might seem as if we could improve the foregoing J-rule without
reference to cognitive processes. One of the following two revisions might
be proposed:
If L is any self-euident truth of logic, then M (V / B(L)).
Inserting the qualifier 'obvious' improves the chances of the rule being cor-
rect. But it also surrenders the hope of providing a purely logic-based rule.
Again it is imperative to invoke a psychological notion, making tacit refer-
ence to human cognitive powers. This again reinforces the unavoidability of
reference to cognitive factors in getting accurate rules of justifiedness or
unjustifiedness.
An analogous point holds concerning possible permissive J-rules. Suppose
we waive Harman's problem about the relation between deductive validity
and right J-rules. Couldn't we then say that while arbitrary valid argument
forms do not generate right J-rules, s a n e valid argument forms do. In par-
ticular, doesn't modus ponens have a right J-rule associated with it? Now
this might be correct. But if it is, it is only because modus ponens represents
a fundamental, primitive cognitive operation. Only this psychological status
can make modus ponens have an associated right J-rule. Once more, what
makes for rightness is not just a matter of logic; it is also a matter of psychol-
ogy. After all, modus ponens is no more valid than other valid argument
forms. It is just psychologically simpler than others?
properly acquired (or properly tested after its original acquisition), it can-
not yield a justified belief. The mere fact of usage is not enough. We must
now determine what kinds of acquisition processes are proper. Of course,
some reference may again be made to learned algorithms or techniques
suitable for appraising the soundness of other methods. But the same ques-
tion will arise for them. What is their pedigree in the cognizer's mind?
Were they suitably acquired? Ultimately, the question of justifiedness must
rest on some unlearned processes, on some psychologically basic processes
that are never themselves encoded or stored as recipes, algorithms, or
heuristics. These processes are just executed, not 'applied' in the way that
an algebraic or statistical technique is 'applied'.
We arrive in this fashion at the same conclusion tentatively put forward
earlier. Justifiednesscentrally rests on the use of suitable psychological pro-
cesses; and by psychological processes I here mean basic, elementary pro-
cesses, not acquired techniques that are mentally encoded and applied."
Now suppose statistics indeed provides us with 'ideal' methods. (It is
questionable whether statistics incorporates methods of belief formation, or
credal-state formation. However, various forms of statistical inference do
seem to prescribe credal-state-forming methods. So let us allow this suppo-
sition to pass.) Should we then conclude that unless our native cognitive ap-
paratus instantiates this very set of methods, we never have justified beliefs?
This demand would almost certainly lead to skepticism (in the justification
theme), for it seems highly unlikely that our native cognitive architecture
would realize these methods. In fact, certain psychologists have recently
been engaged in arguing that native inference propensities do not match
those of 'normative' statistical methodology.'2 If our native processes do not
match these statistical methods, then, first of all, ordinary beliefs formed
without the methods cannot be justified. But, worse yet, even beliefs based
on method utilization cannot be justified. Utilization of the methods does
not confer justifiedness unless the methods are acquired by justification-con-
ferring processes. Ultimately such acquisition must utilize native processes
(including second-order processes). But, ex hypothesi, these native processes
do not include the 'ideal' statistical methods. So it will be humanly impossi-
ble to form justified beliefs.
Is the prospect of landing in the skeptical ditch enough to refute the
foregoing proposal? Perhaps not. That would involve too strong an anteced-
ent commitment to skepticism's falsity. But there are independent reasons
for thinking that the requirements of a refined, or optimal, procedure
should not be imposed at every level of cognition. Suppose we can imagin-
atively construct perceptual devices that are superior (say, in reliability) to
our own. Does this mean that none of our perceptual beliefs are justified?
Suppose we can devise a memory system that is better (again, say, in relia-
bility) than ours. Does it follow that none of our memory beliefs are justi-
Justification and Reliability 93
fied? Surely not. So why should the fact that we can design more refined in-
ference procedures than our native ones show that the latter cannot yield
justified beliefs? Another parallel may be instructive. Suppose that before
Russell presented his paradox, all existing set theories were paradoxical.
Does it follow that nobody before Russell had any justified beliefs about
sets? Surely not. Just because more refined conceptions of sets were later
generated is not enough to show that all earlier beliefs about sets were
unjustified.
All I have been arguing here is that the norms of ideal statistical method-
ology need not be imposed on basic cognitive processes. But it remains to be
seen just what the proper J-rules do require, and just what criterion of
rightness should be chosen to fix correct J-rules. Although I have here
fended off one specter of skepticism, I do not mean to preclude the possibil-
ity that skepticism might enter through another door. Once we choose a cri-
terion of rightness, it might still turn out that no J-rule system that satisfies
the criterion is humanly realizable.
One main result of this section is to emphasize the importance of the dis-
tinction between basic psychological processes, on the one hand, and vari-
ous sorts of algorithms, heuristics, or learnable methodologies, on the other.
Let us call the former simply processes and the latter methods. I shall now
argue that a good theory of justifiedness should systematically separate
these two classes. There is a family of different conceptions of justifiedness,
each having some pull on intuition, which can only be dissected with the
help of this separation.
5.4. Consequentialism
In earlier sections I endorsed a process format for J-rules, arguing that right
J-rules (at the primary level) can only permit basic cognitive processes. This
is really a partial specification of a rightness criterion for J-rules. An ac-
ceptable criterion must, at a minimum, require J-rules to be process rules,
96 Justificationand Reliability
not state-transition rules, for example. What further conditions should ap-
pear in an acceptable criterion remains to be seen. But let us pause and re-
flect on this constraint.
This constraint plays a critical role in vindicating the claim that psychol-
ogy has a role to play in the theory of justifiedness. If right J-rules must spec-
ify basic psychological processes, then cognitive psychology has an
important role to play in determining right J-rules. For cognitive psychol-
ogy (or, more broadly, cognitive science) is the field that explores the
human repertoire of basic cognitive processes. Unless and until these pro-
cesses are identified, no selection can be made of the ones that should be
licensed by a right rule system.
Of course, further conditions must be included in a rightness criterion, As
I develop my account of an acceptable criterion, the argument may con-
ceivably go astray at any point. But if my contention up to this point is cor-
rect-that right J-rules must be basic process rules-then I have already
substantiated one central theme: the importance of psychology for individ-
ual epistemology (at least at one level). I stress this point because the ratio-
nale for investigating psychological processes in Part I1 does not hinge on
all my epistemological theses of Part I. Although Part I1 does investigate
psychological processes with an eye to the truth-linked criteria I embrace in
Part I, a proper place for psychology is already established by the theses
defended thus far, independent of truth-linkedness. Thus even if the rest of
my account of justifiedness (and power) should fail, psychology's role is
secure.
One objection needs to be considered in this connection, though. The
proposed partnership with psychology in the epistemological enterprise-
specifically in the theory of justifiedness-assumes that part of epistemol-
ogy's task is the choice of specific J-rules. But some readers might take ex-
ception to this assumption, asking: Why shouldn't epistemology stop at the
selection of a rightness criterion? Why should epistemology include the job
of identifying the J-rules (or J-rule systems) that actually satisfy the
criterion?
There are three answers here. First, epistemology has traditionally been
concerned not simply with abstract criteria for good methodology or pro-
cedure, but with the endorsement of particular methods or procedures. This
is especially clear in the philosophy of science wing of epistemology. Since
psychological processes, I have argued, are the suitable class of 'procedures'
in primary justifiedness, it is appropriate for primary epistemology to ac-
tually specify the approved processes, not simply to lay down a criterion
they must meet.
Second, it is just part of a full theory of justifiedness to say which J-rules
(or J-rule systems) are right. To stop epistemology at stage 2, the level of the
criterion, is to abort it at an unnatural juncture. Admittedly, the job of arti-
Justificationand Reliability 97
culating a detailed J-rule system goes beyond the horizon of this book; none-
theless, that job is still an epistemological one.
Third, once a criterion of rightness is chosen, it may be an open question
whether any J-rules satisfy it. Perhaps no human psychological processes
meet the constraints of the criterion. In that case skepticism (in the justifi-
cation theme) is victorious: it is not humanly possible to have justified be-
liefs. But since the prospects for this sort of skepticism stand or fall with the
existence of nonexistence of J-rules satisfying the criterion, examining the
satisfiability of the criterion is certainly an assignment for epistemology.
With these points in mind, let us now work toward a more specific right-
ness criterion. The next step is to distinguish two possible approaches,
familiar from ethical theory: consequentialim and deontology. Consequen-
tialism holds that rightness is a function of consequences; in this case it
would hold that rule(-system) rightness is a function of the consequences of
conforming with, or realizing, the rules (or rule system). In opposition to
consequentialism is deontology, which makes rightness rest either wholly or
partly on nonconsequential factors. Criteria discussed earlier are examples
of justificational deontology, namely, the criteria appealing to truths of
logic--(Cl) and (CIh)-and evidence proportionalism-(C4). Evidence
proportionalism is a particularly good example. It says that justificational
rightness is a matter of proportioning strength of conviction to weight of
evidence, no matter what the consequences. This is analogous to the deon-
tological theory of punishment, which says that punishment should be pro-
portioned to the seriousness of the crime, no matter what the effects
may be.
No deontological criterion examined thus far has proved satisfactory.
This does not prove, of course, that none will. But I cannot construct other
good prospects, so I shall look elsewhere. I suspect that some form of conse-
quentialism is correct, and I will now explore varieties of epistemic conse-
quentialism to find the most promising. Of course I restrict my attention to
rule consequentialisms, for I am interested in rightness of rules (or rule sys-
tems). I ignore entirely the suggestion that the justificational status of each
belief is a function of that very beliefs consequences. I also assume that we
are only interested in the direct consequences of rule conformity, not in-
direct ~ o n s e ~ u e n c e s . ' ~
Consequentialisms can differ along two dimensions. First, they may differ
on the issue of which consequences have (positive or negative) ualue. Sec-
ond, they may differ on which meusure of valued and disvalued conse-
quences is relevant to rightness. One familiar measure is mxirnimtion of
net value; another possibility is some sort of satisficing measure. Let me
begin with the question of value. It should be emphasized that I am not
concerned with epistemic value in general, but only with justificationally
relevant value, for I am presently interested in rightness of J-rules.
98 ]ustification and Reliability
not think this is an appropriate determinant of value for the theory of justi-
fication. It may be appropriate for other epistemic terms of appraisal, but
not for justifiedness. Consider two J-rule systems R, and R2. Conformity
with R, would yield true beliefs with greater content than conformity with
R2, but in other verific respects (such as truth ratio) the rule systems are
equivalent. It is unclear to me that the processes licensed by R2 are justifi-
cationally inferior to the processes licensed by R,. Rz7sprocesses may be in-
ferior on other dimensions of appraisal, but not on the dimension of
justifiedness.
Having excluded its principal competitors, I conclude that the verific ap-
proach to justificational value-in the objective consequentialist style-is
the most promising. True belief is the value that J-rules should promote-
really promote-if they are to qualify as right. But so far, this is just a theory
of value and a choice of objective over subjective consequentialism. We still
do not have a determinate criterion of rightness.
Thus far I have narrowed the class of acceptable criteria, first to process
criteria, then to consequentialist process criteria, and finally to uerific con-
sequentialist process criteria. But this still does not pick out a unique crite-
rion. The task of this section is to narrow the class of acceptable criteria still
further. I will advocate a particular criterion as best, but also acknowledge
several related criteria as legitimate conceptions of justifiedness. Thus, in-
stead of plumping for a uniquely correct criterion, I will endorse a small
family of criteria, all in the 'truth-ratio' category.
Within the class of verific consequentialisms, there are two important
subclasses to notice: criteria invoking number of true beliefs, and criteria
invoking ratio of true beliefs. It is clear that J-rule rightness is not solely a
function of the number of truths that would be produced. The 'process' of
believing everything one can think of, including each proposition and its .
negation, might produce as large a total of true beliefs as any other process.
A criterion that required (just) maximization of the number of true beliefs
could be satisfied by a J-rule that permits such a process. But, obviously,
beliefs so formed are not justified.
This strongly suggests that what is critical to J-rule rightness is ratio of
true beliefs, not total number. The trouble with the foregoing process (as-
suming it is psychologically feasible, which is doubtful) is that its ratio of
true beliefs to total beliefs would only be .SO.These considerations prompt a
reliability approach to justifiedness. A justificationally permitted process must
be one that yields a high truth ratio, higher-perhaps appreciably higher
-than .50. Further support for reliabilism comes from a survey of cases.
Among the prototypical cases of justification-conferring processes are: (1)
104 Justification and Reliability
human beings. Human beings might possess no set of basic cognitive pro-
cesses that allow the criterion to be satisfied. Less dramatically, such a crite-
rion might exclude belief permission rules for one class of processes, for
example, inductive processes, though not exclude belief permission rules al-
together. Human perceptual processes might be capable of generating be-
liefs with a truth ratio of .90, but perhaps no human inductive processes
could generate such a high truth ratio.
Notice that a maximizing, resource-relative criterion could have implica-
tions that are either stronger or weaker than some specific resource-inde-
pendent criterion. Take an (absolute) resource-independent criterion that
sets the acceptable truth ratio at .80. If it turns out that human resources
include processes capable of a truth ratio higher than .80, then a maximiz-
ing, resource-relative criterion would require (at least) that higher ratio.
But if human resources are weaker, a maximizing resource-relative criterion
might only require a truth ratio of, say, .70. In principle such a criterion
might even count a rule system as right with a truth ratio less than .50.
Is a criterion that permits a truth ratio of less than .50 a 'reliabilist' crite-
rion? Probably not. This means that not all the criteria that fall into the
truth-ratio category qualify as reliabilist criteria. Perhaps only the absolute,
resource-independent criteria should be so labeled. In that case among the
family of criteria that I am prepared to count as legitimate-that is, as cap-
turing s o n e intuitively satisfying conception of justifiednes-not all are re-
liabilist criteria. They are, however, all truth-ratio criteria.
The satisficing, resource-relative approach has its attractions. In many
domains we make evaluations based on what the 'common man' does, or
what it is reasonable to expect people to do given their capacities. Plausi-
bly, this is what we (sometimes) do in appraising belief formation too. We
expect people to use better, more reliable, processes as long as they are rela-
tively easy, or natural, processes. We do not expect people to use very diffi-
cult, or sophisticated, cognitive process sequences, even if these could
increase the truth ratio. Omission of such processes does not preclude justi-
fied belief, on one plausible conception of justifiedness. In Part I1 I will have
several occasions to distinguish the relative difficulty of cognitive feats and
to link judgments of justifiedness to these distinctions. The satisficing, re-
source-relative criterion would neatly accommodate such judgments.
All three of the aforementioned approaches have their merits. In most of
what follows, though, I shall opt for an absolute, resource-independent cri-
terion of justifiedness. One virtue of this approach is that it makes the chal-
lenge of skepticism both serious and credible. If some fixed level of truth
ratio is necessary for justifiedness, it is an open question whether human
processes ever yield justified belief, whether they are even capable of yield-
ing justified beliefs. By contrast, the maximizing, resource-dependent crite-
rion guarantees that some human processes can yield justified beliefs. And
106 Justijication and Reliability
question. Is a J-rule system that is right in one possible world also right in all
possible worlds? In other words, is rightness a 'rigid' designator? Or can a
system vary in rightness across different possible worlds? Rightness is easily
relativized to a possible world by saying that, for any world W, a system is
right in W just in case it has a sufficiently high truth ratio in W. A different
theory would be chauvinistic toward the actual world. It might say that a
rule system is right in any world W just in case it is sufficiently reliable in
the actual world. This kind of actual-world chauvinism would of course
imply that each rule system has the same rightness status in all possible
worlds.
I think that neither of these answers is correct. At any rate, neither seems
to capture our intuitions about justifiedness in hard cases. The answer that
best accords with our intuitions-as will emerge in the next section-runs
as follows. We have a large set of common beliefs about the actual world:
general beliefs about the sorts of objects, events, and changes that occur in
it. We have beliefs about the kinds of things that, realistically, do and can
happen. Our beliefs on this score generate what I shall call the set of n o m l
worlds. These are worlds consistent with our general beliefs about the ac-
tual world. (I emphasize 'general', since I count worlds with different par-
ticular episodes and individuals as normal.) Our concept of justification is
constructed against the backdrop of such a set of normal worlds. My pro-
posal is that, according to our ordinary conception of justifiedness, a rule
system is right in any world W just in case it has a suffjcientlyhigh truth
ratio in n o m l worlds.24 Rightness is rigidified for all worlds; but it is
rigidified as a function of reliability in normal worlds, not reliability in
the actual world. Rightness of rules-and hence justifiedness-displays
normal-world chauvinism.
Obviously, the notion of normal worlds is quite vague. But I do not find
that objectionable. I think our ordinary notion of justifiedness is vague on
just this point. But let me expand on the approach by putting it in a more
general, nonepistemological, context.
The idea of actual-world chauvinism is related to the causal theory of
names and natural kind terms proposed by Saul Kripke and Hilary Put-
nam.25On this view the reference and/or meaning of a term is derived from
the actual sample of objects to which the term was applied in some original
baptismal acts, or meaning-giving contexts. Words like 'gold' or 'catY,for
example, acquire reference and/or meaning from the class of things to
which they were originally applied. Objects sharing the same nature or es-
sence as that sample, or class, are also gold pieces, or cats. Without review-
ing the many discussions of this theory, let me turn to a problem case raised
by Peter Unger, a case which also provides rationale for normal-world
~hauvinism.~
In an early paper Putnam introduced an example in which the things we
108 Justijication and Reliability
have always taken to be cats have really been inanimate robots, placed on
earth by Martian scientist^.^' Unger now introduces the case of feline robots
in the past and feline animals in the future. Starting now, the Martian scien-
tists (unbeknownst to us) replace all the feline robots with feline animals.
According to the causal theory, the feline animals would not correctly be
called 'cats'. The theory implies that 'cats' would refer to the nature or es-
sence of the things originally dubbed 'cats', namely, feline robots. Since the
feline animals do not share that nature or essence, they are not cats. But this
is intuitively incorrect. Surely the feline animals would be cats.
Furthermore, I find it intuitively wrong to say that the feline robots are
cats, even though they are what humans in the example have been calling
'cats' all along. This suggests that the reference and/or meaning of 'cats' is
fixed not by the actual nature of objects so called, but by their presumed
nature. What we believe about the things we call 'cats' determines the
meaning.
Similarly, I suggest that the meaning of the term 'justified' (in its episte-
mic use) is fixed by certain things we presume about the world, whether we
are right or not. Specifically, beliefs are deemed justified when (roughly)
they are caused by processes that are reliable in the world as it is presumed
to be. Justification-conferring processes are ones that would be reliable in
worlds like the presumptively actual world, that is, in normal worlds.
Let me trace an implication of this theory. Suppose it turned out that the
actual world is not as we think it is. Suppose it is a Cartesian demon-style
world, in which we are radically deceived about 'external' things (though
we are not deceived about the nature of our cognitive processes). This
means that our cognitive processes are not reliable in the actual world. At
least our perceptual processes are unreliable, and perhaps our inductive
processes as well. It does not follow, on my theory, that our actual-world
beliefs are uniustijied. The cognitive processes we use may still be reliable
in n o m l worlds, in worlds similar to the presumed actual world. This pos-
sibility follows from normal-world chauvinism vis-a-vis J-rule rightness.
I do not suppose that the fundamental world regularities that define the
class of normal worlds extend to properties of our own cognitive processes.
Thus, what we believe about our cognitive processes in the actual world
need not hold in (all) normal worlds. For this reason, my proposal does not
imply that the processes believed to be reliable (in the actual world) are re-
liable in normal worlds. Consequently, the processes we believe to be justi-
fication-conferring need not really be justification-conferring. We could be
mistaken in thinking that certain cognitive processes of ours are reliable in
normal worlds. For example, we may presume that our processes of esti-
mating the chances of future events are quite accurate, but we may be quite
mistaken about this. Hence, it is still a contingent and empirical question
whether the processes we take to be justification-conferring really are so;
Justification and Reliability 109
seriously impaired. Despite this belief, Millicent forms other beliefs that
endorse the testimony of her visual appearances. Intuitively, these beliefs
are not justified. But doesn't our reliability theory imply that they are justi-
fied? No. To appreciate this, we must simply recall the nonundermining
clause of our framework principle (P3*).The nonundermining provision is
violated in the Millicent case; at least that is how the provision should be
understood.
Recall the initial discussion of permission undermining in section 4.1. A
sufficient condition for undermining was suggested there: the cognizer be-
lieves--justifiably or not-that he or she is not permitted to hold the target
belief. Now Millicent does not fulfill this condition, we may suppose. Let us
say she does not possess the concept of permission by a right J-rule system,
and therefore possesses no beliefs featuring this concept. What is true of
Millicent, though, is the following. She believes that her visual powers are
impaired, so that the belief-forming processes that utilize her visual powers
must not be reliable. What she believes, then, is such that if it were true,
the beliefs in question (her visually formed beliefs) would not be permitted
by a right rule system. Satisfaction of this condition, I now propose, is suffi-
cient to undermine permittedness. In other words, it is sufficient for under-
mining that a cognizer believes that certain conditions obtain which, if they
did obtain, would entail that the target beliefs are not permitted by a right
rule system. The cognizer need not actually have any beliefs about rule sys-
tems, rightness, or criteria of rightness.
Millicent's visually formed beliefs are permitted by a right J-rule system.
But according to my framework principle, the beliefs still are not justified,
since their being so permitted is undermined. This yields the result needed
for handling the knowledge question: Millicent's visually formed beliefs are
not pieces of knowledge because they are not justified. And they are not
justified despite the fact that the producing processes are indeed reliable.
Let us now turn to BonJour's examples. A difference between BonJour's
cases and the Millicent case is that the characters in BonJour's examples do
not believe they are not using reliable processes. However, undermining
should be so construed that the permittedness of their target beliefs is un-
dermined nonetheless. For although they do not in fact believe that the tar-
get beliefs are unreliably caused, they are justified in so believing (in a sense
to be specified below).
Let us look at these cases in detail. Maud has a completely reliable clair-
voyance power, and uses it in forming a belief N that the President is in
New York. Maud believes herself to have the power of clairvoyance. But
Maud also has been inundated with massive quantities of apparently cogent
scientific evidence that no such power is possible. Intuitively, then, Maud is
justified in believing that she has no reliable clairvoyance power, although
she does not believe this. And she is justified in believing that her belief in N
112 Justification and Reliability
is not reliably caused, although again she does not believe this. Since she is
justified in believing the latter, however, that should suffice to undermine
the permission of her believing N. (Use of her clairvoyant powers would be
permitted by a right rule system because they are reliable.)
To make this proposed treatment stick, I need to go a bit beyond the the-
ory of justifiedness offered thus far. The account I have offered is a theory of
when an actual belief is justified. It is a theory of what I call (in "What Is
Justified Belief?") ex post justifiedness. But the present undermining deals
with being justified in believing a proposition one does not actually believe.
The idea is that it is a proposition one would (or could) be justified in be-
lieving, given one's present cognitive state, although one does not in fact
believe it. This is what I call ex ante justifiedness. A theory of ex ante justi-
fiedness can be constructed, I think, along lines very similar to my account
of ex post justifiedness, though I will not pursue all the details. One differ-
ence is that we may have to require right rule systems to feature obligation
rules as well as permission rules. Thus, we might suppose that a right rule
system would require Maud to utilize certain reasoning processes, processes
that would lead her from the scientific evidence she possesses to belief in
the proposition that she does not have any reliable clairvoyance power. And
similarly, she would be required to infer that her belief in N is not caused by
any reliable processes. On the proposed account she would be justified (ex
ante) in believing that her belief in N is not reliably caused. And this suffices
to undermine the permittedness of her belief in N.
BonJour's case of Casper would be handled similarly. Casper has at-
tempted to confirm his allegedly clairvoyant beliefs on numerous occasions,
but they have always turned out apparently to be false. Given this evidence,
he ought to use processes that would lead him to believe (permissibly) that
he lacks reliable clairvoyant processes. This undermines the actual permissi-
bility of his belief in N.
BonJour's next case, the case of Norman, is arguably a little tougher, and
indeed it is introduced to show that even a nonundermining condition still
does not produce an adequate reliabilist theory. BonJour describes this case
as one in which Norman possesses no evidence or reasons of any kind for or
against the general possibility of clairvoyance, or for or against the thesis
that he possesses it. But it is hard to envisage this description holding. Nor-
man ought to reason along the following lines: 'If I had a clairvoyant power,
I would surely find some evidence for this. I would find myself believing
things in otherwise inexplicable ways, and when these things were checked
by other reliable processes, they would usually check out positively. Since I
lack any such signs, I apparently do not possess reliable clairvoyant pro-
cesses.' Since Norman ought to reason this way, he is ex ante justified in be-
lieving that he does not possess reliable clairvoyant processes. This
undermines his belief in N. Thus, the nonundermining clause of (P3") han-
dles BonJour's cases.
Justification and Reliability 113
Let us turn now to the Cartesian demon example. This case is intended to
show that reliability is not necessary for justifiedness. Intuitively, a cognizer
being massively deceived by an evil demon is still forming justified beliefs.
But in an evil demon possible world, the cognitive processes being utilized
just are not reliable. So reliability is not a necessary condition for
justifiedness.
The solution to this counterexample is straightforward. The counterex-
ample rests on the assumption that the justificational status of a belief in
possible world W is determined by the reliability of the beliefs processes in
W. But this is what I have denied (in the previous section). The justifica-
tional status of a W-world belief does not depend on the reliability of the
causing processes in W. Rather, it depends on the reliability of the processes
in normal worlds. Now an evil demon world is a paradigm case of a non-
normal world. So it does not matter that the processes in question are highly
unreliable in that world. It only matters whether they are reliable in normal
worlds, and that is apparently the case. (At any rate, that is what our intui-
tive judgment of justifiedness presumes, I think.)
The same treatment of the putative counterexample applies if the coun-
terexample is revised to make it an actual-world case. Imagine, the objec-
tion goes, that our actual world turns out to be an evil demon world. (Or
imagine we are actually brains in a vat being deceived by scheming scien-
tists.) Intuitively, our beliefs would still be justified; yet the belief-forming
processes being deployed are not reliable. Again the case is easy to handle.
Its apparent strength rests on the assumption that the justificational status
of the beliefs is determined by the reliability of their causal processes in the
actual world. But this does not accord with our theory. Reliability is mea-
sured in normal worIds; and in this case, the actual world is an abnormal
world! So reliability in the actual world just does not matter.
The reader can now see how the normal-worlds account of reliability fits
our intuitions very naturally in the evil demon class of cases. Indeed, reflec-
tion on these cases (partly) prompted the normal-worlds approach. Once
formulated, though, it seems to work naturally for other cases as well.32
5.8. Truth-Linkedness
The rightness criteria I have proposed are all verific consequentialist cri-
teria, and these are a species of truth-linked criteria. Many people a.re suspi-
cious of such criteria. There are charges of circularity and worries about the
difficulty of applying a truth-linked criterion. This section is intended to an-
swer the charges and allay the worries.
Let us begin with a charge of definitional circularity. This charge (in
slightly different words) is leveled by Putnam:
Truth, in the only sense in which we have a vital and working notion of it, is
rational acceptability . . . But to substitute this characterization of truth into
the formula 'reason is a capacity for discovering truths' is to see the empti-
ness of that formula at once: 'reason is a capacity for discovering what is . . .
Justij5cation and Reliability 11 7
must already assume some system as right. Some system must be presup-
posed in order to form beliefs at all, including beliefs about which system
satisfies the criterion's requirements. This is apparently viewed as objec-
tionable on grounds of circularity.
Does the suggestion that a rule-system must be assumed, or presupposed,
mean that one must already have beliefs about what rule system is right?
That is not so. One need only use cognitive processes (and perhaps methods)
to decide whether a target rule system satisfies the criterion. Using cogni-
tive processes does not in general-nor in this case--require beliefs in the
epistemic properties of rule systems.
Firth is also wrong in suggesting that application of a reliabilist criterion
requires prediction of the future. One can identify a truth propensity in
normal worlds-at least an approximate truth propensity-without worry-
ing about the exact course of events in the future segment of the actual
world. Consider, for example, assessment of the truth propensity of visual
mechanisms. By studying their underlying structure, one can determine
that they would breed illusory impressions in a certain class of stimulus
conditions. If one has a rough idea of the frequency of this class of condi-
tions in normal situations, one can assess the degree of unreliability this
property engenders. Some inductive reasoning is involved in the study of
any mechanism's properties, but no prediction of individual future events is
required.
The next point to notice-and emphasize-is that truth-linked criteria
are no worse off than any other criteria in the matter of application. To
apply any sort of criterion, some reasoning and other cognitive processes
are required. How else could one decide whether a particular rule-system
satisfies a given criterion? This fact does not reflect adwersely on truth-
linked criteria as compared with other criteria.
Perhaps the intended challenge is not whether a truth-linked criterion
can be applied at all, but whether it can be applied either (a) correctly, or
(b)rationally (or justifiably). Firth's formulation concerns (b), but let us start
with (a).
First, again note that the issue does not reflect adversely on truth-linked
criteria as compared with other criteria. There can always be a question of
whether antecedent cognitive processes or methods will work correctly in
trying to apply a criterion. Consider a criterion that says that a rule system
is right if and only if conformity with it would maximize 'positive coher-
ence'. An attempt to apply such a criterion would first involve an assess-
ment of just how much coherence would be produced by each of a number
of specified rule systems (given some measure of positive coherence). Next
one would have to determine that all possible rule systems had been sur-
veyed, at least all strong candidates. Clearly, all sorts of mistakes might be
made in this selection process. Even much weaker criteria would feature
Justijkation and Reliability 119
our native cognitive processes and currently accepted methods are too de-
fective, they will not allow non-truth-linked criteria to be applied correctly
either. Indeed, if our processes and accepted methods are so flawed, almost
any worthy intellectual endeavor is doomed to failure.
Some philosophers would insist that such a dire scenario can be excluded
a priori. Proper principles of 'interpretation', or belief ascription, they
would say, guarantee that most of a person's beliefs are true. So belief-
forming processes cannot be too defective. I can live with this suggestion,
but I do not wish to insist upon it. In particular, I want to stress a strong
contingent component in process reliability (see Chapter 8). But a modest a
priori component could be allowed as well, if the argument for it could be
made good.
Although I am leery of the thesis that our belief-forming processes could
not be too defective, I acknowledge that nobody could believe that his or
her own processes were defective in a wholesale way; at least nobody with a
modicum of logical acuity and reflectiveness. How could anyone think that
all his or her belief-forming processes are radically defective? A little re-
flection would indicate that this very belief is thereby undermined, since it
too must have been formed by putatively defective processes. Still, while
wholesale self-deprecation is excluded, a person can certainly believe that
some of his or her belief-forming processes are faulty.
Let us now turn to a slightly different argument, which states that there is
no point in seeking a right rule system under a reliabilist criterion, because
there is no possibility of improvement. If we unfortunately start out with an
unreliable set of processes, any attempt to identify proper processes and
methods (right rules) will fail. So we would be stuck in an epistemic quag-
mire. If, however, we are fortunate to start with a reliable set of processes,
any attempt to identify proper processes will be pointless, since we will al-
ready be using such processes.
It is evident, though, that this argument is a flop. First of all, even if we
natively use (mostly) reliable processes, it is another matter to identify what
those processes are. Second, no convincing argument has been given against
the possibility of improvement, of epistemic bootstrapping. The main sort
of bootstrapping scenario runs as follows. We start with a set of available
processes with varying degrees of reliability. We use the more reliable pro-
cesses to identify good methods. We then use the more reliable processes,
together with some of the good methods, to identify the various processes
and their respective degrees of reliability. The superior specimens are so
identified, and their use is said to be justification-conferring. The inferior
specimens are so identified, and their use is said to be non-justification-con-
ferring. If the faulty processes are subject to direct or indirect control, we
try to avoid those processes. Or we try to devise methods to minimize their
impact. In this manner epistemic melioration is possible.
Justification and Reliability 121
bat and relieve ignorance, not simply error.' To be sure, an intelligent crea-
ture may be lazy, and may therefore remain mired in ignorance. But an in-
telligent creature is at least capable of relieving its ignorance.
It would be a mistake to equate intelligence with power, at least if power
is equated with number of truths that can be obtained. First, it is not en-
tirely clear that number of truths, rather than, say, complexity, is the appro-
priate standard of intelligence. Suppose system A is capable of getting more
true beliefs than B, but B can solve more complex problems. B might be
credited with more intelligence. This could be accommodated by building a
complexity parameter into a measure of power. But I have no proposals for
a metric of complexity.
Second, a system may be capable of acquiring numerous truths, but
mostly incidental ones, unrelated to its needs or interests. For example, a
system may be capable of indefinitely expanding its true-belief set by add-
ing disjuncts to truths it already believes. But a system that just cranks out
more and more disjunctive true beliefs, oblivious to other pressing questions
left unanswered, might well be considered very stupid.
The last point is really twofold. First, intelligence probably has some-
thing to do with breadth of answerable questions, not simply their number.
Second, a system oblivious to the questions it wants to answer, which
blindly computes answers to irrelevant questions, is not very intelligent. In-
telligence is partly a matter of goal-responsiueness.
If we seek to crystallize a standard of (intellectual) power that is reason-
ably close to intelligence, we need to distinguish several conceptions of
power. One simply makes power a function of the number of true beliefs a
cognitive system can generate. A second makes it a function of number plus
breadth. A third makes power a function of the proportion of questions it
wants to answer that it can answer (correctly), or the proportion of prob-
lems undertaken that it can solve correctly. Only the third conception
makes power a goal-responsive notion, so it is the best conception in this
group. Notice, though, that we should restrict attention not to questions or
problems actually posed by the cognitive system, but rather to those it
could pose.
Further comments on goal-responsiveness are in order. Biological cogni-
tive systems are typically deployed to help the organism satisfy its practical
needs or ends. This involves the choice of actions, designed to achieve spe-
cified goals. Good choices of action-ones that lead to a high level of goal
satisfaction-typically depend on good information, that is, true beliefs or
relatively high confidence in truths. Thus, cognitive mechanisms are used to
try to get true beliefs on pretargeted questions, questions relevant to the
practical tasks at hand. Even when no practical task is at stake, cognitive
activity involves prespecified questions. Purely intellectual questions com-
monly engender a host of subsidiary problems, solutions to which become
124 Problem Soluing, Power, and Speed
sure, I would cheerfully endorse it. But I know of no such formula. We must
live for now with an irreducible multiplicity of standards. This may be a bit
disappointing, but not surprising. Consolation may be taken from ethics and
the law, where analogous complications arise. It is hard to balance the
competing claims of justice and mercy, of public good and individual rights.
We should not expect an easy formula for epistemic evaluation either. Dis-
tinct standards need to be treated separately. That is why I qualified all talk
of rules in Chapters 4 and 5 as iustijicational rules: they were not rules for
overall epistemic status. Still, all appropriate standards need to be kept in
mind. Thus, my assessments of cognitive mechanisms in Part I1 will heed
power and speed as much as reliability.
But then their question is: What are some candidate answers to that ques-
tion? and there again a correct answer is desired, namely, things that are
candidate answers.
Given this definition of having a problem, the natural definition of a
problem solution is:
Proposition A is a solution (answer) to problem (question)Q if and only if
(1) A is a potential answer to Q, and
(2) A is true.
Erotetic logic studies the logical relations between questions and answers.
This includes the study of which propositions are potential (relevant) an-
swers to which questions. By a potential answer, I mean roughly what Nuel
Belnap and Thomas Steel call a "direct" answer: a proposition that answers
the question completely, but just completely? 1 will not survey or endorse
any theories of erotetic logic. I will just presuppose our ability to recognize
potential answers.'
Given the controversy over the truth-condition, it is worth underlining its
appropriateness further. Its suitability is especially clear for the terms
'problem' and 'solution'. In common usage one does not solve a problem
unless one gets a correct solution: a real solution is a correct solution. No-
body is credited with solving a mathematical problem unless he or she finds
(is credited with finding) a true solution. Nobody is credited with solving a
game problem or puzzle, such as Rubik's cube or cannibals-and-mission-
aries, without getting a correct solution. A detective does not solve a mys-
tery without figuring out who really committed the crime. In crediting
someone with a solution, we mean that he or she has a true solution-al-
though we can, of course, be wrong about this.
Is a solution to a problem necessarily a proposition, as our definition im-
plies? In the Rubik's cube puzzle, or other such puzzles, a solution might
appear to be a specified end state, such as a configuration of the cube; and
this is not a proposition. However, the propositional formulation can read-
ily handle this and similar cases. A solution may be construed as a proposi- '
can mistakenly believe that there are. Philosophers used to study a special
class of such questions, 'pseudoquestions'. Pseudoquestions were said to be
meaningless, or irresoluble, philosophical questions that have a semblance
of meaningfulness and answerability. This status was invested with some
philosophical importance. Of course, questions can also lack true answers
for philosophically uninteresting reasons. What is a proof of this formula?
has no true answer if the formula is not a theorem. Where is a good restau-
rant in town? has no true answer if the town has no good restaurants. ('No-
where' is not a true answer so much as a denial of the question's
presupposition. Typically, questions have presuppositions, the falsity of
which can imply the absence of any true answer.)"
Some questions delimit, either explicitly or implicitly, a fixed set of po-
tential answers. Yes-or-no questions are a special case, where the two can-
didate answers are 'yes' and 'no' (or corresponding full sentences). Other
questions enumerate a list of possible answers, for example, Who is the old-
est: Sally, Mary, or Beverly? Many questions, however, delimit no finite or
fixed set of possible answers. These include, What explains the moon illu-
sion? and What is a cure for cancer? For epistemological purposes, this is
the most interesting class. To find an answer to such questions, a cognizer
must construct possible answers. He cannot simply assess truth-values of the
specified candidates. This cognitive task is clearly more complex. Admit-
tedly, truth-value assessment is often difficult enough, but not always. It
may be easy to determine whether a proposed proof of a putative mathe-
matical theorem really is a proof. So it is easy to answer the question, Is this
putative proof a genuine proof? But constructing or inventing a genuine
proof may be extremely difficult, requiring lots of creativity. So it is much
harder to answer, What is a proof of this statement?
Science characteristically asks questions for which it is difficult both to
devise plausible answers and to verify their truth-value. Because of verifi-
cational difficulties, there are often deep methodological disputes over
principles of verification. Does this mean that tnlth should be dispensed
with as a scientific desideratum? Surely not. Just because something is hard
to get, even hard to recognize if and when you've got it, does not imply that
it should not be sought. There is ample reason, moreover, for understanding
scientific inquiry as a quest for truth, like other inquiries. One significant
scientific activity is the devising of new kinds of equipment, technologies,
and methodologies that can facilitate theory testing. The natural explana-
tion of such activities is that better devices, better statistical methodologies,
and so on, will enable researchers to identify, or get closer to, the truths
they seek.
I will say more about truth seeking in science in Chapter 7. But let me
add a few remarks here about some authors' explicit rejection of truth in the
conception of scientific problem solving. Laudan, in particular, offers a
130 Problem Soloing, Power, and Speed
tells the cognizer always to seek and choose a move that goes up the present
hill, brings the cognizer closer to the goal state. The notion of closeness is
frequently fairly natural. In the cannibals-and-missionariesproblem, for ex-
ample, a move that results in two cannibals and two missionaries being on
the other side of the river gets one closer to the goal than the initial state.
(The initial state is no cannibals or missionaries on the other side; the goal
state is all three cannibals and all three missionaries on the other side.) Hill
climbing has its defects as a general strategy. For example, it can get one
stuck on a local hill (a local maximum), and give no method for ultimately
climbing the mountain (getting to the real goal state). But it illustrates the
idea of a strategy that results in a series of search steps2'
A cognitive system must operate sequentially, by a series of steps. (I do
not mean to exclude parallel processing.) In trying to solve a problem, it al-
ways faces the implicit question: What should be done next? Furthermore,
it must perform its operations in real time, not in some idealized situation
where time is suspended. Since cognitive tasks often have time con-
straints-especially when linked to practical exigencies-the exact tem-
poral sequence of operations is critical. The epistemic adequacy of
cognitive 'heuristics', or short cuts, must be assessed in light of these
constraints.=
Traditional epistemology and philosophy of science have largely ne-
glected the temporal and operational constraints on the intellect. Discus-
sions of justification and rationality have typically proceeded in terms of
logical and probabilistic relations between theory and evidence; but these
sorts of relations abstract from temporal and operational considerations.
The so-called historical wing of philosophy of science, to be sure, empha-
sizes diachronic factors. But this tradition has not attended to cognitive op-
erations, in the sense intended here. It seems clear, though, that an
adequate primary epistemology must be centrally concerned with the real-
time situation of a cognitive system and the sequences of operations it exe-
cutes. Cognitive science, of course, studies cognition in precisely this .
fashion.
The importance of sequential operations does not just result from our
concern with the epistemic standard of speed. To reason effectively even
without deadlines, a cognitive system must ensure that relevant premises
are accessed at the time they are needed. Suppose that I periodically think
of the fact that the university library opens on Sundays at noon. That will
not be of much help if I neglect to access this fact on Sunday mornings,
when I actually contemplate going to the library. If I fail to retrieve this
fact on Sunday mornings, I may make some futile trips. Thus, the timeliness
of mental operations is an important facet of cognitive life, which must not
be neglected in a comprehensive primary epistemology. But it is readily ne-
glected without detailed study of cognitive operations.
I have focused predominantly on stages of inquiry prior to the doxastic
136 Problem Soluing, Power, and Speed
decision stage, primarily because past epistemology has riveted its attention
on the latter stage. Past emphasis on the doxastic decision stage may partly
be rooted in the dominant interest in justification, which seems mainly con-
cemed with this stage. But if intelligence, or power, is of equal importance,
there is no good rationale for this concentration. Arguably, a cognizer can
have justified doxastic attitudes just by using suitable doxastic decision pro-
c e ~ s e sBut
. ~ ~one cannot solve problems effectively just by using good dox-
astic processes. Doxastic decision processes only serve to select from the
available candidate solutions. At best, selection can reject false candidates
and accept a true candidate if one is generated. But if no true solution is
generated, even the best doxastic decision process cannot produce belief in
a true solution. To be a good problem solver, then, one needs good hypoth-
esis-constructing processes. Similar remarks apply to processes governing
sequential search.
generate systems of status, authority, and (social) power. When those ac-
corded a high status get things wrong, their false doctrines may gain wide
currency and may be difficult to eradicate, even in the face of striking
counterevidence. The influence of authority figures may congeal ideational
frameworks and inhibit novelty and invention.
Setting aside these issues, which are too large for present scrutiny, notice
that there are questions to be raised about what it even means for a group to
solve a prob1e.m. Does it mean that all members of the group believe a true
answer? Presumably that is much too strong. Exactly which members, then,
or how many, need to believe a true answer? Once this issue is raised, it be-
comes obvious that a finer-grained treatment of the topic is in order. Take
the case of governments. Certain experts in the State Department, or the
Defense Department, may know the truth about a certain question, but
cannot persuade their higher-ups of this truth. Perhaps they cannot even
have timely communication with decision makers to get the truth heard. If
the decision makers lack (belief in) the truth, it is arguable that the govern-
ment does not have the solution, even though some of its members do.
In any case, the distribution of information (or belief) within an organiza-
tion is certainly of paramount importance. This topic should be placed on
the agenda of social epistemology. But the point is also of interest because
of analogies with (primary) individual epistemology. As noted in section 6.3,
the mind also has functional locations, such as working memory and long-
term memory. Just as it matters to an organization where information is lo-
cated within it, so it matters to the mind where information is located. If
information in an organization resides only in some lower-echelon official,
and never gets persuasively communicated in a timely fashion to higher of-
ficials, its value to the organization may be nil. Similarly, information only
stored in LTM,and never accessed at all or in a timely fashion, has little
value to a cognitive system, either for purposes of behavioral decision or
further problem solving.
Evidently, a critical facet of social problem solving is the process of com- .
munication. This brings into prominence the units of communication by
which individuals hope to shape the thinking of others. These units include
oral discourses, books, articles, research reports, theorems (or proofs), legal
opinions and briefs, and various forms of fiction. Let us call all such works
intellectual products.
I have said that epistemology's aim is intellectual evaluation. Thus far,
our attention has been engaged by the evaluation of mental events and pro-
cesses, such as, beliefs and belief-forming processes. In fact, intellectual
evaluation far more frequently centers on what I am calling products than
on internal states and operations. In everyday intellectual and cultural
commerce, it is not mental acts that comprise the chief objects of appraisal;
it is discourses and published works. It would be a lean epistemology indeed
that did not feature such objects in its purview. Social epistemology, at a
138 Problem Solving, Power, and Speed
minimum, must be concerned with these objects, since they are the life-
blood of cooperative intellectual endeavor. So let us momentarily shift our
focus from belief epistemology to product epistemology.
However, I doubt that product epistemology can be divorced entirely
from belief. A principal aim of most intellectual products is to inform an au-
dience, to induce beliefs in putative truths. Several dimensions of evalua-
tion flow naturally from this aim. First, a premium is placed on products
that present new truths, truths not already circulating in the public forum.
If some true contents have previously been aired, people have had the op-
portunity to learn and believe those contents. So their significance is re-
duced. But this needs some qualification. Even if a given truth has
previously been presented, a product that presents it more cogently, per-
suasively, or comprehensibly is still highly valued. Thus, second, the mode
or manner of presentation counts substantially in a product's appraisal.
Third, value is placed not simply on truth per se, but on truths that interest
the audience. Where individuals or groups place high priority on answering
certain questions, products that succeed in answering those questions-or
are judged to answer them successfully-receive special plaudits.
Still another determinant of product evaluation may also be traced to
mental factors. The grading of a product typically involves the question of
originality, or creativity. By how much does this product go beyond previ-
ous work in the field? How difficult was it to produce this work, given the
common knowledge and tools of the discipline? A certain mathematical
lemma, for example, may never have been stated or proved before, but it
might still be an obvious consequence of a well-known theorem. The mag-
nitude of an intellectual achievement is partly a function of the difficulty of
generating it. Clearly, such judgments of difficulty presuppose some metric
of originality or creativity, and the question arises of the basis of such a met-
ric. It can only be based, I believe, on some conception of normal, human
cognitive capacities, a matter on which cognitive psychology may be able
to shed some light. I shall return to this theme in Chapter 11. Meanwhile, I
content myself with the remark that even the evaluation of products in the
social arena may be illuminated by reference to human psychology.
Other specialized aims and vehicles account for specialized values at-
tached to diverse traditions and genres. Although language and symbolism
are the medium of the sciences, humanities, law, and the arts, the aims and
uses of language and symbolism differ widely, and the values emphasized
differ accordingly. In law and philosophy, disputation and argumentation
play a central role. The skills of the debater and the controversialist are
emphasized. In other fields, such as the fictive arts, different kinds of rhe-
torical styles are emphasized; disputation plays little role. While the law-
yer, the philosopher, and the playwright may all aim at some sort of truth,
the formats and techniques for expressing their respective ideas are very di-
vergent. Similarly, canons of science and mathematics require precision and
rigor, whereas the literary fields call for expressions that are evocative and
rich in connotation, possibly precluding precision and rigor.
It may indeed be suggested that different disciplines and genres have
fundamentally different modes of conveying truths. Science, scholarship,
and philosophy concentrate on the assertive mode of expression, whereas
literary works may try to convey truths by showing, or depicting, them. The
novelist shows what it is like to be a certain kind of person, to live in a cer-
tain era or culture. This is a commonly valued way of getting understanding
of human and cultural facts; important information is somehow conveyed.
Epistemology must be concerned with such diverse modes of communica-
tion, and it must countenance multiple values specifically keyed to these
diverse modes. Even among purely assertive communications, differences in
style are crucial to the mental or doxastic upshots in an audience. Style af-
fects perspicuity, which affects how well a message is understood, compre-
hended, or represented by the audience. This too is reflected in critical
evaluations of intellectual products.
Epistemology has focused heavily on statemental units of communication
and on propositional units of mentation. But this may be an inadequate ap-
proach if epistemology seeks to deal with understanding or comprehension.
How does a cognizer comprehend a large body of data or a highly complex
theory? Can such comprehension be well represented by a list of believed
propositions? Does incomprehension simply consist in failure to grasp the
relevant propositions (whatever 'grasping' is)? Recent work in psychology
and artificial intelligence posits different sorts of mental structures-sche-
mata, frame-systems, scripts-that encode complex informational struc-
tures. An adequate belief epistemology may want to make use of such
mentalistic constructs. At the same time, a product epistemology may want
to appraise, or elucidate appraisals of, discourses in terms of how well they
succeed in effecting suitable sorts of mental structures. Thus, true 'belief
may be too restrictive a conception of epistemic aims, even when other ad-
junct and speciaIized values are set aside.
chapter 7
volve is that statements in the given class relate to some reality that exists
independently of our knowledge of it, in such a way that that reality renders
each statement in the class determinately true or false, again independently
of whether we know, or are even able to discover, its truth-value."'
There are two components in this characterization of realism. First, there
is the principle of biualence: every statement in the class is determinately
either true or false. Second, there is the principle of uerijication-transcen-
dent truth: a statement is true or false independently of our knowledge, or
verification, of it (or even our ability to verify it). The second of these theses
will figure prominently in my discussion. But the first, the bivalence princi-
ple, is tangential.
It is not clear just how firm a criterion of realism bivalence should be.
Dummett regards it as necessary for realism, though not ~ufficient.~ But it is
not even clear that it should be accepted as necessary. Why can't a philoso-
pher be a realist about a given class even if he takes the Frege-Strawson line
that a singular statement containing an empty proper name-such as, 'The
present king of France is bald'-is neither true nor false? Why can't he be a
realist even if he acknowledges vague statements within the given class,
which are neither true nor false?4
Setting these points aside, suppose we grant Dummett the label 'antireal-
ist' for philosophers who deny the principle of bivalence for a given class of
statements. Is it important for my purposes to defend the corresponding
kind of realism, as a general claim? Not at all. There is nothing in my episte-
mology that requires defense of the principle of bivalence for all classes of
statements. I could readily admit the failure of this principle for any num-
ber of classes of statements: statements about the future, subjunctive condi-
tional statements, even statements of mathematics. What is critical is that
when any such statement i s true (or false), what makes it true (or false) is
independent of our knowledge or verification. Thus, the second of the two
realist theses explained above, verification-transcendence, is critical. To put
it another way, truth must not be an epistemic matter. But the issue of bi- .
valence does not affect my epistemology.
Admittedly, the two theses are interconnected in some philosophies.
Some denials of bivalence are consequences of a denial of verification-tran-
scendence. The constructivist in mathematics denies verification-transcen-
dence, saying that a mathematical statement is true only if there exists a
proof of it, and is false only if there exists a proof of its negation. The denial
of bivalence readily follows. However, people may also deny bivalence for
other reasons. Nothing in my epistemology requires any quarrel with biva-
Ience per se.
My concern with realism, then, is a concern with truth; with what makes
a statement, or belief, true, if it is true. So I shall primarily be concerned-
in much of the chapter, at any rate-with the theory of truth.
144 Truth and Realism
ject propositions. Clearly, (a) is much too strong. It could certainly be true
that Julius Caesar had a mole on the nape of his neck although it is false that
all of us now have conclusive evidence for this. Similarly for (b):this could
be true even though nobody now has conclusive evidence for p. Even (c) is
still too strong. This minor historical fact about Julius Caesar may be one for
which nobody will ever have conclusive evidence. Even (d) is too stringent.
Perhaps nobody now has even prima facie evidence for this fact'. This would
not keep the proposition from being true.
Not only is (d) too strong, it is also too weak. Suppose some historian does
have prima facie evidence for this proposition. That surely does not entail
that the proposition is true. For example, suppose a Roman text of the pe-
riod attributes a remark to Brutus that seems to indicate the presence of a
mole on Caesar's neck. That evidence is not sufficient for the truth of the
proposition. The same twin problems arise for (e), for similar reasons. It is
both too strong and too weak.
Turning to the 'possibility' formulations, (f3 is too weak for the same rea-
sons as (d) and (e). The mere possibility of prima facie evidence for p does
not guarantee p's truth. The difficulties for (g) are slightly harder to assess.
They depend on how 'conclusive evidence' is interpreted. If it is interpreted
very strictly-so that conclusive evidence entails truth-then it is doubtful
whether it is possible for somebody to have conclusive evidence for any
proposition, at least any physical object proposition. So (g) would be too
strong. If 'conclusive evidence' is interpreted less strictly, so that it does not
entail truth, then (g) is too weak.
There is another plausible objection against (g), if it is interpreted strictly.
On the strict interpretation we are to understand 'p is conclusively verified
(for some person S)' to entail (by virtue of its meaning) 'p is true'. But if this
is right, truth is covertly contained in the analysis of 'p is conclusively veri-
fied'. And then it would be circular to employ the latter locution in the
analysis of 'p is true'!
This point is worth underlining because it should also figure in the assess-
ment of the epistemic account of truth endorsed by Putnam. Putnam
writes:
Truth is an idealization of rational acceptability. We speak as if there were
such things as epistemically ideal conditions, and we call a statement 'true' if
it would be justified under such conditions."
of the hypothesis. But in that case who would suppose that this technical
notion has anything to do with the intuitive notion of evidential support?
Second, take the notion of epistemic 'access' to a state of affairs, such as
perceptual access or memory access. Again we have the notion of better
and worse epistemic situations. And again this can only be spelled out in
terms of truth. Perceptual access to a situation is better just in case it is eas-
ier, in that situation, to form true beliefs about the objects or events in the
scene. The optimal distance and viewing conditions are those that promote
the greatest chance of truth and accuracy.
Third, the history of science and philosophy is filled with controversies
over methodology. Contending schools have vied over the right procedures
for fixing belief. It is hard to make sense of these debates save as debates
over which methodology is most conducive to true convictions. However
difficult it may be to settle such questions, they seem to presuppose the no-
tion of truth. So truth is conceptually prior to any operational specification
of epistemic methods or standards. Antirealists, who seek to define truth in
epistemic terms, are proceeding precisely backwards.
We must not ignore, however, an underlying impetus behind many
moves toward an epistemic account of truth. Antirealists are doubtless
worried about the epistemological consequences of a realist conception of
truth. If truth is definitionally detached from evidence, either actual or
possible, is there really any prospect that we can have epistemic access to
the truth? Isn't there a threat of skepticism? Some antirealists might be
prepared to concede that the ordinary, naive conception of truth is none-
pistemic. But they may feel that replacement of this naive conception is re-
quired to avert epistemic catastrophe.
Given this problem, we must determine whether candidate epistemic ac-
counts of truth really help, whether antirealist proposals really improve our
epistemological opportunities. The answer is: not necessarily. Putnam's
proposal, (h), is a case in point. Truth is there equated with what people
would be justified in believing if they were in epistemically ideal conditions.
But since none of us ever is in epistemically ideal conditions with respect to
any proposition, how are we supposed to tell-from our actual stand-
point-what a person would be justified in believing in that hypothetical
situation? To assess the truth of p according to this proposed equivalence,
we would have to assess the truth-value of the indicated subjunctive condi-
tional. But that is no mean task! If anything, it imposes an even nastier
epistemic predicament than the one feared from realism.
The nonepistemic nature of truth is well described by Brian ~ 0 a r .Our l~
ordinary (realist) conception of truth can best be understood, he says, in
modal terms. The idea of a proposition's being true is the idea of a state of
affairs such that it could happen (or could have happened) that it be true
even though we are not in a position to verify it. Furthermore, he asserts, it
Truth and Realism 149
is part of our theory of nature that this is generally so. Whether p concerns
objects in the garden, in Antarctica, or in the Andromeda galaxy, p's veri-
fiability by us depends upon various contingent circumstances not entailed
by laws of nature. For us to verify such propositions, light, sound, or some
such energy must reach our sense organs. But the regions of space through
which such energy must travel might have (or might have had) distorting
properties. There might be black holes in the relevant regions that 'soak up'
the light. Also, for us to verify such propositions (indeed, any propositions),
our brains must be constructed in certain ways. But they might not have
been suitably constructed. So it might have happened that these proposi-
tions are true although we are not in a position to verify them.
As this way of explaining truth indicates, a sharp distinction must be
drawn being a proposition's being true and the proposition's being verified.
The latter, but not the former, involves processes by which the truth is de-
tected or apprehended. (Here I have in mind both extracognitive pro-
cesses-for example, the traveling of light from object to retina-as well as
intracognitive processes.) Indeed, only such a sharp distinction-character-
istic of realism--can make good sense of certain of our verifying
procedures.
In seeing how realism makes sense of verification, I follow the suggestions
of William Wimsatt, who in turn credits Donald Campbell and R. ~evins.'"
Our conception of reality is the conception of something robust, an object
or property that is invariant under multiple modes of detection. The use of
multiple procedures, methods, or assumptions to get at the same putative
object, trait, or regularity is commonplace in ordinary cognition and in sci-
ence. Wimsatt lists many examples, including the following: (a) we use dif-
ferent perceptual modalities to observe the same object, or the same sense
to observe the object from different perspectives and under different obser- .
vational circumstances; (b) we use different experimental procedures to (try
to) verify the same alleged empirical relationships, that is, alternative pro-
cedures are used to 'validate' one another; (c) different assumptions, models,
or axiomatizations are used to derive the same result; (d) we use agreement
of different tests, scales, or indices as a sign of validity of a trait-construct;
and (e) we seek matches of theoretical description of a given system at dif-
ferent levels of organization. In short, our verification procedures, when
they are careful, seek to 'triangulate' on the objects or relationships under
study.
We can best make sense of a need for triangulation on the assumption
that the truths, or facts, about the object or system under study are sharply
distinguished from the processes of verification or detection of them. Any
particular method might yield putative information partly because of its
own peculiarities-its biases or distortions. This is a potential danger as
long as the properties of the studied object are distinct from the process of
150 Truth and Realism
question, for any specified type of apparel, whether a specific token of that
type fits a particular customer's body. This question of fittingness is not just
a question of the style of garment. It depends specifically on that customer's
body. Similarly, although the forms of mental and linguistic representation
are human products, not products of the world per se, whether any given
sentence, thought sign, or proposition is true depends on something extra-
human, namely, the actual world itself. This is the point on which realists
properly insist.
In inventing or evolving sartorial styles, people devise standards for
proper fittingness. These may vary from garment to garment and from fash-
ion to fashion. Styles specify which portions of selected bodily parts should
be covered or uncovered, and whether the clothing should hug snugly or
hang loosely. This is all a matter of style, or convention, which determines
the conditions of fittingness for a given type of garment. Once such fitting-
ness-conditions are specified, however, there remains a question of whether
a given garment token of that type satisfies these conditions with respect to
a particular wearer's body. Whether it fits or not does not depend solely on
the fittingness-conditions; it depends on the contours of the prospective
wearer as well.
The case of truth is quite analogous. Which things a cognizer-speaker
chooses to think or say about the world is not determined by the world it-
self. That is a matter of human noetic activity, lexical resources in the
speaker's language, and the like. A sentence or thought sign, in order to
have any truth-value, must have an associated set of conditions of truth. Ex-
actly what determines truth-conditions for a sentence or thought sign is a
complex and controversial matter. But let us assume that a given utterance
or thought, supplemented perhaps with certain contextual factors, deter-
mines a set of truth-conditions. The question then arises whether these con-
ditions are satisfied or not. The satisfaction or nonsatisfaction of these
conditions depends upon the world. Truth and falsity, then, consists in the
world's 'answering' or 'not answering' to whatever truth-conditions are in
'
question. This kind of answering is what I think of as 'correspondence,'. No-
tice that which truth-conditions must be satisfied is not determined by the
world. Conditions of truth are laid down not by the world, but only by
thinkers or speakers. This is the sense in which the world is not precate-
gorized, and in which truth does not consist in mirroring of a precate-
gorized world. Unlike correspondencel, correspondencez is compatible
with what is correct in the constructivist themes of Kant, Goodman, and
Putnam.
While these philosophers are right to emphasize constructivism, they
carry it too far. There is a strong suggestion in Goodman's writing that
"versions" are all there is; there is no world in itself. Although Goodman
speaks of truth as being a matter of "fit," it is not fitness of a version to the
154 Truth and Realism
chased set of tires, the induction would be weak. (This response may not be
conclusive, but it should give us pause.)
Setting aside the force of the meta-induction, we need to look more care-
fully at the implications of my theory of justification for the present ques-
t i ~ nThere
. ~ ~ are several reasons why a poor track record of positive belief
in scientific theory need not ruin scientists' chances for justified theoretical
belief according to my account. By 'positive' theoretical belief let us mean a
belief that a theory is true; by 'negative' theoretical belief let us mean a be-
lief that a theory is false, that is, a belief in the denial of a theory. Since the
same methods are used to arrive at negative theoretical beliefs as positive
theoretical beliefs, even if scientific methodology (assuming, inaccurately,
that a single methodology has continuously been in use) has consistently
generated false positive theoretical beliefs, it may still have generated lots
of true negative theoretical beliefs. Moreover, the latter class is probably
much larger than the former. Think of all the scientific theories we now re-
ject and of the innumerable theories that past scientists rejected. When the
entire class of outputs of scientific methodology is included, both negative
and positive, the overall truth ratio may be rather impressive.
The next point concerns the distinction between primary justifiedness
(P-justifiedness) and secondary justifiedness (S-justifiedness). Recall that
P-justifiedness turns on the truth ratio of basic psychological processes; and
S-justifiedness turns on properties of methods. Now the basic psychological
processes used in assessing scientific theories probably are no different from
those used for many other tasks; only the methods, at most, are distinctive.
So even if these processes have a poor track record in this very special do-
main, it does not follow that they have poor reliability in general.
When we turn to S-justifiedness the matter is up in the air, because I did
not fully commit myself to a criterion of J-rightness at the secondary level.
While I am certainly attracted to an objectivist, truth-ratio criterion here
too, I feel a tug in the direction of a culturally relative criterion (see section
5.3). Perhaps a belief is S-justified if it results from methods that are ap- ,
I draw this discussion to a close, having shown why realism about truth is
a pIausible and defensible doctrine. Truth is not an epistemic notion; and in
a properly softened form, even a correspondence theory of truth is tenable.
We have seen how core scientific realism, while distinct from realism about
truth, is an attractive view, and ultimately compatible with my general
epistemology.
chapter 8
temporary science. Any generalizations made with their help are very
rough indeed. Nonetheless, there is little cause to deny the referential status
of these predicates. There are trees, automobiles, living things, wealthy
people, and holders of Ph.D.'s. Similarly, even if words like 'belief and 'de-
sire' have modest predictive utility---compared with projected predicates
of a maturing cognitive science-this should not be taken to establish their
referential bankruptcy. The thesis that the genuinely referring terms are
only those that occur in generalizations with maximal predictive power
(within a given domain) is very dubious.
Another challenge to mentalistic content is presented by Stephen Stich,
who sometimes seems to make a more cautious claim, namely, that the be-
lief construct is unsuitable for cognitive science. At other times, however, he
seems to call the very existence of beliefs into question. Stich claims that
cognitive science must pay a very heavy price if it adheres to the program
of a "strong representational theory of mind (RTM)," that is, a program that
couches generalizations in terms of content. Predicates of the form 'believes
that p' are both vague and context sensitive. The folk language of belief
characterizes a person's cognitive state by comparing it to our own. So peo-
ple who differ fairly radically from us in doctrine, or ideology, will fall out-
side the reach of belief description altogether. But some of these subjects
may have minds that work very much as ours do. Thus, Stich argues, if there
are important generalizations that cover both us and such exotic folk, a
cognitive science embodying strong RTM is bound to miss them."
One of Stich's examples is that of aged Mrs. T, who suffers from severe
memory loss. Whenever she is asked 'What happened to McKinley?' she
says, 'McKinley was assassinated', but she is quite unable to say what an as-
sassination is, or what its consequences are. It seems inappropriate, says
Stich, to describe Mrs. T as believing that McKinley was assassinated. In-
deed, no definite content ascription seems available to us. Thus, Mrs. T's
states are beyond the reach of a cognitive theory that cleaves to descrip-
tions in terms of content.
Is Stich's conclusion warranted? It is doubtless true that we often have
inadequate resources for describing the content of a person's belief. In par-
ticular, simple sentences of English (and other natural languages) cannot be
counted on to get things right. Our language marks certain distinctions but
not others. This means that it is difficult to phrase a set of distinctions drawn
by a given cognizer but not well marked in English. This holds, for example,
of Daniel Dennett's case of a six-year-old child who says that his daddy is a
doctor.12 The child does not yet have a full understanding of 'doctor'; he
may not know that this precludes being a quack or an unlicensed practi-
tioner. So perhaps the English word 'doctor' does not exactly capture the
content of the child's belief. Still, this should not imply that the child has no
belief at all.
168 The Problem of Content
There are additional reasons for doubting that belief content can be cap-
tured with simple sentences of a natural language, even one the believer
speaks. A person's perceptual representation of an object should probably
be included in the content of a belief concerning that object. But there may
be no satisfactory way of identifying this perceptual content with any sim-
ple sentence of a natural language. Of course, since we often want to com-
municate about the content of perceptual beliefs, we need some vehicle of
communication, especially a handy, efficient one. So we employ simple
sentences of a natural language. It is a mistake to suppose, however, that
such sentences can provide anything but a very rough approximution of be-
lief contents. Even in expressing one's own beliefs, it is doubtful that one's
utterances convey them very fully or accurately. None of this, however,
constitutes negative evidence against the existence of contentful beliefs.
Let us return now to Stich. Stich's worries are not exhausted by the diffi-
culty of pinpointing belief contents in puzzle cases. He presents two other
problems for the viability of the belief construct in cognitive science. First,
he claims that folk psychology's belief construct requires that beliefs be the
cognitive causes of both verbal and nonverbal behavioral effects. But cer-
tain psychological experiments show that verbal and nonverbal behavior
have distinct cognitive causes. Hence, he claims, the right thing to say is
that there are no such things as beliefs.
Second, Stich contends that beliefs should be identical with "naturally
isolable" parts of the cognitive system, but there are possible network
models of the mind within cognitive science that would violate this isolabi-
lity (or modularity) requirement.
These arguments are effectively countered by Terence Horgan and James
woodward.'' I briefly summarize their principal responses. First, even if
the dual-control thesis is partly right, it is still plausible to posit beliefs as
the two sets of causes. (Notice that Stich himself concedes that there would
be "belief-like" states.)18 Second, in m n y cases people display integrated
verbal and nonverbal behavior: for example, lecturing while doing a logic
problem on the blackboard. Here it seems inescapable that the same states
must be causing both verbal and nonverbal behavior. The control systems
must not be wholly independent. So belief states causing both verbal and
nonverbal behavior must be posited at least for this class of cases.
In reply to Stich's second argument, Horgan and Woodward reveal the
implausibility of the natural isolability, or modularity, requirement, point-
ing out that it is an enormously unattractive intertheoretic compatibility
requirement. There is no reason why beliefs could not be identical with
highly complex, even 'gerrymandered', conglomerations of lower-level
events. Such complex events could still play the causal role that folk psy-
chology imputes to them.
170 The Problem of Content
formally, once the contents of these states are assigned, one would be in a
position to assess the truth-conduciveness of the processes, In other words,
even if cognitive psychology does not participate in the assignment of
truth-conditional contents, it can still play the role in primary epistemology
that I have assigned it, that of identifying-in some descriptive terms or
other!-the various belief-forming processes available to the organism. This
completely undercuts (4), which comprises the challenge to my project
being considered in this section.
Although this completes my main response to this challenge, a few other
comments are in order, apropos of premise (1). First, one might reject prem-
ise (1)as descriptively false. Many parts of psychology in fact deal with the
world-mind interface. Perceptual psychologists study the relationship be-
tween physical inputs and perceptual representations, in order to deter-
mine when and why perceptual illusions occur. And even when cognitive
psychologists do not study environment-organism interactions, they com-
monly do employ content descriptions to characterize internal events.28
But suppose (1)is construed as a prescription, rather than a description of
current practice. Perhaps it is a prediction of psychology's practice in some
ideal, future incarnation. On what would such a prescription be based?
One consideration, of course, is the computational model, which has in-
spired many of the themes of cognitive psychology. But notice that even the
computational model makes use of extrinsic descriptions. When we de-
scribe states of a corpputer in terms of numbers and the arithmetic function
addition, we are not giving intrinsic descriptions of those states; we are in-
terpreting the states in terms of some external, abstract objects. So the
computational model itself makes use of semantic interpretation.
A second possible reason for a purely formal psychology is that the true
laws of the system can be formulated only in syntactic, intrinsic terms. But
is this correct? It is admittedly plausible to suppose that if there are any
laws of the system statable in extrinsic terms, there must also be laws stat-
able in intrinsic terms.29But this would not show that no laws are statable in
extrinsic terms. Furthermore, some writers argue that one cannot state all
laws of the system without including laws formulatable in semantic termi-
n o ~ o ~Laws
~ . ~stated
' in intrinsic terminology would be different laws. So if
psychology wants to include all laws of the system, it must work at the se-
mantic level as well as the intrinsic.
A different kind of defense of psychology's use of content descriptions
would reject the heavy emphasis on laws. According to a recent model of
psychological explanation offered by Robert Cummins, the main style of
psychological explanation is not subsumption under laws.31 Rather, it is
"functional analysis": explaining a system's possession of one property by
appeal to properties of its components. If this model of psychological expla-
nation is correct, psychology would not be forced to abandon content de-
The Problem of Content L 73
scriptions even if it turned out that the best (or only) laws of the system
were statable syntactically. Psychology might still want semantic descrip-
tions of the organism at certain levels of description, which would ulti-
mately, of course, be explainable in terms of lower-level organization.
withhold belief ascriptions for this reason. Russell showed that the axiom of
comprehension leads to inconsistency; but this does not mean that people
who previously (ostensibly) accepted the axiom of comprehension did not
really have any beliefs.
Of course, Davidson does not claim that believers must always be rational
or consistent. He only says that propositional attitude attributions commit
us to finding a "large degree" of rationality and consistency in the subjects
of our attributions. But then it emerges that there is no a priori determina-
tion of the specific logical operations that believers must always, or even
regularly, employ. So, as I wish to contend, it presumably rests with em-
pirical science to determine just what these deductive operations are.
I may put the point this way. Since perfect logicality is not a presupposi-
tion of belief attribution, belief attribution at best presupposes minimal log-
icality. This point has been emphasized by Christopher cherniake4' But
what, exactly, does 'minimal logicality' imply? Precisely which deductive
inferences must a cognizer make, as a prerequisite of having beliefs at all? I
do not see how any plausible conditions can be laid down a priori. Formal
logic catalogs all sorts of argument patterns. How should we pick from
these the ones that a believer, qua believer, must (always?) execute. Even if
we laid down a disjunction of such patterns as requisite for believing, no
conclusion could be drawn about which of these disjuncts human beings
regularly, or typically, or easily, execute. As far as human inference capaci-
ties are concerned, the matter is surely an empirical one.
The point can be brought home another way. Before teaching a course in
logic, it is not obvious to the trained Ph.D. exactly which inferences fresh-
men will find easy or hard, natural or unnatural. This is manifestly an em-
pirical matter, and quite appropriately studied by the psychology of
reasoning. (See Chapter 13 for further discussion.)
I have failed to identify any plausible a priori logicality constraints, even
in the domain of deductive logic. But a priori constraints of rationality, if
there are such, should also include inductive logic. Here one hardly knows
where to begin. Principles of inductive logic are very controversial, in the
first place. Assuming we could settle on some such principles, are they
really to be required as conditions for belief ascription? What about princi-
ples of probability and statistics? Are creatures who violate such principles
necessarily devoid of propositional attitudes? That is extremely dubious.41
Empirical studies strongly indicate that people have little native grasp of
probability (as I will discuss in Chapter 14). But this hardly keeps them
from qualifying as believers.
It is time to turn to the second form of the present challenge, the one
concerning reliability. The first comment I need to make is that, contrary to
what Davidson avers, our procedures of content attribution do allow the
ascription of belief systems that are 'massively' false (in the ascriber's opin-
176 The Problem of Content
ion). Cartesian demon scenarios are apt for illustrating this suggestion. But
let me give a slightly more realistic variant of Descartes's demon.
Imagine a child who grows up normally and develops a mastery of
English. Several cognitive catastrophes then befall him. He is stricken by
diseases that distort his visual powers and his memory-except his memory
for language. As a result, he now describes his visual environment in ways
we regard as massively false. And when he recalls things from his personal
past, his accounts are systematically erroneous. To ensure his health and
well-being, he is placed in the hands of a guardian, who feeds and treats him
well except for instruction in cognitive matters. The diabolical guardian
reads him specially concocted newspapers and encyclopedias full of false-
hoods. So the child gradually acquires a radically false picture of the world.
The guardian even teaches him fallacious inference patterns, which take
him even further from the truth. His misconceptions are expressed in avow-
als of belief, and since there is nothing amiss with his linguistic memory, we
have every reason to interpret his utterances homophonically. We would
then be strongly disposed to ascribe to him a system of belief that is mas-
sively false.42
Suppose, nonetheless, that we concede a kernel of truth in the notion that
some sort of truth constraint is presupposed by the ascription of content,
though of a more modest nature than those Davidson endorses. What would
be the repercussions of this concession? A theory of interpretation still
would not fix a belief set's reliability, so the determination of just what that
global reliability is would still be an empirical matter.
Furthermore, even if Davidson were right in claiming that massive error
is excluded a priori, this would not damage my theory. Suppose belief
ascription commits us to an accuracy rate of 50 percent. The criterion of
justifiedness presumably requires a higher truth ratio, say 80,90, or 95 per-
cent. Can people's belief-forming processes attain such higher levels? This
would be an empirical question, not foreclosed by the theory of
interpretation.
Notice next that even if the theory of interpretation requires a high truth
ratio across the corpus taken as a whole, there might be small slums of de-
fective belief-forming processes, perhaps infrequently employed. The task
of identifying a right system of J-rules would include the identification and
rejection of such defective processes. More generally, for purposes of the
theory of justifiedness, it is not enough to know the global truth ratio of a
cognizer's actual beliefs. One needs to know the specific cognitive processes
that are employed and available for employment, since it is only such pro-
cesses that a right J-rule system will specify. Identification of the available
processes is an empirical task.
Finally, all of the foregoing questions concern reliability. But epistemol-
ogy is also interested in power and speed. Evaluation of basic cognitive pro-
The Problem of Content 177
.
cesses along these dimensions is not at all affected by the theory of
interpretation. (At least, no argument has been given for any such thesis.)
Thus, many questions of importance to primary epistemology would re-
quire contributions from cognitive science-ven if some of Davidson's
more radical suggestions were correct.
I conclude that any plausible interpretation theory is compatible with
the role assigned by my framework to empirical psychology. Let us pro-
ceed, then, to concrete illustrations of the contributions that psychology
might make.
part II
ASSESSING O U R C O G N I T I V E
RESOURCES
Perception
Even an initial survey of the field and sampling of the possibilities will oc-
cupy us at length. Finally, a truly complete set of J-rules would have to reg-
ister an interplay between different components, or functions, of the
mind-brain. I shall have enough on my hands in identifying relevant pro-
cesses individually, and exploring their properties.
For all of these reasons, Part I1 presents only some first steps toward a
full-fledgedprimary epistemics: a prolegomenon, if you will, to the subject.
Clearly, a full development of primary epistemics will require continued
research and collaboration by many people from several disciplines.
I have said that in this part of the book I will illustrate prospective con-
tributions of cognitive science to epistemology. The term 'illustrate' needs
emphasis. It is not crucial to my view that any particular item in Part I1
should prove to be correct. I am not wedded to any particular candidate as
a basic cognitive process; nor do I have a significant stake in the normative
assessment of any such candidate. My aim is to show what might be among
the basic cognitive processes and what the epistemological implications of
this may be. Relative to this aim, no specimen hypothesis I adduce about
psychological processes is critical. The structure of the inquiry concerns
me, not the correctness of specific details. However, I naturally select
themes taken seriously by at least part of the cognitive science community.
And I pay some attention to the evidence supporting various hypothesized
processes, although I do not do this systematically or rigorously. Weighing
evidence is the task of cognitive science itself (using currently accepted sci-
entific methods). The strict role of primary epistemology is to borrow the
results of cognitive science and assess the epistemic repercussions of these
results.
While the principal contribution of cognitive science is to identify basic
belief-forming processes, there is another contribution as well: refinement
of the descriptive categories with which epistemology operates. For exam-
ple, epistemology commonly deploys the category of belief. But belief is a
very coarse-grained category from the standpoint of mental science. Pri-
mary epistemology needs a more subtle and flexible set of doxastic state cat-
egories. Drawing on psychological materials, I shall introduce a few such
categorial refinements among doxastic states, distinguishing 'perceptual'
from 'central' beliefs (this chapter), and 'activated' from 'unactivated' be-
liefs (Chapter 10).The whole topic of uncertainty or degrees of confidence,
standardly handled with the category of subjective or 'judgmental' probabi-
lities, also needs to be addressed with increased psychological sophistica-
tion. I shall draw on psychological materials to tender a new approach to
accept-anceand uncertainty in Chapter 15.
Of course, any complication in the array of doxastic categories threatens
to complicate the picture of normative assessment. When activated and
unactivated beliefs are distinguished, questions arise about the meaning of
Perception 183
\\
'
\\
'
Figure 9.1
Perception 187
and grammatical constraints. That is, they were able to use their back-
ground knowledge about meaning and grammar to facilitate correct
identification.
Cognitive psychologists generally believe that perception uses a mixture
of bottom-up and top-down processing. The occurrence of top-down pro-
cessing is the basis for the widely made claim that perception is 'intelligent'.
All one's knowledge about the world--or a lot of it, at any rate--can aid the
construction of a percept.
One dissenting voice is that of Jerry odor.^ Fodor defends the thesis that
so-called input systems-including perceptual systems-are modular, and
one significant feature of these modules is that they are "informationally
encapsulated." This means that pieces of information represented else-
where in the organism are not available for use by such systems. Obviously,
the thesis of informational encapsulation conflicts with the claims of many
proponents of top-down processing.
However, Fodor's statements of his position make it less radical than it
initially appears, as is clear from this passage:
The claim that input systems are informationally encapsulated is equivalent
to the claim that the data that can bear on the confirmation of perceptual
hypotheses includes, in the general case, considerably less than the organism
may know. That is, the confirmation function for input systems does not have
access to all of the information that the organism internally represents.'
Fodor does not deny, then, that some top-down processing takes place.
Rather, he wishes to register three points. First,, he argues that some infor-
mation which the organism represents-in the 'central' systems-cannot
feed back on perception. That is, this information cannot influence the con-
struction of a percept. In a Miiller-Lyer illusion, one line looks longer even
when we know that the two lines are equal in length. What we know ap-
parently does not (in Zenon Pylyshyn's terminology)'0 cognitively "pene-
trate" the perceptual process. Second, Fodor disputes the interpretations of
some of the experiments standardly cited in support of top-down pro-
cessing. There is a question, for example, of whether subjects are really re-
porting how things look or sound, rather than, say, guessing. Third, Fodor
claims that where there is top-down information flow, it occurs within the
perceptual module. The input system cannot access outside information,
but it can deploy internal high-level information to help construct a
percept.
It seems hard to deny that memory information influences perception.
The most striking cases are with fragmented or incoherent figures, for ex-
ample, pictures that initially appear to be an assortment of dots, globs, and
unrelated lines. Once we succeed (perhaps with the help of a hint) in iden-
tifying such a figure as a particular object or group of objects, such as a dal-
188 Perception
over low-level determinants only when exposure times are brief, or when
there is a lot of noise, so that no perceptual conclusion can be reached by
bottom-up processing alone. Given more time and/or less noise, bottom-up
processing may predominate. For example, when one looks leisurely and
carefully at the stimulus in Figure 9.1, one sees clearly that the second let-
ters of each word are identical.
It is possible, then, that even if theory-ladenness sometimes occurs, it can
be excluded in scientific contexts by appropriate experimental designs. In
the case of scientific observation, there is (in general) no bar to ample ob-
servation time and noise reduction. Furthermore, the context of observation
can often be manipulated, so the impact of context can be minimized or
eliminated.
Finally, even when a percept is affected by categories stored in memory,
it is normally possible to switch back to a prerecognition stance: to see the
form independently of the familiar category. For example, even if one nor-
mally sees a drawing of a cube as a cube, with appropriate perspective, it is
also possible to see it in a prerecognition, two-dimensional mode, namely, as
a square and two parallelograms joined together at their edges, as suggested
in Figure 9.2. This again illustrates how perception is capable of releasing
itself from the influence of theory.
I now turn to the more central issue of this section: the impact of top-
down processing on the reliability of perception. Primary epistemology is
concerned with this point for the following reason. Let us assume that per-
ceptual processes can occur with variable amounts of top-down processing,
ranging from none at all to substantial amounts. If there is a substantial dif-
ference in reliability among these different processes, this could well have
an impact on approved systems of J-rules. Suppose, for example, that heav-
ily top-down perceptual processes-those using large amounts of contextual
information-are much less reliable than purely bottom-up processes. Then
perhaps no rules permitting these top-down processes would be included in
any right system of J-rules.
We have already encountered reasons for being suspicious of top-down
processing. In the phoneme-restoration effect, for example, subjects hear
Figure 9.2
190 Perception
legislatures even though the sound corresponding to the first s has not ac-
tually occurred. Only some nonspeech sound, such as a pure tone, has oc-
curred. Clearly, reliance on context can breed error. It does not follow,
however, that top-down processing generally promotes error. On the con-
trary, the contexts used by speech perceivers may generally be reliable in-
dicators of the target phonemes, and hence may generally foster reliability.
Some further experiments make this point clear. One important line of
research on top-down effects comes from experiments on letter identifica-
' ~ D. D. wheeler.'' In these
tion, especially experiments by G. ~ e i c h e r and
experiments subjects were given a very brief presentation of either a letter
(such as D) or a word (such as WORD).Immediately afterward they were
given a pair of alternatives and instructed to report which they had seen. If
they had been shown D, subjects might be presented with D or K as alter-
natives. If they had been shown WORD, they might be shown WORD or
WORK as alternatives. The two word choices, of course, differ only in the D
and K letter. Results of the experiments showed that subjects were about 10
percent more accurate in the word condition. They more accurately discri-
minated between D and K in the context of a word than as letters alone,
even though they had to process more letters in the word context.
Similar effects have been obtained at the multiword level. In an experi-
ment by E. Tulving, G. Mandler, and R. Baumal, for example, subjects were
given varying amounts of context before being shown a target word in a
brief exposure.17 The greater the number of context words preceding the
critical word, the higher the percentage of correct identifications.
As these experiments clearly indicate, top-down processing can increase
reliability, at least for some kinds of tasks. It is noteworthy, indeed, that the
initial example of legisla.tures, in which top-down processing leads to per-
ceptual error, was a deliberately contrived case (like all experimental
cases). One might well speculate that in more ecologically natural settings,
top-down processing does not often produce such errors. At present, then,
we have no reason to think that top-down processing per se is a negative
characteristic of perception, at least from the vantage point of reliability
and justifiedness.
It is worth elaborating this point in a slightly different way. In the litera-
ture on perception it is common to distinguish realist and constructivist
theories of perception. Roughly, constructivism holds that the mind adds
things to the incoming stimulus in the course of producing a percept,
whereas realism tends to deny such activity. For my purposes, the label 're-
alist theory7 is likely to be misleading. So let me substitute a preferable
phrase suggested by Rock: stimulus theory.18 On Rock's characterization
stimulus theories hold that nothing more is necessary by way of explanation
of perception than the relevant stimulus. Such theories would of course
allow that various processes transduce the physical stimulus, encode it
Perception 191
neurally, and transmit it deeper into the brain. But as long as these pro-
cesses are "perfectly correlated" with the sensory input, what goes on in-
ternally is a "direct function" of the incoming stimulus, and can therefore
be ignored by perceptual psychology. The most promising forms of stimulus
theory would stress, however, that proper identification of the effective
stimulus is a subtle matter. The stimulus for perceived size need not be the
size of the object's visual angle subtended at the eye: it could be the rela-
tionship (the ratio) of the visual angle of the object to that of other neigh-
boring objects. This emphasis on higher-order features of the stimulus
derives from the work of J. J. Gibson, the most prominent perceptual realist,
or in my terminology, stinlulus theorist.''
Contrasting with the stimulus theory are the various sorts of constructi-
vist theories. As indicated, the common ground of such theories is that some.
internal constructive process occurs that mediates between the incoming
stimulus and the percept; something gets added to the stimulus.
We might say, very roughly, that constructivist theories tend to postulate
top-down processing, whereas stimulus theories postulate purely bottom-up
processing. Furthermore, the contrast between constructivist and stimulus
theories seems to invite conclusions about reliability. If the perceptual pro-
cesses add things to the stimulus during perceptual processing, isn't this
likely to distort the perceptual output? Won't there be a tendency for the
percept to get things wrong? Conversely, if stimulus theories are correct,
doesn't this bode well for the accuracy of perceptual processes? Indeed, this
prospect of accuracy, or truth, is already intimated by the very term 'realist'
(which is why I prefer to avoid it).
However, these conclusions are completely unwarranted. As the experi-
ments I have cited indicate, the deployment of information obtained from
sources other than the target stimulus need not be distorting. On the con-
trary, proper use of accurate contextual information can enhance reliabil-
ity. Furthermore, real-world regularities may have been incorporated into
perceptual systems in the course of evolution. So if, as constructivism often
suggests, perceptual systems automatically add such information in the
course of processing--or reveal innate preferences for certain kinds of an-
swers or solutions to perceptual questions-this need not be detrimental to
reliability.
Until now my discussion has been restricted to the standard of reliability.
But other standards of appraisal should also be explored. Let me turn first to
the standard of speed (or, rather, the combination of reliability and speed).
The relevant literature is readily interpreted as demonstrating the contri-
bution of top-down processing to increases in speed, the ability to make
correct perceptual judgments in less time. In the Tulving, Mandler, and
Baumal study, for example, subjects were shown a target word, such as dis-
order, for varying brief exposure times. In one condition this word was pre-
192 Perception
slow, it would take too much time to make them all. From a pragmatic
vantage point, then, some accuracy is worth sacrificing for speed.
Now I have not thus far endorsed pragmatic values, such as need satisfac-
tion or desire satisfaction, as species of epistemic values. But an appeal to
such pragmatic factors certainly makes sense. Although I rejected this prag-
matic approach to iustijiedness (in section 5.4), this does not bar the prag-
matic dimension from appearing in a more inclusive survey of epistemic
values. It is worth asking, then, how well the perceptual systems perform
from this pragmatic standpoint.
In particular, let us consider how perceptual pattern recognition may be
influenced by affect: by strong desires and aversions, or other emotions.
Suppose that pattern recognition proceeds in terms of matching sense-
derived materials to a stored pattern representation. Then make three other
assumptions: (1)A match can be effected even if there is only partial corre-
spondence with the stored pattern. (2) Greater activation of a stored pattern
facilitates the matching of that pattern. (3) Patterns with high affective
value--either positive or negative-receive a higher level of activation.*
Given these assumptions, there will be an enhanced tendency to perceive
stimuli in high-affect categories. For example, a young child who has a
strong desire to see her mother will recognize lots of women in the distance
as her mother. The parent in the shower, listening for its infant's cry in the
next room, may hear such a cry when in fact it is only the noise of the
shower. A gardener fearful of snakes will see a snake when it is really only a
piece of garden hose.
These characteristics of a cognitive system probably deserve high marks
from a pragmatic standpoint. Pragmatically, erring on the side of false posi-
tives is prudent. Nothing much is lost if the child mistakes another woman
for her mother, if you mistake your garden hose for a snake, or shower noise
for your infant's cry. It is cost-effective to accept some false positives in
order to minimize errors of omission.
But processes that apply desire and aversion parameters are not condu- ,
A
Figure 9.3
called basic beliefs--few theorists would hold that they must be incorrigi-
ble. The essential requirement for basicness, then, remains to be specified.
Another facet of Descartes's F-ism held that the class of foundational be-
liefs consists in beliefs about one's own current mental states (or one's own
existence-as a purely mental entity). And this aspect of Cartesianism is
also unfashionable nowadays. F-ism could allow other kinds of beliefs to be
basic, for example, perceptual beliefs about physical objects or events in the
environment. Current epistemologists tend to formulate F-ism as a doctrine
that is neutral about the specific subject matter of the basic beliefs. I will
follow this neutralist mode of formulation.
Even within this core, neutralist conception of F-ism, there are two dis-
tinct variants: strong F-ism and weak F-ism. Strong F-ism says that the jus-
tificational status of basic beliefs is wholly independent of the rest of one's
doxastic corpus. Basic beliefs are justified no matter what other beliefs the
cognizer currently holds. Weak F-ism says that the justificational status of
basic beliefs is partly independent of the rest of one's doxastic corpus. There
is some nondoxastic source of justificational status, or warrant-but perhaps
some doxastic source as well. This characterization would still allow a dis-
tinction between basic and nonbasic beliefs, since nonbasic beliefs would
get whatever warrant they have entirely from other beliefs.25
We come now to the formulation of C-ism. One approach to C-ism sees it
as the denial of F-ism, another sees it as an autonomous doctrine. Viewed in
the first way, C-ism is the doctrine that there are no basic, or foundational,
beliefs. But this doctrine admits of different interpretations, given different
versions of F-ism. Thus, one version of C-ism is the denial of strong F-ism,
another is the denial of weak F-ism. The former doctrine would hold that no
beliefs receive their justificational status wholly from nondoxastic factors.
The latter would hold that no beliefs receive their justificational status even
partly from nondoxastic factors: all justificational status derives from rela-
tions with the cognizer's other beliefs.
The second approach to C-ism, the autonomous approach, would avoid
characterizing the doctrine as merely the denial of F-ism. It might give the
following sort of generic formulation:
A belief is justified if and only if it belongs to a coherent system of beliefs of
the cognizer.
This leaves at least two questions to be answered: (a) what is it for beliefs to
'cohere'? and (b) which system of the cognizer's beliefs must cohere in order
for a member belief to be justified?
Conceming question (a), answers might take a negative or a positive
form.26A negative answer might be: beliefs cohere just in case they are not
inconsistent. A positive answer might be: beliefs cohere just in case they
mutually support one anotherF7 Concerning (b), answers might again be of
Perception 19 7
two sorts: holistic and nonholistic. Holistic C-ism says that the relevant sys-
tem for purposes of the coherence theory is the cognizer's entire doxastic
system. Nonholistic C-ism would admit proper parts of an entire doxastic
system.
Under the influence of Quine's holism, many current formulations of
C-ism tend to be holistic. But there is a problem with this approach. Sup-
pose that the construal of 'coherence' implies that the lack of inconsistency
of the relevant corpus is a necessary, if not sufficient, condition for coher-
ence. Then if one requires holism, coherence of the entire doxastic corpus,
someone with an inconsistent set of beliefs cannot have any justified beliefs.
This is strikingly counterintuitive. If some inconsistency has arisen in one
comer of your doxastic corpus, it should not disqualify all the rest of your
beliefs from being justified. So it is doubtful that an optimal formulation of
C-ism is one that requires holism. Notice too that C-ism viewed as the de-
nial of F-ism need not involve holism.28Clearly, not all varieties of C-ism
need be holistic.
Tuning now to my own theory of justifiedness, the question is, What is it
for a given J-rule system to be F-ist or C-ist in character? Again, the differ-
ent conceptions of F-ism and C-ism sketched above will dictate different
replies. On one conception the critical question is whether any rules in the
set of J-rules permit the use of belief-forming processes whose inputs are
wholly nondmastic. If so, this set of J-rules would instantiate strong F-ism.
If not, it would instantiate C-ism viewed as the denial of strong ~-isrn."
On an autonomous and holistic conception of C-ism, however, the fore-
going would not be the critical question. According to holistic C-ism, a set
of J-rules will be C-ist only if the rules authorize only holistic processes,
processes that take as inputs all current doxastic states. In other words, a
process must be (in Fodor's terminology) "Quinean," in the sense of being
sensitive to the cognizer's entire belief system.30
My reliable-process criterion of rightness implies that whether a set of
J-rules is right depends on two things: (1)the cognitive processes available
to human beings, and (2) their reliability. Whether there are any right F-ist
J-rule systems, or any right C-ist J-rule systems, also depends on these facts.
Hence, whether right J-rule systems are F-ist or C-ist in character depends
on psychological facts.
Let me apply these points to the area of perceptual processes. It should
be recalled, in this context, that I interpret a percept as a belief, a percep-
tual system's judgment about environmental objects (see section 9.2).Such a
belief can be rejected by the central system. This is just what occurs when
vision 'says' that the stick immersed in water is bent, and the central system
'says' that it is straight. Admittedly, calling perceptual outputs 'beliefs' di-
verges from ordinary usage; but it is theoretically fruitful. It makes sense of
the intuitively compelling feeling that the central system disagrees with the
198 Perception
Memory
p' entails the truth of p; and 'S remembers event E' entails that E really oc-
curred. For present purposes, though, I set aside this factive sense. When I
speak of remembering a proposition or event, I do not mean to imply that
the recalled proposition is true, or the recalled event actually occurred.
'Memoryy will refer, roughly, to ostensible memory.
Questions about the power and reliability of human memory will be
treated in later sections. First I wish to address another issue to which the
psychology of memory is highly germane. This concerns the concept of
belief.
One way psychology can impinge on primary epistemology, I have sug-
gested, is by refining its descriptive resources. In this chapter I shall illus-
trate this point by arguing that memory research demands a revision--or
enrichment-in the concept of belief. Twentieth-century epistemology has
largely rested content with the conceptual tools of ordinary language, in-
cluding the belief concept. But this concept is really psychologically inade-
quate. In the first two sections of this chapter, I will identify this
inadequacy and suggest how the psychology of memory can improve the
situation.
Many recent writers have worried about the content aspect of the belief
concept, as we saw in Chapter 8. The inadequacy of the belief concept on
which I focus here is quite different. There are (at least) two different kinds
of belief, and failure to distinguish these kinds and heed this distinction sys-
tematically undermines the attempt to provide satisfactory and realistic
epistemic principles.
In Chapter 9 I proposed a dichotomy between perceptual and nonper-
ceptual beliefs. In this chapter, though, I will ignore perceptual beliefs en-
tirely. The distinction examined here is a further distinction, wholly within
the more commonly acknowledged class of beliefs, so-called central-system
beliefs.
To believe a proposition at a given time, a person need not think of it, or
entertain it, at that time. Five minutes ago you doubtless had a belief about
where you were born, but you probably were not ruminating on your birth-
place five minutes ago. Yet a person often does entertain beliefs; they oc-
cupy attention and get actively invoked in cognitive tasks. These two
classes of belief are sometimes noted by philosophers of mind, who use the
phrases 'dispositional belief and 'occurrent belief to mark the two classes.'
But cognitive psychology would have us think of these two classes in a
slightly different way.
According to the popular 'duplex' theory of memory, the mind has two
memories: long-term and short-term (or actiue) memory. (A third kind of
memory, sensory information storage, is ignored here.) Long-term memory
(LTM) is a place where information can be stored for very long periods of
time, perhaps indefinitely. Information in LTM decays very slowly. Short-
term memory (STM), by contrast, has a much faster decay period. Unless
refreshed, or rehearsed, information fades fairly rapidly from STM. There is
also a big difference in the a m n t of information retainable in LTM as
compared with STM. LTM is virtually unlimited in capacity, whereas STM
has a very restricted capacity. Furthermore, whereas information lies dor-
mant in LTM and is not capable of influencing cognitive activities, STM is a
sort of 'warkplace', where various operations and transformations can be
performed on data, and new items can be constructed. In order to utilize
material stored in LTM, that material must be retrieved and copied into
STM. That's what it is to 'think of something one has previously known or
believed.
Although this duplex portrait of memory has been popular for about a
decade and a half, experiments during the 1970s have posed difficulties for
Memory 201
it, particularly for the notion of STM as a single entity. The properties used
to define STM have been split apart, so it is no longer clear which of them
(if any) should be taken as constitutive of a short-term store.2 ow ever, I
shall stick with the older model in my arguments. If anything, the new re-
search suggests that belief-state phenomena are even more complicated
than the older model suggests, an idea that would strengthen my case for
making distinctions. For expository purposes, it will be convenient to rest
content with the older model. It should be recalled, throughout, that my
discussion is only illustrative of the potential impact of psychological re-
search. Not all details of contemporary findings need detain us.
John Anderson's recent treatment of memory in The Architecture of Cog-
nition departs slightly from the duplex theory.3 Anderson views active
memories not as events occupying a special location, but as activated states
of LTM. Activation is something like a light bulb lighting up. Indeed, on
Anderson's theory, it is like a light bulb governed by a dimmer switch: acti-
vation can take on a continuous range of levels, not just on and off. For pre-
sent purposes, though, I will confine my attention to on and off states:
'activated' and 'unactivated' states. These can be identified with the two
sorts of belief states I spoke of earlier, which I will call activated beliefs-
A-beliefs-and unactivated be1iefs-U-beliefs4
Two objections might be lodged against the attempt to identify so-called
dispositional beliefs with states of LTM. First, it may be claimed, not all
states of LTM are belief states. Second, not all dispositional beliefs are
states of LTM.
The first objection has three elements. (1)Not all LTM states are belief
states because some LTM states are goal states. Our long-range goals, or in-
tentions, are stored in LTM; but these are not belief states. (2) Some things
may be stored in LTM, but neither as beliefs nor as goals. For example, a
sentence may be salted away in LTM because we like the sound of it, or just
because it has a sort of staying power in the imagination.5 This does not
mean that we believe the sentence. (3)Some contents stored in LTM are so
weak-whether through decay, repression, interference, or what have
you-that they are no longer retrievable, even with the best of retrieval
cues. It is not very plausible to say that these contents are still believed.
Argument (2) is not entirely persuasive. When a sentence is stored, there
may always be one or more beliefs associated with it, such as 'I have heard
this sentence before', or 'This sentence was uttered on such-and-such an oc-
casion'. However, elements (1) and (3)seem decisive enough. But what do
these points establish? Only that not all states of LTM are belief states; we
may still retain the view that some-presumably many-states of LTM are
belief states.
Turn now to the second objection, which claims that not all dispositional
beliefs are LTM states. Daniel Dennett, for one, is inclined to say that we
202 Memory
have indefinitely many beliefs that are not explicitly stored in LTM (or else-
where). His examples include: "New York is not on the moon, or in Vene-
zuela"; "Salt is not sugar, or green, or oily"; "Tweed coats are not made of
salt"; "A grain of salt is smaller than an elephant."6 These are all proposi-
tions a person may never think of, and hence are not stored in LTM. Still,
they are things he probably believes.
Personally, I find this claim unconvincing. If a person has never thought
whether New York is in Venezuela, then he does not already believe that
New York is in Venezuela. If he has never asked himself whether zebras
wear codpieces in the wild, then he does not already believe that zebras
don't wear codpieces in the wild. He may be disposed to assent to these
propositions unhesitatingly, the moment they are queried. But this just
shows he has beliefs from which these conclusions would be readily in-
ferred. It does not show he already believes each conclusion, prior to the
question being raised.
I acknowledge, however, that many philosophers do not share my intui-
tions. They hold there are non-occurrent beliefs that also are not explicitly
represented in LTM.' Whether or not they are right is not critical here,
however. I can accept the notion of nonexplicitly represented beliefs, as
long as we also retain the two categories of explicitly represented beliefs,
activated and unactivated beliefs. Since I want to restrict attention to these
two categories, I will cheerfully leave the issue of nonexplicitly represented
beliefs unresolved.
My main point about the distinction between A-beliefs and U-beliefs
concerns normative epistemic principles; I will come to this in section 10.2.
First let me comment, though, on the descriptive importance of this dis-
tinction. If the notion of belief is to be even moderately useful for predictive
and explanatory purposes, the distinction between A-beliefs and U-beliefs
must be observed.
Here is how a failure to observe this distinction can derail a prediction.
Suppose Melanie is deciding, one Sunday morning, whether to go to the
university library to work. She seldom works there on Sundays, but today
she needs to read a certain important article there. Now Melanie knows
that the library normally opens at 7:00 A.M. She has also noticed, on many
occasions, that the library's Sunday hours are different: it only opens at 1:00
P.M.If she were to recall the latter piece of knowledge, she certainly would
not make the trek to the library this morning; it would be a waste of time.
Alas, she forgets this piece of information and decides to go there after all.
When, upon arrival, she sees the posted hours, she upbraids herself as fol-
lows: 'Wasn't that silly of me! I knew the library doesn't open on Sundays
until 1:0OY.
Now suppose we worked with a belief-desire model of human decision
that countenanced only one kind each of belief and desire. Within each
Memory 203
class the only differences in efficacy would rest on strength. Then such a
model would presumably predict that Melanie would decide not to go to
the library this morning. After all, her corpus of beliefs and desires include
the following:
(B1) The library normally opens at 7:00 A.M.
(B2) Today is Sunday.
(B3) On Sundays the library opens at 1:00P.M.
(Dl) It is most desirable to read a certain article soon, which is only ob-
tainable at the library.
(D2) It is desirable to do a good bit of other work too today.
(D3) It is undesirable to waste time going places that don't allow essential
work to be done.
Since Melanie believes today is Sunday, and believes that the library does
not open before 1:00 on Sundays, the model should predict that she will de-
cide not to go this morning. This would be a wrong prediction, of course,
but one to which such a model seems committed.
Once we draw the distinction between A-beliefs and U-beliefs, and take
account of the fact that only A-beliefs are efficacious in decision making,
this prediction can be reversed. While the items listed above constitute
U-beliefs and U-desires-they are part of Melanie's LTM store-they do
not all get activated at the time of decision. At the moment of decision,
Melanie fails to retrieve the belief that today is Sunday; or fails to retrieve
the belief that the library opens at 1:00 on Sundays. The critical activated
belief at the time is the belief that the library normally opens at 7:00 A.M.
Now if a belief-desire model incorporates the A-belief/U-belief distinction,
and if it portrays decisions as governed by A-beliefs only, such a model
could correctly predict Melanie's decision to go to the library. Clearly, this
would be a better variant of the model.
It is a familiar occurrence that people often do what Melanie does-they
fail to access part of what they know, and therefore perform silly actions.
Failure to distinguish U-beliefs and A-beliefs makes it impossible to account
for these occurrences.
Turn now to theoretical inference, where the situation is perfectly analo-
gous. Which inferences a cognizer makes is determined not by his beliefs
stored in LTM, but by those that get activated, by his A-beliefs. A person's
U-belief corpus is enormous, and only a relatively small portion of it is acti-
vated at any one time. Cognizers often fail to make connections between
elements in this corpus, even though they are relevant to the theoretical
tasks at hand.
For example, at least a decade before Fleming's discovery of penicillin,
many microbiologists were aware that molds cause clear spots in bacteria
cultures; and they knew such a bare spot indicates no bacterial growth. Yet
they did not consider the possibility that molds release an antibacterial
Memory 205
both inability to recall and errors of recall. But in this section I shall con-
centrate on two other memory processes that psychologists have investi-
gated: elaboration and reconstruction.
J . D. Bransford, J. R. Barclay, and J. J. Franks studied the ability of sub-
jects to recognize previously presented sentences, such as The ants were in
the kitchen, The ants ate the sweet jelly, The ants in the kitchen ate the jelly
which was on the table, and The jelly was sweet.l4 When subjects were sub-
sequently asked to try to recognize sentences, some of which had been pre-
sented and others of which had not, they tended to respond on the basis of
how much of the total information a test sentence conveyed. They were
more likely to misrecognize a nonpresented sentence involving all the pre-
sented semantic information-for example, the ants in the kitchen ate the
sweet jelly which was on the table-than a shorter sentence which had in
fact been presented. Apparently, subjects encode the sentences in a seman-
tic structure that goes beyond the individual sentences actually presented.
An interpretation is constructed that elaborates upon the information ac-
tually presented.
Another example of this elaborative process occurs in an experiment by
R. A. Sulin and D. J. ~ o o l i n ~ .One
" group of subjects read the following
passage :
Carol Harris was a problem child from birth. She was wild, stubborn, and
violent. By the time Carol turned eight, she was still unmanageable. Her par-
ents were very concerned about her mental health. There was no good insti-
tution for her problem in her state. Her parents finally decided to take some
action. They hired a private teacher for Carol.
A second group of subjects was given the same passage, except that the
name Helen Keller was substituted for Carol Harris. A week after reading
the passage subjects were given a recognition test in which they were pre-
sented with a sentence and asked to judge whether it was in the passage.
One of the critical sentences was She was deaf, dumb, and blind. Only 5 .
percent of the subjects who read the Carol Harris passage accepted this
sentence, but 50 percent of the subjects who read the passage as about
Helen Keller thought (mistakenly) they had read this sentence. Apparently,
the latter subjects had elaborated the story they actually read with facts
they previously knew about Helen Keller.
These experiments leave it indeterminate whether construction occurs at
the time material is originally encoded into LTM or whether a reconstruc-
tive inference occurs at the time of test. Other experiments, however, seem
to demonstrate that both of these phenomena commonly occur.
An additional experiment on the Carol Harris passage by Dooling and
R. E. Christiaansen shows that inferences can be made at the time of test.16
They had subjects study the passage and told them a week later, just before
210 Memory
test, that Carol Harris really was Helen Keller. Subjects still made many
errors, accepting such sentences as She war deaf, dumb, and blind. These
subjects were clearly making an inference at the time of test (or recall).
But elaboration at the time of encoding is also well accepted and docu-
mented. T. S. Hyde and J. J. Jenkins had subjects read groups of twenty-four
words presented at the rate of 3 seconds per word." One group was asked
to check whether each word had an e or a g. The other group was asked to
rate the pleasantness of the words. It is reasonable to assume that the pleas-
antness rating was a task that required more elaborate processing than the
letter verification task. In rating a word for pleasantness, subjects had to
think about its meaning, which gave them an opportunity to elaborate on it.
For instance, a subject presented with duck might think, 'Duck--oh yes, I
used to feed the ducks in the park; that was a pleasant time'. Subjects rating
for pleasantness showed much higher recall than those with the letter veri-
fication task. This was apparently due to elaborative processing.
In the experiments just reviewed elaboration and reconstruction often led
to false recall judgments. So elaboration and reconstruction seem to foster
unreliability. This fact is important not only in abstract, theoretical ap-
praisal of memory mechanisms, but also in practical contexts, where we
need to decide whether to trust someone else's memory judgments. Such
practical exigencies arise in the courtroom, where decisions appeal in part
to the memories of eyewitnesses. The memory mechanisms I have been de-
scribing offer opportunities for errors in eyewitness testimony, as Elizabeth
Loftus has documented.18
One problem in eyewitness testimony is simply the period of retention.
R. N. Shepard tested thirty-four clerical workers for recognition of pictures
after intervals of two hours, three days, one week, and four months.'' He
found that retention of the picture material dropped from 100 percent cor-
rect recognition after two hours to only 57 percent correct after four
months. The latter figure essentially represents mere guessing. However,
not concerned at the moment with retention intervals, I want to look at the
impact of reconstruction and elaboration on recall. Loftus cites relevant
work on this topic, including some of her own studies. Specifically, she dis-
cusses how post-event information-information given after a witnessed
event occurred--can distort one's recollection of that event.
An early illustration of this phenomenon was presented by C. ~ i r d . ~ '
During the course of a routine classroom lecture, the instructor was discuss-
ing the results of a series of experiments. A well-meaning but not very
thoughtful reporter on the local newspaper printed an account of the lec-
ture riddled with errors. Many students read the newspaper account, and
nearly ail of these thought it was accurate. The instructor gave an exam at
the end of the week and after the usual set of exam questions asked students
to indicate whether they had read the press account. Those who had read
Memory 21 1
the article made many more errors on the exam. They remembered the er-
roneous information from the newspaper, assuming they had learned it
from the instructor's original lecture.
In several experiments Loftus and colleagues tested subjects to see
whether post-event misinformation would distort their later recollection. In
one experiment nearly 200 subjects viewed a series of thirty color slides de-
picting successive stages in an accident involving a car and a pedestrian.21
The car was a red Datsun shown traveling along a side street toward an in-
tersection, where there was a stop sign for half of the subjects and a yield
sign for the others, and where the accident occurred. Immediately after
seeing the slides, the subjects were asked some questions, one of which was
critical. For half of the subjects the critical question was, Did another car
pass the red Datsun while it was stopped at the stop sign? The other half of
the subjects were asked the same question with stop sign replaced by yield
sign. For some subjects, the sign mentioned was the one they had actually
seen: the question gave them consistent information. For the remaining
subjects, the question contained misleading information.
Later a recognition test was administered. Pairs of slides were presented
and subjects had to indicate which member of each pair they had seen be-
fore. The critical pair was a slide depicting the Datsun stopped at a stop
sign and a nearly identical slide depicting it at a yield sign. When the inter-
vening question had contained consistent information, 75 percent of the
subjects responded accurately. When the question had contained mislead-
ing information, only 41 percent responded accurately. If subjects had been
simply guessing, they would have been correct about 50 percent of the
time; so the misleading information contained in the question reduced their
accuracy below the guessing level,
Apparently, people tend to use post-event information to reconstruct
episodes they have observed. Furthermore, they are not always able to dis-
entangle the original sources of their information. So even though they ac-
quire certain information (or misinformation) after the event, they often
think it was contained in what they saw.
Is this another piece of evidence for the unreliability of the reconstruc-
tion process? Superficially, this is so. But we should proceed cautiously. In
Loftus's cases, error is introduced into the subjects' cognitive systems by
misleading questions provided by the experimenter. The initial fault, then,
does not reside in memory, at least not in the unreliability of memory. It is
true that the mistakes could be averted if the subjects confidentIy recalled
exactly which sign-a stop sign or a yield sign-they originally saw. If so,
they presumably would not be misled by the question, for they would reject
what the question presupposes. But failure to recall this accurately is a defi-
ciency in memory power, rather than in memory accuracy. Similarly, any
failure to disentangle or isolate the original sources of information could
212 Memory
In storing all the above items in LTM, the subject creates a much more
complicated memory structure, which incorporates the elaborations on the
original sentence.
How can these elaborations lead to better memory? In two ways. First,
they provide additional retrieval routes for recall. Suppose that at the time
of test, the subject is given the word doctor and asked to retrieve the origi-
nal sentence. The link from the node representing doctor to the target node,
representing the sentence The doctor hated the lawyer, may be too weak to
revive the latter. But because elaborations have led to additional associative
links, recall may still be possible. From the prompt doctor, the subject might
recall the proposition that the lawyer sued the doctor for malpractice. And
from here the subject might be able to recall the target node. Thus, an al-
ternative retrieval route may succeed if the more direct one fails.
A second way in which elaboration aids memory is that it can enable
someone to infer the target information. For example, the subject who can-
not immediately recall the target sentence from the prompt doctor might
think as follows:
I cannot remember the target sentence but I can remember conjecturing
that it was caused by the lawyer suing the doctor for malpractice and I can
remember it was a sentence with a negative tone.
From this he might infer that the target sentence was The doctor hated the
lawyer. L. M. Reder and R. J. Spiro have argued that in real life a great deal
of inference is involved in Talk of retrieval or activation may be
somewhat misleading, then, if the so-called remembered beliefs are heavily
influenced by inferential reconstruction.
In looking for epistemological consequences, we should not rush to the
conclusion that elaboration and reconstruction yield unjustified beliefs.
Whether the processes would be licensed by a right J-rule system depends
on their contribution to such a system's truth ratio. If both elaboration and
reconstruction are well constrained by prior information in LTM, if the lat-
ter information is true, and if the inferential processes are suitable, the con-
tribution to the rule system's truth ratio may well be acceptable. (Of course,
this will also depend on the chosen degree of reliability needed for
justifiedness.)
Notice too that there are all sorts of amounts and styles of elaboration and
reconstruction that could be discriminated. Some of these amounts and
styles would have a more deleterious effect on truth ratios than others. A
214 Memory
right J-rule system might license processes incorporating some amounts and
styles of elaboration and reconstruction, but not others.
The processes of elaboration and reconstruction exemplify a situation
that we have encountered before and will encounter again: a trade-off be-
tween reliability and power. It is clear that these processes (can) purchase
increases in memory power, though perhaps at some cost in reliability.
Our native, uncritical procedures seem content with the purchase price,
and nothing in my principles of epistemological appraisal dispute its
appropriateness.
Actually, though, it is not clear that there is always a net trade-off, that
power and reliability always conflict. Although I often speak of acquiring
power at the cost of reliability, power can also foster reliability in a J-rule
system. At least a case for this suggestion is readily made, in connection
with inductive inference. First, the more experiences a person correctly re-
calls, the larger the number of true beliefs available as premises for induc-
tive inference. Second, the larger the number of true premises employed in
inductive inference (that is, the greater the total evidence that gets used),
the greater the likelihood of arriving at true conclusions. So there are ways
in which recall power can promote the overall reliability of a cognitive sys-
tem, or the reliability of a set of J-rules. To some degree, then, power sup-
ports reliability, and does not conflict with it.
-
Actors Observen Actors Observers Actors Observers
Estimated Predicted Rated Task
lni tial Future Ability
Success Success
Figure 10.1
such powers. She can also easily find causal explanations of her perform-
ance: her familiarity with the writings of a famous novelist who recently
committed suicide, or her part-time job as a paramedical assistant, or her
open relationship with her parents all might explain her high level of ability
at a task requiring social sensitivity. Once these additional facts or explana-
tions are generated, they are found convincing, and they remain to support
her assessment of her ability even after she receives a debriefing from the
experimenter.
Although there is no direct support for this account of the perseverance
phenomenon, I agree that it is quite plausible. In the remainder of my dis-
cussion I will assume that this sort of process underlies the phenomenon.
Notice, though, that this process can be redescribed as an instance of elabo-
ration. Jane is finding other things in memory with which her apparent good
performance on the suicide note task coheres. She is finding a meaning in
this good performance by linking it up to other (putative) facts in her life or
traits she possesses.
Let us now turn to the normative appropriateness of this sort of process.
Nisbett and Ross are highly critical of it, seeing it as an instance of two
habits, or propensities, which they deem normatively wrong. First, they see
it as an instance of confirmation &us. They cite a rich research literature
"that shows the operation of a variety of encoding and decoding biases that
favor confirmation of prior hypotheses or beliefs over disconfirmation . . .
People tend to recognize the relevance of confirming cases more readily
than that of disconfirming ones, and therefore tend to search for such cases
in evaluating their hypotheses."29 They cite experiments by P. C. Wason
and P. N. Johnson-Laird, M. Snyder and W. B. Swann, and Snyder and M.
Cantor, in support of this claim.30
Second, they see the hypothetical process as an instance of searching for
causal explanations, which are too readily generated and accepted even
when the evidence for them is skimpy or nonexistent. Such explanations
have a marked effect, they say, even when later undermined. They cite an
experiment by L. Ross, M. R. Lepper, F. Strack and J. L. Steinmetz in which
subjects were asked to place themselves in the position of a clinical psychol-
ogist trying to understand and predict a patient's b e h a ~ i o r . Subjects
~' were
given an authentic clinical case history, and some were asked to use this
case history to explain a critical event in the patient's life, whereas other
subjects were not asked to give any such explanation. After the explanation
task subjects were informed that the events they had been asked to explain
were purely hypothetical, and that there was no available information
about the later life of the patient. Then subjects were asked to assess the
likelihood of a number of possible events in the patient's later life, inchding
the critical events. The subjects' likelihood estimates revealed a marked ef-
fect of the explanation task. Subjects who had explained an event rated it as
218 Memory
more probable (even after debriefing) than subjects who had not explained
it.
We should not accept Nisbett and Ross's characterizations of people's
habits too uncritically. Concerning confirmation bias, for example, they
themselves admit that some research indicates that confirming instances are
not always more 'available' to a theory holder than disconfirming ones. As
R. Hastie and P. A. Kumar noted, surprising or incongruent events may be
attended to and stored in memory more often than expected or hypothesis-
confirming events.32 Concerning causal explanation, it is not clear just how
readily people manufacture causal explanations. That they readily invent
causal explanations when given such a task by an experimenter hardly
shows that they always invent them spontaneously, even on the slimmest of
evidence. Clearly, much empirical work remains to be done on these
matters.
Turning to the normative question, suppose that people regularly search
for other material in memory that coheres with, or even explains, some
newly acquired belief. Is this good or bad? An epistemologist might disagree
with Nisbett and Ross's negative assessment on the grounds that search for
coherence is a desirable epistemic trait, and that the search for possible
causal explanations is one piece in such a coherence search. Such searches
are a legitimate-perhaps even a required-phase of belief formation. Un-
less a cognizer checks for coherence, his chances of true belief formation are
greatly reduced. (For some further discussion of this point, see Chapter 15.)
Of course, we should distinguish between attempts to assess coherence
and the specific procedures used for assessing coherence. Perhaps Nisbett
and Ross mean to be critical of the specific procedures people use, not their
proclivity to investigate the question of coherence. But three responses
should be made here. First, since many debriefed subjects do not suffer from
belief perseverance, as the studies show, these subjects may be using per-
fectly acceptable procedures (even by Nisbett and Ross's standards). Sec-
ond, among the subjects who do persist with their initial belief, despite
debriefing, their coherence-determining procedures may be all right.
Maybe they legitimately found reasons to bolster this belief (even if the be-
lief is false). Third, we should be quite prepared to learn that subjects differ
from one another in their coherence-determining procedures; that some of
them fail to conform to any right system of J-rules whereas others do so
conform. After all, subjects differed in their post-debriefing test responses,
so they should not all be lumped together. Nisbett and Ross have a tendency
to make blanket statements about human traits, without full allowance for
observed variation in subject responses.
However, there is a further point, specially connected to memory, that I
have yet to stress. I remarked above that the tendency to look for anteced-
ent beliefs which cohere with a new item of information may be classified as
elaboration: the cognizer seeks to connect the new item with other things
previously stored in memory. In the previous section we saw that elabora-
tion serves to strengthen memory, that is, to improve recall at later junc-
tures. This increase in memory power is surely an epistemically desirable
outcome. It should be weighed as one factor in the overall normative assess-
ment of the process in question. If it is significant enough, it may render the
process, on balance, normatively acceptable.
There are several complications that need mentioning here. First, I have
not offered any principles for overall normative assessment. I have only of-
fered a criterion of justifiedness, and this does not incorporate the element
of power. (Arguably it should; but for now let's put that aside.) However,
we also saw at the end of section 10.3 that power can positively contribute
toward reIiability. So elaboration may enhance reliability via its enhance-
ment of power, and this could bolster its claim for admissibility by a right
system of J-rules. Nisbett and Ross, although they speak frequently of
"human judgmental failings" and even "perversity," offer no principles for
normative evaluation^.^^ Perhaps they would not count power as a relevant
factor at all, for normative purposes. And perhaps they would not accept
the truth-ratio consequentialism that I have proposed. This just underlines
the need for general criteria of epistemic evaluation.
Suppose we focus, though, on my own criterion of justifiedness. It is a
very delicate matter to decide just which processes, among those presently
at issue, should be approved by a system of J-rules. Since it is always possi-
ble that a newly acquired belief is false, how much should such beliefs be
strengthened in memory? How much should they be combined with other
material in memory to generate new beliefs? In the case of Jane, should the
subject combine the (apparent) fact that she has performed well on a cer-
tain task with other facts in memory to make inferences about her general
traits (for example, her social sensitivity)? When focusing on these experi-
mental cases, where it is a datum-to the reader!-that the newly acquired
belief is false, it is all too easy to say that the belief should not have been
combined with others to generate further inferences. We are inclined to say .
this because we know that the belief is false. But a J-rule system needs to
authorize general processes. A general process of finding coherent beliefs
and making new inferences may have a high truth ratio, as long as the class
of input beliefs has a high enough truth ratio. In short, a right J-rule system
might well sanction just what many (even most) of the subjects in Ross's ex-
periments do. Many of those subjects arrive at errors; but these errors are
only a tiny sample of the set of cases that determine the relevant truth ratio.
I have been discussing the normative status of belief perseverance in
terms of the (hypothetical) processes that are actually at work. But perhaps
this is the wrong angle. Instead of focusing on what subjects do that they
shouldn't, perhaps we should focus on what subjects fail to do that they
should.
At time t,, Jane forms a belief (based on feedback) about her task per-
220 Memory
formance; call this p. At time t2, she infers from p (plus other beliefs that p
activates) something about her general ability in a certain domain; call this
q. At time t3, she is debriefed: she is told that p is really false, that her per-
formance at the task did not have the property she was led to believe. What
should she do now? There is an initial feeling that she should go back and
revise both her belief in p and her belief in q. In section 10.5 I will raise
queries about what such revision could consist in. The question to be ad-
dressed now is: Is Jane obliged to go back in order to revise her belief in p
and her belief in q?
When Jane is debriefed, this activates her belief in p, and it is plausible to
hold that this belief should be revised. But the debriefing may not activate
her belief in q, or her beliefs in any other propositions that are partly trace-
able to her belief in p. So it may not occur to her to revise any of these other
beliefs. Is this a culpable omission? Does her failure to backtrack and re-
cover those other beliefs, and then revise them, render those beliefs
unjustified334
Notice that my basic theory of justifiedness does not imply that merely
because Jane's belief in q at t, is justified, it is also justified at t3. My frame-
work does not guarantee automatic continuity in justifiedness, even for a
belief that remains inert in LTM. Possibly, a right system of J-rules would
require a cognizer to backtrack wheneuer new beliefs are acquired, to make
selected revisions in antecedently entrenched beliefs. If these old beliefs are
retained, contrary to this requirement, they become unjustified.
But would a right rule system really make such a requirement? It should
first be stressed that a rule system cannot simply say: 'Revise all U-beliefs
that have been undermined by new evidence'. For this injunction does not
specify any basic cognitive processes, and only such processes are allowed in
the rules. What a rule may do is instruct the cognizer to search LTM for old
beliefs that could be undermined by the new evidence, or perhaps to search
LTM for old beliefs that were originally generated by beliefs one has now
abandoned. The rule could then say that any beliefs so identified should be
abandoned.
A rule formulated in one of these ways would certainly be admissible. But
one could conform with such a rule and not succeed in weeding out all (or
even many) of the beliefs that are the proper targets of the search. First,
even if one searches for old beliefs under the heading 'beliefs that the new
evidence undermines', the search might not yield much. Second, if one
searches under the heading 'beliefs that originated from beliefs now aban-
doned', then again the search might yield sparse results. For then one must
be guided by beliefs about the origins of one's beliefs; yet one might not
keep track of the origins very well, or might not activate these beliefs. (Re-
call my discussion to this effect in section 10.3.)There might be any number
of beliefs satisfying this description that one does not identify as such and
therefore does not identify as a belief that needs to be abandoned. In short,
a person could abide by rules of this sort, fail to abandon some of the in-
directly undermined beliefs, yet not be justificationally remiss.
Although not uncongenial, this is a result that may seem counterintuitive
until one considers the feasibilities of available cognitive operations. When
one considers feasibilities, it is an open question whether a right rule system
would oblige a cognizer continually to search for old beliefs in LTM that
might be weeded out in light of new evidence. Since one acquires massive
quantities of new evidence (new perceptual beliefs) at every waking mo-
ment, conformity to such a requirement would be very costly in cognitive
resources; so costly that it could well reduce the overall truth ratio of one's
belief corpus. And it would almost certainly put a crimp in other cognitive
tasks. Hence, it is a dubious normative precept.35
Notice that much more modest belief revision principles are possible
than the ones scouted above. A simple rule might say that if one activates an
old belief in q, and if one (actively) believes that this belief wholly stems
from now abandoned evidence, then one is required to abandon q. But this
is a far cry from the general maxim to search LTM continually for possibly
discredited beliefs. Here is another case, then, in which facts about the
structure of memory, and what it takes to activate materials in memory,
have direct bearing on normative precepts.
If this is right, what does belief revision, or change in belief, consist in?
Surely there is some phenomenon of belief revision, eSen if it is not literally
the erasure or replacement of an old belief. What it is, I suggest, is the 'su-
perimposition' of new content over, or alongside, old. New associative links
are established that are intended to be stronger than the old, and to have
contrary contents. For example, suppose you believe for a long time (as all
the books used to show) that brontosauruses had very small heads. Later
evidence is unearthed to show that they did not have such small heads after
all. What happens, when you learn this, is that a new associative path is
added, leading from the node for brontosaurus to the node for moderate-
sized head. Perhaps you also add a note, or correction, to the old small-hea-
dedness node, which says: 'No, no, this is wrong'. Later, if asked the size of
brontosauruses' heads, you might initially retrieve 'small head'. But then
you will retrieve the correction as well and inhibit any such verbal re-
sponse. When 'moderate-sized head' is also retrieved, that response is made
verbally. Still later, after the new material is frequently repeated, for exam-
ple, the 'small-head' node may not be activated at all; it might be totally
interfered with by the 'moderate-sized head' node.
All this conforms with the discussion by Rumelhart, Lindsay, and
Norman:
If material cannot be erased from LTM, then it will contain contradictions
and "temporary" structures. The intermediate incorrect steps one follows in
solving a complex problem (one that takes a considerable amount of time
and thought) will be remembered permanently, along with the final, correct
path. The retrieval processes must evolve to deal with these contradictions
and irrele~ancies?~
If the connection between 'brontosaurus' and 'small head' is still there, in
LTM, is it a belief? We better not say so, else we will have to ascribe in-
consistent beliefs to the cognizer, and this is not the sort of situation that
invites ascription of inconsistency. But this is not a serious problem. As long
as the connection between 'brontosaurus' and 'moderate-sized head' is
stronger than the old connection, and as long as the note, or correction,
pathway remains strong, only the latter materials deserve to be called be-
liefs. But don't we need a label for the LTM state that consists in the old,
uneradicated connection between 'brontosaurus' and 'small head'? Al-
though there is no good provision for this sort of state in ordinary language,
I have frequently stressed the inadequacies of ordinary-language taxono-
mies of the mental. I will call this sort of state a credal residue.
Credal residues are important to acknowledge because, like a phoenix,
they can be reborn-as beliefs. Suppose a recently added correction and
newly added content become weaker than the old associative pathway.
That is, upon first hearing the new evidence, you resolve to make the new
Memory 225
content be your real opinion, but unfortunately, your memory does not co-
operate; it does not retain the new content in sufficient strength. 'Bronto-
saurus' cues now activate the old node ('small head') without activating
either the correction ('No, no; this is wrong') or the new node ('moderate-
sized head'). The new belief is forgotten, and the credal residue resumes its
ascendancy. What can we now say about this credal residue? I think the
only reasonable thing to say is that it has reacquired its old status as a belief.
What is the justificational status of a credal residue that reemerges as a
belief? There is some temptation to call it unjustified. After all, its content
has been rejected on a previous occasion, and only the failure to retrieve
this rejection allows the content to reemerge in an ascendant form. This an-
swer would make sense if it were possible to expunge beliefs. A theory of
justification might say that striking such a belief from one's corpus is man-
datory; and if this is not done, then the preserved belief is unwarranted.
However, since we have denied the psychological feasibility of belief era-
sure, it is not clear how we could accommodate the view that the reascen-
dant credal residue is unjustified.
One way of trying to rationalize the label 'unjustified' within our frame-
work is to point out that the target case of belief recreation crucially in-
volves a process of forgetting (forgetting some other material). It might be
urged that the process of forgetting generally inhibits reliability, and there-
fore would not be authorized by a right system of J-rules.
How would it be argued that forgetting inhibits, or reduces, reliability?
There are really two forms, or manifestations, of forgetting. One manifesta-
tion is simply a failure to retrieve a once stored item. This manifestation
does not involve the production of a false A-belief; rather, it merely in-
volves the absence of an A-belief. The second manifestation of forgetting,
though, is erroneous retrieval. This arises from interference. Suppose a sub-
ject has studied a paired-associate list, including the item cat-43. Later he
studies another list that includes cat-82. Learning of the second list inter-
feres with the first. So when he is subsequently asked to recall items on the
first list, he erroneously retrieves cat-82. This process of interference is re-
sponsible for error.
The present case of credal residue is really a form of such interference
(only it features 'proactive' inhibition rather than 'retroactive' inhibition).
Earlier learning of the connection between 'brontosaurus' and 'small head'
interferes with the subsequent learning of the connection between 'bronto-
saurus' and 'moderate-sized head', so that the latter is no longer retrieved.
When it is recognized that such a process of interference works against reli-
ability, we can see that a rule permitting such a process might be a poor
candidate for inclusion in a right system of J-rules. Hence, beliefs formed by
such a process would not be justified.
All this is a way of supporting the claim that a reascendant credal residue
could be counted as an unjustified belief. But there is a problem with this
analysis. Although interference clearly can produce false A-beliefs, it can
also work to promote true beliefs. According to my earlier account, the way
in which the mind revises its old beliefs is by adding new informational
pathways and getting them to dominate the old pathways by means of in-
terference. In many of these cases the beliefs resulting (in part) from intet-
ference will be true. So it is not evident whether interference generally
promotes unreliability, or how seriously it breeds unreliability.
It is thus premature to expect that interference processes generally would
be excluded by right J-rule systems. Hence, it is questionable whether my
theory implies that reascendant credal residues are always unjustified. I do
not think it obvious that this is the intuitively correct judgment anyway.
The whole matter needs further reflection and investigation. At this junc-
ture I leave this as one among many problems for future explorations of pri-
mary epistemics to pursue.
chapter 11
Constraints on Representation
Language acquisition ,however, is not the main area that I shall explore, for
several reasons. First, the literature there is vast and technical, and would
take us far afield. Second, precisely because it is so intensively researched, it
may be instructive to choose examples from less familiar domains. Third,
language acquisition may have very specialized properties, which reduce its
usefulness for illustrative purposes. So I shall concentrate my efforts else-
where. Nonetheless, a few remarks on grammatical representation will be
ventured later in the chapter.
Virtually all theories of mental representation assume that the mind
somehow decomposes experience into parts, that it somehow partitions or
segments the world into convenient units. It is also assumed that the mind
engages in 'synthesizing' activities: constructing new units and structures
out of antecedently formed units. Kant, of course, laid great emphasis on
the mind's synthesizing activity, specifically the mind's propensity to struc-
ture experience in terms of material objects. In this section I shall focus on
some selected unitizing and constructional operations that cognitive psy-
chologists are investigating. In particular, I shall look at some hypotheses
Constraints on Representation 229
about .the ways in which human beings partition objects and events into
subunits, hypotheses which (in many cases) attribute a strong preference for
groups of representations that admit of hierarchical organization.
Let me begin with hierarchical structuring. The literature is replete with
descriptions of data structures, or cognitive units, that are organized hierar-
chically. Units are embedded in higher-level units, which are in turn em-
bedded in yet higher-level units.
Many examples of hierarchically structured representations are linked to
natural language.
(I) The phrase structure of sentences is hierarchical. A sentence is the
highest level unit, subordinating noun phrases and verb phrases, which in
turn subordinate nouns, verbs, and other grammatical units. While these
units are categories of word sequences, not of mental units, the universality
of these categories in public codes implicates a hierarchical organization of
associated mental representations.
(2) Another domain in which hierarchical structuring is widely postulated
is semantic memory. The popular semantic memory 'networks' stress a
class-inclusion structure in commonsense concepts.= Class-inclusion arrays
such as <physical object - living thing - animal - mammal - dog - collie>
are ubiquitous in natural languages. These taxonomical systems are proba-
bly manifestations of deep-seated representational mechanisms of the
human cognitive system.4
Hierarchicalization is by no means confined to representations associated
with natural language. It is found in the representation of all sorts of visual
and temporal materials.
(3)In the case of visual cognitive units smaller visual parts are embedded
in larger ones.'
(4) In the case of temporal strings shorter units are subordinated to longer
ones.6
Detailed discussion of these domains of representation will follow shortly.
Many other domains are also the subjects of postulated hierarchicalization.
(5)In story understanding plots are said to be represented hierarchically,
with smaller episodes embedded in larger ones.'
(6) The organization of behavior is achieved by hierarchical networks of
motives and submotives, plans and subplans.'
(7) Problem-solving tasks are structured by means of problems and
subproblems?
Turning now to some details, let us begin with visual materials. It is easy
to portray visual representation-in both perception and imagery-in
terms of multiple layers of embedding. Take the representation of a stand-
ing person. The whole is an elongated, ellipse-shaped object, oriented ver-
tically. At a finer level of resolution, the parts of the body are delineated: a
head, a torso, two arms, and two legs. Each of these parts has global proper-
230 Constraints on Representation
ties. But each can also be considered a whole with further parts. The head,
for example, has eyes, ears, a mouth, and a nose.
An example from A. Stevens and P. Coupe testifies to the storage of imag-
istic information by hierarchically layered parts.10When asked which is far-
ther west, Reno or San Diego, the typical subject mistakenly chooses San
Diego. The natural explanation is that San Diego and Reno are encoded as
subunits of California and Nevada respectively, and California is pictured
as west of Nevada on one's mental map of the continent.
Several of the experimental studies I shall review are not only concerned
with hierarchical organization; they also postulate preferences for patterns
that accord with Gestalt principles. So before turning to these studies, let
me mention and illustrate these well-known Gestalt principles, originally
postulated for perception. Some of the principles are illustrated in Figure
11.1. In part A we perceive four pairs of lines rather than eight separate
lines. This illustrates the principle of proximity: elements close together
tend to organize into units. Part B illustrates the same principle. Five col-
umns of dots are seen rather than five rows, because the dots in a column are
closer together than the dots in a row. Part C illustrates the principle of
similarity: objects that look alike tend to be grouped together. In C we tend
to see five rows of alternating 0's and X's, even though the rows are spaced
similar to those in B. Part D illustrates the principle of good continuation.
We perceive part D as two lines, one from x to y and the other from w to z,
although there is no reason why this could not represent another pair of
lines, one from x to z and the other from y to w. But the line from x to y
displays better continuation than the line from x to z, which has a sharp
turn. Part E illustrates the principles of closure and good form.We see the
drawing as one circle occluding another circle, although the occluded ob-
ject could have many other possible shapes.
It is time to consider some representative studies of hierarchical organi-
zation and partition of visual materials. G. H. Bower and A. L. Glass studied
subjects' styles of partitioning unfamiliar line drawings." In principle, a
given line drawing could be mentally dissected into all kinds of parts. But in
fact, they proposed, only certain subpatterns are 'naturally7 represented as
parts. Bower and Glass predicted that a natural part should serve as a strong
retrieval cue for the original, whole pattern; whereas an unnatural frag-
ment, even one of equal size, should lead to poorer recall of the original
whole. After their subjects studied a set of original drawings, they were
tested for recall on fragments which the experimenters had categorized as
'good', 'mediocre', or 'bad' (misleading).The experimental results were that
good cues had about five times more retrieval power than bad cues.
Bower and Glass chose the good, bad, and mediocre fragments by appli-
cation of Gestalt rules of common direction and minimal angle. As we have
seen, the rule of common direction, or good continuation, states that line
Constraints on Representation 231
0 0 0 0 0
0 0 0 0 0
0 0 0 0 0
0 0 0 0 0
0 0 0 0 0
X X X X X
X X X X X
Figure 11.1
Figure 11.2
into the parts shown in D. Using this theme, Bower and Glass constructed
picture fragments of C such as E, F, and G. E should be a good cue, F a
mediocre cue, and G a bad cue. Experimental results confirmed this
prediction.
A similar set of results was reported by Stephen palmer.12 Like Bower
and Glass, Palmer suggested that only certain subparts are 'natural', or
'good', decompositions of a larger figure. He devised a measure of goodness
based on Gestalt principles of grouping: proximity, closedness, connected-
ness, continuity, and so forth. Figure 11.3displays some of the whole figures
and respective parts he used, where the parts are rated for degrees of good-
ness: high-(H), medium high (MH), medium (M), medium low (ML), and low
(L). Palmer then conducted a series of experiments to see whether observed
results conformed with the part ratings his scheme predicted. One experi-
Constraints on Representation 233
.
ment asked subjects to divide figures into their natural parts. A second task
was to rate goodness of identified parts within figures. A third experiment
was a reaction-time experiment. A fourth experiment involved a 'mental
synthesis' task, in which subjects were instructed to construct a figure men-
tally from parts. Palmer found substantial evidence for the kinds of selective
organization in perception and imagery that he had postulated.
Although one might suppose these styles of representation are peculiar to
vision, quite similar styles are found in the representation of sequences of
events. With an eye toward studying hierarchical representation of serial
patterns, Frank Restle devised a formalism congenial to such representa-
tion.13Let E = 112 3 4 5 61 be a set of elementary events, where 'l', '2', and
so on, comprise the alphabet for representing these elementary events. Let
X be a sequence of events from this set; for example, X = <1 2>. A set of
operations on this set is then introduced: repeat (R) of X , mirror image ( M ) of
X, and transposition +1 (T) of X. Each operation generates a new sequence
from a given sequence and then concatenates the old and the new. For ex-
ample, R<1 2> = <1 2 1 2>; relative to the alphabet specified above,
M<2 3 4> = <2 3 4 5 4 3>; and T<2 3> = (2 3 3 4>. In this fashion a
lengthy and complex series can be represented economically by multiple
embedding. For example, T<1> = <1 2>; R<T<l>> = <1 2 1 2>;
T<R<T<l>>> = <12 1 2 2 3 2 3>; and M<T<R<T<l>>>> = <12 1 2
2 3 2 3 6 5 6 5 5 4 5 4>. Using sequences constructed in this way, Restle and
Brown found evidence that subjects mentally encode temporal patterns in
accord with a hierarchical structure indicated by the recursive format.14
A number of writers have recently proposed that musical cognition re-
veals hierarchically organized representations. F. Lerdahl and R. Jacken-
doff, for instance, hold that at least four components of the experienced
listener's musical intuitions are hierarchical.'' 'Grouping structure' ex-
presses a hierarchical segmentation of a piece into motives, phrases, and
Whole
Figures H MH M ML L
Figure 11.3
234 Constraints on Representation
Figure 1 1.4
sections. 'Metrical structure' expresses the intuition that the events of the
piece are related to a regular alternation of strong and weak beats at a num-
ber of hierarchical levels. And so on.
Diana Deutsch and John Feroe also suggest that people represent pitch
sequences in tonal music in abstract form.'' To illustrate their argument,
consider the pitch sequence shown in Figure 11.4. A musically sophisticated
performer (or listener) would represent this sequence at two or more levels
of description. At the lowest level there is a representation of the entire
presented sequence. At the next (higher) level the representation would
feature an arpeggio that ascends through the C major triad (C-E-G-C). This
more abstract representation registers the fact that the entire sequence can
be viewed as the four notes of this triad each preceded by a neighbor em-
bellishment. At this higher level of representation, the embellishments are
deleted. The resulting hierarchical structure may be presented in tree form
as in Figure 11.5.Drawing on ideas of earlier authors, Deutsch and Feroe
propose that tonal music is represented with the help,of several different
(mental) 'alphabets': major scale alphabets, minor scale alphabets, the chro-
matic alphabet, and the triad alphabets. Different alphabets can be used at
C- E-G- C
Figure 11.6
continuity is the line in the crook of the elbow. The rule does not define a
unique partition of the block. As predicted, there are three plausible ways
to cut the block into parts, all relying on the contour defined by the parti-
tioning rule.
The foregoing rule does not help with entirely smooth surfaces. Hoffman
therefore develops the following partitioning rule: divide a surface into
parts at loci of negative minima of each principal curvature along its asso-
ciated family of lines of curvature. As applied to the shape in Figure 11.7,
this yields the indicated partitioning contours (dashed lines).
Figure 11.7
Constraints on Representation 237
Let me close this section with a slightly different topic: not constraints on
the representation of parts, but constraints on the representation of motion.
Shimon Ullman has studied principles that underlie the visual system's in-
terpretation of a sequence of two-dimensional stimuli (for example, dots of
light on a movie screen).21How does a subject 'see' such a succession of
stimuli? One issue is how the subject 'matches' the parts of one frame in a
motion picture with parts of a succeeding frame. Which parts of the subse-
quent frame are viewed as identical to parts of a preceding frame? And
does the subject interpret these event sequences as representing a single
three-dimensional object in motion or in change? If so, what governs the
choice of a three-dimensional interpretation?
Ullman claims that the visual system displays at least three innate prefer-
ences: (1) In interpreting which dot elements match in successive arrays,
there is a preference for seeing motion to nearest neighbors. (2) In deciding
what matching to impute, motion along straight lines is preferred. (Notice
that these first two principles echo Gestalt principles of proximity in space
and common direction.) (3) Any set of elements undergoing two-dimen-
sional transformation that has a unique interpretation as a rigid body mov-
ing in space should be so interpreted. This third principle is a version of the
preference for seeing visual arrays as solid 'bodies', a theme that many phi-
losophers have stressed.22
Earlier work had shown how humans can retrieve structure from motion.
For instance, if a transparent beach ball with tiny light bulbs mounted ran-
domly on its surface is set spinning in a dark room, one immediately per-
ceives the correct spherical layout of the lights. When the spinning stops,
so does the perception of the spherical array. How does one see the cor-
rect three-dimensional array when infinitely many three-dimensional struc-
tures are consistent with the moving two-dimensional retinal projection?
Ullman showed mathematically that if the visual system exploits the
laws of projection, and 'assumes' that the world contains rigid objects,
then in principle a correct interpretation can be obtained. In particular,
three views of four noncoplanar light bulbs are enough to solve the
problem.
However, in cases of 'biological motion' studied by Gunnar Johansson,
one does not have four noncoplanar points.23When small light bulbs are at-
tached to the major joints of a walking person, there are at best only pairs of
rigidly connected points, such as the ankle and the knee, or the knee and the
hip. Nonetheless, the visual system does recognize a walking person from
such information. Hoffman and Bruce Flinchbaugh suggested that if the vis-
ual system makes a further assumption-the 'planarity' assumption--dis-
plays of gait of this sort can be correctly interpreted from three snapshots of
just two points.24Thus, the visual system appears to have very specific con-
straintjprobably innate computational mechanisms-for constructing
shape representations.
238 Constraints on Representation
state 2 (below); or would they prefer to pour from A into B, thereby yield-
ing state 3?
Twice as many subjects preferred the move to state 3 as preferred the move
to state 2. Notice that state 3 is quite similar to the goal, since the goal is to
have 4 cups in both A and B, and state 3 has 3 cups in A and 5 cups in B. By
contrast, state 2 has no cups of water in B. This illustrates a tendency to
move to states that are as similar as possible to the goal state.
Recent work on scientific theory construction, both by cognitive scien-
tists and by historians of science, also explores the role of analogy, or simi-
larity. For example, several authors trace the development of Darwin's
theory to his finding an analogy between Malthus's theory of human popu-
lation growth and the growth of species.32
Analogy is often treated as a separate representation-forming device, and
we could certainly place it alongside the other operations tentatively listed
earlier. In other words, where it is assumed that one representation (be it a
concept, a theory, an idea for a problem solution, or what have you) is ob-
tained from another by some sort of operation, one candidate operation
might be that of analogy. This might be regarded as a primitiue, irreducible
operation. But I want to advance the proposal that the analogy operation is
analyzable in terms of other operations.
One way of trying to flesh out this proposal would be to treat analogical
representation formation as a substitution operation. Take Bohr's model of
the atom, for example. This representation might be derived 'by analogy'
from a representation of the solar system. Wherein does the analogical con-
struction consist? The suggestion would be that it consists in substituting (a)
a nucleus for the sun, (b) electrons for the planets, and (c) nuclear forces for
the gravitational force.
Is this approach plausible? A criterion of adequacy for any such theory
would seem to be this: if representation ROis obtained from R 'by analogy',
then R* (or, rather, its putative referent) should be judged quite similar, or
analogous, to R, at least by the cognizer in question. But not every (mental)
substitution yields a representation of a subjectively similar object. If R is a
representation of a swan, and ROis obtained by first deleting everything in
R except (a representation of) the eyes and then substituting (a representa-
tion of) a tiger, the new representation ROis hardly very similar to, or analo-
gous to, the original.
What further constraints can be placed, then, on substitutions to yield a
match with intuitive judgments of resemblance or similarity?One possibil-
ity is to employ the notion of 'natural parts'. The greater the naturalness of
242 Constraints on Representation
the parts preserved under replacement, the greater the subjective similarity
of the resulting representation.
I doubt, however, that judged analogies or similarities can be adequately
recovered by the substitution operation, even conjoined with this additional
device. Other high-ranking or 'preferred' operations, in addition to substi-
tution, might have to be invoked. For example, assume that reversal, or in-
version, relative to some natural axis or alphabet is a high-ranking
transformation. Then an object obtained from another object by such a re-
versal would be judged rather similar to it.
With this in mind, we might try to generalize the approach along the fol-
lowing lines:
The contents of representations R and R' are similar-receive a high simi-
. larity rating-if and only if the content of one is obtainable from the content
of the other by application of a high-ranking cognitive transformation (or
combination of transformations).
Lance Rips points out, in a personal communication, that one source of evi-
dence for this account of similarity is that different transformations yield
representations similar to the original one but not themselves similar. For
example, the concept 'cold' might be considered similar to 'hot' (from one
perspective), via an 'opposites' transformation. Also, 'warm' is similar to
'hot', via a 'next-to' transformation. But 'warm' and 'cold' are not subjec-
tively similar.33
Notice the analogy between this idea and one employed by topologists.
In topology two figures are said to be topologically equivalent (homeomor-
phic) just in case one can be obtained from the other by some sort of topo-
logical operation.
If this line of analysis proved promising, it might be exploited by turning
it on its head. (I owe this suggestion, and the example that follows, to
Charles Chastain.) Instead of studying similarity in terms of transforma-
tions, one could study transformations in terms of judged similarities. For
example, given perceived similarities among faces, one might try to identify
high-ranking transformations by seeing which would account for those simi-
larity judgments.
These are some tentative thoughts about analogy. Their relevance stems
from my interest in representation-forming operations. If analogy is related
to such operations in the manner proposed, we gain new avenues for getting
at the repertoire of operations. Furthermore, given the apparent ubiquity of
similarity, or analogy, in cognitive phenomena, identification of this reper-
toire-and the most preferred members of it-is quite important for cogni-
tive science.
Constraints on Representation 243
not 'fit'. To execute such a plan, however, a cognizer can partition the plan
into subplans. If the relevant subplans are activated at suitable junctures,
the entire plan can be executed piece by piece. But this requires that the
larger, superordinate plan be suitably linked in memory to appropriate
subplans. This is achieved by hierarchical representation of plan and sub-
plans, and the retrieval of microunits from macrounits in accordance with
this hierarchical organization.
Musical performance is a vivid case in point. The piece to be performed
is represented in LTM at various levels, encompassing different sized seg-
ments of the work. As the musician proceeds through the performance, rep-
resentations appropriate to each passage are accessed. These are detailed
enough for guidance of behavior but small enough for accurate retention in
working memory. Deutsch and Feroe underline this point as follows:
Several investigators have shown that for serial recall of a string of items,
performance levels are optimal when such a string is grouped by the ob-
server into chunks of three or four items each . . . When segments of tonal
music are notated on the present [hierarchical]system, there emerges a very
high proportion of chunks of three or four items each . . . As pitch sequences
become more elaborate, they are represented as on a larger number of hier-
archical levels, but the basic chunk size does not appear to vary with changes
in sequence complexity. This chunking feature therefore serves to reduce
memory load.37
Thus far I have mentioned only virtues of the partitioning and hierarchi-
calization mechanisms, But one might ask, aren't there any vices? For ex-
ample, doesn't the presence of restricted styles of decomposing wholes into
parts reduce power and flexibility? This issue may be discussed in terms of
Fodor's notion of "epistemic boundedness": the inability of a cognitive sys-
tem even to represent-and a fortiori its inability to believe--certain
truths. If a cognitive system has a fixed set of representation-forming opera-
tions, there must be representations it is incapable of generating. This ap-
pears to imply a limitation on power.
Actually this is not so, at least on my definition of power. On my defini-
tion a system is completely powerful if it can answer (correctly) every gues-
ti072 it can ask. This does not imply that it can answer every question. In
other words, complete power does not entail unboundedness. Therefore,
boundedness does not entail the absence of complete power.
Setting this point aside, how serious a defect is the absence of total
power? Not much. When talking about finite minds, it is not an alarming
defect that omniscience is beyond reach. That's just the nature of the beast.
The more important question is: Is it capable of getting true answers to the
sorts of questions it actually wants to answer, to the sorts of problems it ac-
tually wants to solve? On this point there is no glaring deficiency in the re-
presentational operations under discussion.
246 Constraints on Representation
Notice, in this connection, that the Gestalt principles might be best in-
terpreted as preferred operations or default procedures, rather than manda-
tory operations. Take the wholes and parts that Palmer used in some of his
experiments (shown in Figure 11.3).The parts are rated for goodness as
either high, medium high, medium, medium low, or low. Subjects gave rat-
ings to the parts that accorded with Palmer's Gestalt-inspired metrics. But
subjects can 'see' each of them as a part of the whole, even the lowest rank-
ing parts. It is just harder to see them as parts. So people are not incapable
of parsing the whole into the inferior segments; those parsings are just less
'natural' or obvious.
Still, isn't it a vice of the system that it has preferences for certain pars-
ings? Doesn't it inevitably betoken a distortion in the system's noetic rela-
tion with the world? Doesn't it mean that the mind can't mirror the world,
and doesn't this necessarily breed error as well as ignorance?
Representational preferences do preclude mirroring, but mirroring is not
required for truth. You do not have to get every truth to get some truths.
Partitioning and synthesizing preferences may indeed imply that some
'correct' parts and wholes do not get constructed. But the preferred parts
and wholes may be perfectly accurate as far as they go.
Sometimes, however, preferences do produce error, not just partial igno-
rance. Gestalt preferences induce the visual system to complete, or fill in,
the stimuli it encounters, thereby achieving 'good continuation', even when
the real figure has discontinuities or gaps. This is indeed distortion. Inter-
pretation of two-dimensional arrays as three-dimensional solid objects can
also lead to error-when the real stimuli are only dots of light.
But how serious are these error possibilities? As long as the organism's
environment is full of solid objects, which generally do instantiate good
continuation, these error prospects will not be systematic. In terms of the
actual environment of human beings, the preferences seem to be highly re-
liable. Indeed, the mathematical proofs by Ullman, Hoffman, and their col-
leagues, mentioned in section 11.2, show that under certain conditions like
rigidity and planarity, the hypothesized visual principles are perfectly ac-
curate. (To be more precise, the probability of error has a measure of O-
which does not imply the impossibility of error.)
Even if highly constrained representation-forming devices can lead to
error, there may be compensating gains in speed. This is worth illustrating
in connection with language learning. In acquiring a first language, the
learner may be thought of as generating grammatical hypotheses on the
basis of linguistic 'texts' (discourses) to which he or she has been exposed.
Language-learning theorists in the Chomskyan tradition conjecture that
human dispositions of hypothesis generation-conceptualized as functions
from texts to hypothesized grammar-are quite restricted. Such disposi-
tions could not lead to correct identification of all possible languages.38
Constraints on Representation 24 7
1 1.5. Originality
Although it is scarcely touched by philosophical epistemology, 'originality'
is a widespread term of evaluation in intellectual affairs. It is a dimension
commonly invoked in the evaluation of both intellectual product-books,
articles, ideas, and hypotheseeand intellectual producers. Since the
progress of knowledge depends on the production of new knowledge, and
since this characteristically depends on original ideas (not merely new evi-
dence for old ideas), originality is a proper subject of epistemological study.
248 Constraints on Representation
Of course, originality is also present in the sphere of the arts, which may not
fall under the rubric 'intellectual', and may therefore fall outside epistemol-
ogy. But that needn't perturb us.
What do we mean when we call an idea 'original'? One broad meaning is
that it is an idea nobody else has had before; at any rate, it is an idea that
the thinker did not get from anyone else. But there is another, narrower
sense of the term. Within the class of ideas that are not directly borrowed
from others and are not mere duplicates, some strike us as more inventive,
clever, surprising, imaginative, or creative than others. These we deem
'original'. Other ideas, which are equally nonduplicates, seem pedestrian,
obvious, and trite. The term 'original' is withheld from these. But how do
we distinguish the inventive ideas from the pedestrian, the imaginative
from the hackneyed?
Apparently we have some tacit metric by which we measure originality.
Some ideas seem more 'distant' from socially available ideas, harder to think
up on one's own. Other ideas, even if they are original in the broad sense,
seem easy to think up; they seem 'closer' to what has already been formu-
lated. But what underlies this metric? What makes some ideas seem close to
their predecessors, and easier to think up, and others more distant and diffi-
cult? The answer is related to human representation-forming operations.
Suppose, as I have suggested, that human beings have a fairly delimited
set of representation-forming operations, and a (substantially)fixed ranking
of these operations. In addition, assume that in any scientific or cultural mi-
lieu the stock of ideas antecedently available to cognizers is essentially the
same. Then all new ideas must result from transformations of the same ini-
tial stock of ideas. If this is correct, one should expect certain new ideas to
occur independently to many workers in a given field. Members of the dis-
cipline all have roughly the same stock of initial information and previously
tried ideas; and they share a set of problems they wish to solve. So when the
highest-ranking-or most 'natural'-operations are applied to their initial
stock of ideas, many people will come up with the same (or very similar)
new thoughts. But these will be the most obvious thoughts, the most
straightforward extensions of preexisting ideas.
The sociologist of science Robert Merton has remarked on the prevalence
of multiple discoveries in science: the same discovery being made at
roughly the same time by several independent researcher^.^' In fact, I
would add, the number of historically signijicant multiples surely underes-
timates the prevalence of the phenomenon. Scientists and scholars are con-
stantly coming up with similar ideas; but many of these do not get
published or do not get published by everyone who thinks of them. People's
minds &n in very much the same channels.
How, then, are truly original ideas created? And why do certain individu-
als seem particularly adept at such inventiveness?Four possibilities come to
Constraints on Representation 249
11.6. Justification
Originality and creativity are nonstandard topics for traditional epistemol-
ogy. Other, more traditional, dimensions of appraisal to which the study of
representation-forming operations is also germane include justification,
which can be illuminated by the psychology of representation formation
slightly differently from the reliability considerations already adduced.
It is widely suggested that a person is justified in believing hypothesis p if
p would be a good explanation of certain observed phenomena and there is
250 Constraints on Representation
Internal Codes
Figure 12.1
256 Internal Codes
Since I regard the multiple code theory as extremely natural and attrac-
tive, and since I (therefore) wish to explore the epistemological significance
of multiple codes, I want to put in some good words for the multiple code
hypothesis. In other words, I do not want to merely assume its correctness;
rather, I want to lend support to it. There are different ways of formulating
the central hypothesis, however, and I do not mean to endorse the hypoth-
esis in all--or even the most popular--of its renderings. Much of the debate
has centered on topics I regard as incidental to epistemological purposes. I
shall endorse only a modest variant of the multiple code approach. To ex-
plicate this variant, and set it against the backdrop of the recent contro-
versy, is the task of the next section.
(sense-neutral) thinking presumably does not share that code or those re-
sources. Nor do auditory images. Hence, if the perception-similitude thesis
is correct, there are distinct codes or code-systems. Thus the descriptional-
ist, to the extent that he wants to cleave to a single code, should not be re-
ceptive to the perception-similitude thesis.
For my purposes, I do not care to endorse or defend the pictorial formu-
lation of the imagist position, although this has received the lion's share of
attention. It suffices for my purposes to endorse and defend the perception-
similitude thesis-or, for that matter, any other thesis that guarantees mul-
tiple, infmmationally distinct code systems. However, let us see just what
imagists have been claiming. We will find considerable ambiguity in their
writings, but no matter. It is open to us to select whichever aspects of their
position seem most secure, and most significant from an epistemological
point of view.
The most sustained and detailed defender of the imagist position is Ste-
phen Kosslyn. Many of his writings contain phrases and models that
strongly insinuate an acceptance of the pictorialist thesis. He has intro-
duced the idea that visual images are like displays on a cathode ray tube
(CRT), and he and his colleagues characterize images as "spatial displays"
in active memory.14 In his book I m g e and Mind Kosslyn talks explicitly
about images as occurring in an "internal spatial medium,"15 and he calls
his position "quasi-pictorialist."ls
Yet Kosslyn and colleagues also say that the CRT model is just that-a
model-and they do not mean that images are exactly like a display on a
CRT.In Image and Mind Kosslyn says that picture talk in connection with
imagery is just a metaphor:
Whereas the earlier phrases I quoted strongly suggest the pictorialist the-
sis, these passages--especially the italicized clauses-support only the per-
258 Internal Codes
cent statement, "The Imagery Debate: Analog Media versus Tacit Knowl-
edge," Pylyshyn appears to accept a large chunk of the perception-simili-
tude view:
It is my view that there is only one empirical hypothesis responsible for the
predictive success of the whole range of imagistic models . . . When people
inuagine a scene or an event, what goes on in their minds is in many ways
similar to what goes on when they obseroe the corresponding event actually
happening . . . The claim that imagery is (in some ways) like perception has
predictive value because it enables us to predict that, say, it will take longer
to mentally scan longer distances, to report the visual characteristics of
smaller imagined objects, to rotate images through larger angles, . . . and so
However, Pylyshyn indicates that this (the italicized statement) is not all
that Kosslyn and colleagues maintain. They also claim that imagery in-
volves use of a distinctive form of representation, or a distinctive medium or
mechanism. And he is unpersuaded of the alleged characteristics of such a
form of representation.
Since even the chief antagonist of the imagist position seems to accept
the perception-similitude thesis, perhaps I need not offer an extended de-
fense of it. However, let me cite some of the experimental evidence that
supports a close connection between vision and visual imagery, and sup-
ports the contention that they both utilize some of the same specialized
code system is shared by visual perception and visual imagery, a code not
deployed by nonsensory thought or by perception and imagery in other mo-
dalities. Of course, at some deep level, all internal representations, or code
systems, have a common basis, namely, electrochemical activity of neurons.
But there can be distinct code systems at a higher level, even if there is a
common bottom-level realization. (The same piece of computer hardware
can run distinct programs and implement many programming languages.)
Moreover, for the purposes of (primary)epistemology, it is sufficiently note-
worthy that, at some level, there are code systems with informationally dis-
tinctive properties. That is the idea I shall be elaborating in the rest of the
chapter.
The next questions we need to ask are: What are the informationally dis-
tinctive properties of sense-linked code systems, and what are the episte-
mological ramifications of these properties? How does the visual code
system, for example, differ from the nonsensory code system? Here the
theories developed by Kosslyn and others should prove helpful.
The Ancient Greek Simonides, the story goes, was performing his role as
bard and orator at a banquet when he was called outside to receive a mes-
sage. At just that time, the ceiling collapsed, mangling all the dinner guests
beyond recognition. When asked to recount who the hapless victims were,
Simonides discovered that he could mentally picture the table and "walk
around it," seeing all those seated with his mind's eye. Simonides reportedly
found it quite easy to name all those in attendance by employing this tech-
nique, and went on to develop it as a general way of improving one's mem-
ory. The product of Simonides' and his methodological descendants' efforts is
usually called the "method of loci." The method of loci involves first select-
ing a series of familiar places, among which one can imagine walking in se-
quence Then, when one is trying to memorize a list, one imagines walking
from place to place and "leaving" an image of each item in the list at each
successive location. When later recalling the list, one then walks by the loci,
and "sees" what is present at each one.32
Internal Codes 265
Let us turn now to uses of imagery in the visual buffer alone, without (sig-
nificant) reliance on LTM. In this category are the "mental rotation" tasks
explored by Shepard and colleagues. Given the task of deciding whether
certain line drawings depict congruent three-dimensional objects (see Fig-
ure 12.1), virtually all subjects imagine the rotation of one of the depicted
objects, compare the two imagined objects, and then offer an answer. Sig-
nificantly, subjects' performances on this task are extremely accurate. In
Shepard's initial experiments nearly 97 percent of the 12,800 recorded re-
sponses of the eight sabjects were correct, with error rates for individual
subjects ranging from 0.5 percent to 5.7 percent.33
Shepard also emphasizes the significant record of image-assisted creativ-
ity in the annals of science.% The theory of relativity is said to have had its
inception when the young Albert Einstein performed his epochal Gedanken
experiment of imaging himself traveling along with a wave of light at 186,-
000 miles per second. James Watson's account of how he and Francis Crick
deciphered the double-helical structure of DNA contains still more specific
references to mental performances involving imagery, such as his sudden,
crucial realization that under an appropriate rotation in space an adenine-
thymine pair had the same shape as a guanine-cytosine pair. And the con-
joint inventions of the self-starting, reversible induction motor and the
polyphase system of electrical distribution burst upon the Hungarian-born
inventor Nicola Tesla in imagistic form. Tesla had an almost hallucinatory
vision of the rotating magnetic field that would be induced by a circle of
electromagnets, each energized by the same alternating current but succes-
sively shifted in phase.
These major creative feats are not necessary to establish the significance
of image-assisted thinking in science. It is the common practice of average
scientists both to draw diagrams and to think in terms of imaged diagrams.
Most scientists would be greatly handicapped by any stricture against dia-
grammatic imagery in their scientific thought processes. Outside science,
examples of image-assisted tasks are also plentiful as blackberries. One such .
example is chess. Chess players typically find it convenient to picture the
chessboard configuration that would be generated by a move sequence they
are contemplating. This helps them determine favorable or unfavorable
prospects of such a sequence.
In all of these cases, it is quite clear that imaging helps. It enables one to
get information one could not get otherwise, either at all, or with equal fa-
cility. Not all uses of imagery, though, are helpful or accurate. A familiar
case is the untrustworthiness of imagery as a test of possibility. The fact that
one cannot imagine, that is, picture, a chiliagon does not demonstrate the
impossibility of a chiliagon.
Other cases of unhelpful, or merely pseudohelpful, imaging are presented
by ~ ~ l ~ s hImagine
y n . ~holding in your two hands, and then simultaneously
266 Internal Codes
the proof does not suffice for the believer's being justified. He must at least
believe that such a proof exists. But mere belief is not enough either. He
must believe this justifiably. Such justification might come from someone
else's authoritative say-so, in which case the relevant mental processes are
those that infer the existence of the proof from this testimony plus other
background beliefs. Suppose, however, that his justified belief arises from
firsthand study of the proof. Then some mental operations must be em-
ployed that make him justified in believing that this is a proof of the propo-
sition in question.
The typical case is one where the cognizer reads the proof, and this ob-
viously involves perception. So perceptual processes are among those typi-
cally involved in the acquisition of mathematical knowledge. (This needn't
undermine the familiar claim that mathematics is a priori, since this notion
might be defined in terms of the possibility of nonperceptual knowledge.)
What other mental processes would typically be involved? Clearly, it is im-
portant to remember previous steps of the proof, and to remember the in-
ference rules. These memories might involve visually coded representations
of those steps and rules. One would also have to recognize, of each step, that
it is indeed an application of relevant rules to previous theorems. This sort
of recognition could also involve imagistic representations. In short, a suit-
able-that is, justification-conferring-set of mental operations might well
include imagistic thinking (at least for some people on some occasions).
Nothing in the notion of 'following', or 'understanding', a proof excludes
imagistic representation from a role in such a process.
It is worth remarking in this connection that visually imaged spatial lay-
outs may be used as analogies for an indefinite assortment of problems, in-
cluding logical problems. That such imagistic strategies are quite common
has been hypothesized by Janellen Huttenlocher, among others.* Hutten-
locher presented subjects with tasks of the following sort:
Arthur is taller than Bill.
Tom is shorter than Bill.
Is Arthur taller than Tom?
The subjects in Huttenlocher's experiments describe themselves as pro-
ceeding as follows.
First, they say, they arrange the two items given them in the first premise.
Sometimes Ss [subjects] describe this array as horizontal and sometimes as
vertical. For vertical arrays, Ss claim that they start "building" at the top
and work towards the bottom. For horizontal arrays they claim to start at the
left and work towards the right . . . The goals are to identify the third item
and tsdetermine its position with respect to the other two items.41
Huttenlocher's hypothesis that subjects solve these sorts of logic problems
with spatial imagery is controversial. (It has been challenged, for example,
Internal Codes 271
dicts, and subjects exposed to vivid defense testimony shifted toward not-
guilty verdicts. (However, delayed judgment did not show any effect of the
vividness manipulation for a bad character defendant.)
In an experiment by R. Hamill, T. D. Wilson, and R. E. Nisbett, subjects
were given a description of a single welfare case." The description painted
a vivid picture of social pathology: an obese and irresponsible woman who
had been on welfare for many years, lived with a succession of 'husbands', in
a home full of dirty and delapidated plastic furniture, with cockroaches
walking about in the daylight, and so on. In a second set of conditions this
description was omitted, and subjects were given statistics showing that the
median stay on welfare was two years, and that only 10 percent of recipi-
ents remained on welfare rolls for four years or longer. The pallid statistical
information, however, had no effect on subjects' opinions about welfare re-
cipients; but the vivid description of one welfare family prompted subjects
to express more unfavorable attitudes toward recipients. So a questionably
informative case history had greater effect on inferences than dull statistics.
Anecdotal evidence supporting the vividness-impact hypothesis includes
the following two items (originally given by Nisbett, E. Borgida, R. Cran-
dall, and H. ~eed.)'l (1) Mastectomies performed on Mrs. Ford and Mrs.
Rockefeller in the fall of 1974 produced a flood of visits to cancer detection
clinics, whereas widely disseminated statistics about the lifetime risk for
breast cancer had never produced a comparable impact. (2) While all physi-
cians are aware of the statistical evidence linking cigarette smoking to
cancer, only physicians who diagnose and treat lung cancer victims are
quite unlikely to smoke; and radiologists have the very lowest rate of smok-
ing. So informational vividness seems to influence even sophisticated people
who have been exposed to the most probative data.
Why does vividness have this (putatively) large impact? One factor cited
by Nisbett and Ross is that vivid information is more likely to be stored and
retrieved than pallid information.
Nisbett and Ross cite other possible explanatory factors, beyond good re-
tention. For example, vivid materials may convey more information; or
they may recruit connected information by means of memory links; or they
may remain longer in (conscious)thought.
Before proceeding to reflect on these possible causes of the impact of
sense-linked materials, I should note that many of Nisbett and Ross's con-
jectures about vividness have been challenged on empirical grounds.
Doubts about their empirical support are expressed by Shelley Taylor and
Suzanne Thompson in a review of the l i t e r a t ~ r e Eight
.~ studies, they re-
port, have directly tested the concreteness hypothesis, manipulating vivid-
ness through concrete and colorful language. Seven found no differences in
attitudes as a function of the appeal. Four studies employed photographs as
manipulations of vividness and found no differences between the pic-
ture/no picture conditions. Thirteen studies compared the impact of video-
taped information with identical or similar information presented through
other media (such as oral or written presentations). Of these, six found no
differences in persuasive impact of various media; five found video presen-
tations to be more effective, but only under limited conditions; one found
print to have more impact than either television or audio presentations; and
one found radio to be a more persuasive medium than television.
Concerning imagery, Taylor and Thompson indicate that actual studies
contradict expectations that imageable vivid material will influence judg-
ments more. The one area where the so-called vividness effect is supported
by empirical studies concerns case histories. Case history information does
seem to have a persuasive impact on judgments, when contrasted with
base-rate or abstract information. However, the extent to which this finding
speaks to the vividness effect is questionable. One possibility, they say, is
that subjects are underutilizing base-rate or other statistical information,
which they do not understand well, rather than overusing case history in-
formation. Perhaps subjects can readily discern the causal relevance of case
history information to the judgments they make, but are less able to see the
causal relevance when information is presented statistically. However this
empirical debate is settled, these last points are extremely important for the
nornuztive assessment of the use of imaged materials, to which I now
proceed.
Nisbett and Ross entitle their chapter on vividness "Assigning Weights to
Data: The 'Vividness' Criterion." The implication is that cognizers fre-
quently have both vivid and pallid data equally before their minds, that
they understand them equally, but just assign greater weight to the former.
However, this is an implausible description even of the cases they describe,
and an inaccurate account of even (some of) their own diagnoses. In several
of the studies they cite, subjects are given either vivid or pallid information,
but not both. Thus, they are not preferring one kind of information over the
other. They are just operating separately either with the vivid information
276 Internal Codes
or with the pallid. Now it may well be (as Taylor and Thompson suggest)
that the subjects tend to underutilize the pallid, statistical information. But
this does not show that they are doing anything wrong in deploying their
vivid, concrete, sensorily imageable information. Consider the women who,
upon learning of the well-publicized mastectomies of Mrs. Ford and Mrs.
Rockefeller, visit cancer detection clinics. Surely this is not an obvious cog-
nitive error on their part. Maybe their earlier failure to take similar action
upon reading statistical data on breast cancer was irrational. So why is posi-
tive responsiveness 'to the case histories of Mrs. Ford and Mrs. Rockefeller
counternormative?
Suppose sensorily coded data indeed have more influence over inference
and action than nonsensorily coded data. Does this betoken any epistemic
flaw in the handling of sensorily coded data? That is not obvious. As Nisbett
and Ross themselves emphasize, one factor responsible for this situation is
that sensorily coded data are better retained and recalled. But, surely, there
is nothing counternormative about accurate and efficient recall of genuinely
learned facts! There cannot be any epistemic objection to the memory sys-
tem for being good at remembering sensorily coded information.
Where, then, can the fault lie? It could lie, as Nisbett and Ross's chapter
title intimates, in the handling of the total 'clump' of information available.
But how is that faulty? If the pallid statistical information is not well re-
membered, it seems perfectly proper to rely only on the other information,
which is remembered. The executive portion of the mind, we assume, can
only operate on activated materials. It is blameless if it fails to draw con-
clusions from materials once encountered but no longer available. One
could say that there is a flaw in the memory system: it ought to recall statis-
tical information better. However, even if we sympathize with this wish,
how would it reflect ill on the accurate retention and retrieval of imaged
information?
A further point-also suggested by Taylor and Thompson-should be
stressed. Perhaps statistical information just is not well understood. This
could influence either how well it is remembered (that is, poorly), or how
much it is used when it is remembered. Recall (from Chapter 10) that
strength of retention seems to be influenced by depth of processing. Poorly
comprehended material will not be processed deeply, because one does not
know what connections to establish between it and previous items in.LTM.
This possibility would accommodate the fact that, as Taylor and Thompson
note, many studies do not show appreciable inferiority of printed material.
Perhaps it is only the complexity of statistical materials that make them
p~orly~remembered, not the fact that they are not sensorily coded.
Suppose that statistical information is recalled but poorly understood,
and the executive system fails to draw permissible conclusions from it. Is
this counternormative?It might be argued, on the one hand, that the system
ought to understand statistical data better. Perhaps, but on the other hand,
Internal Codes 277
is there anything wrong with the refusal to draw conclusions from informa-
tion which, as a matter of fact, is not well understood? I should think not.
To draw conclusions in this situation amounts to random guessing or grop-
ing in the dark. That is not likely to promote reliability.
What about the suggestion, then, that the system behaves counternorma-
tively if it fails to understand statistical data? My reaction-to be elabo-
rated upon in later chapters-is that it is unreasonable to require this of the
system's basic architecture. Just because some species of information would
be cognitively useful, it does not follow that a system is poorly designed if it
is not equipped with an innate propensity to grasp that information. (If the
information is in principle ungraspable by the system's architecture, that is
another matter. This is not in question here.) Granted, statistical informa-
tion is something it is good for human beings to learn to understand and uti-
lize. (This might even be a principle of secondary epistemology.) Failure to
understand statistics at the outset, however, is not obviously a flaw in the
system's fundamental architecture.
This topic, however, transcends the issue of sense-linked thinking. Is
there evidence that something is wrong with our cognitive system's use of
sensory codes? For example, is its use highly unreliable, or does it contrib-
ute to unreliability? As I have emphasized, the recall strength of sensorily
coded materials is hardly a defect. The most that Nisbett and Ross are enti-
tled to claim, I think, is that (1)sensory information is likely to concern indi-
vidual, concrete cases, and (2) the system has a tendency to draw
unwarranted conclusions from small samples (for example, isolated case
histories). However, there is no necessary connection between sensorily
coded information and sample size. Imagery can, in principle, carry infor-
mation about large numbers of cases. For that matter, it can encode and
store information about statistical tendencies: it can help represent and re-
member diagrams or charts that present statistical patterns. Moreover, if
the system does tend to make improper inferences from small samples, that
is not a problem with sensory coding per se. It is far from clear, then, that
operations governing sensory codes are deficient or defective.
One moral of this discussion-encountered in previous chapters as
well-is that a very complex bundle of operations needs to be surveyed and
appraised in order to construct even a small piece of a right system of
J-rules. It is because of this complexity that I have eschewed any attempt to
present and defend any particular J-rule system. The reader may be able to
extract from my discussions any number of candidate operations that might
(or might not) be permitted by a system of J-rules. A proper appraisal of
such a system, though, must weigh the consequences of the system as a
whole. This contrasts with the (self-confessed) propensity of Nisbett and
Ross to underline the dangers and costs of certain cognitive operations (in
the present case, deployment of sense-linked materials), not to weigh their
overall performance in the context of a complete cognitive system.
chapter 13
Deductive Reasoning
into the discussion: speed. Suppose that the only effective procedure for
solving a certain class of problems would be very slow for a certain subclass
of those problems, so that use of the procedure for that class of cases would
not yield an answer in the being's lifetime. Is a being who uses a noneffec-
tive procedure instead-a procedure not guaranteed to produce a correct
answer, but which might produce one in a timely fashion-is such a being
irrational? Christopher Cherniak considers a cognizer who wishes to test a
system of propositional beliefs (beliefs involving no quantificational struc-
ture) for consistency, and his whole belief corpus contains (only) 138beliefs.
The truth table method provides an effective procedure for testing this cor-
pus for consistency. But if each line of the truth table for the conjunction of
these statements can be checked in the time a light ray takes to traverse the
diameter of a proton-an appropriate cycle time for an ideal computer-a
truth-tabular consistency test would still take more time than the estimated
twenty billion years from the dawn of the universe to the
As this example illustrates, the theoretical adequacy of a method is hardly
equivalent to practical adequacy or efficiency. Should we say, nonetheless,
that creatures who use noneffective procedures when effective procedures
exist are imperfectly rational? Again, that seems wrong. Switching to the
evaluative term 'intelligent', I feel even greater confidence in saying that
fast but noneffective procedures may be more intelligent than effective but
laborious procedures. (Notice that 'slow' is one approximate synonym of
'unintelligent'.)
The general topic of speed of computation is currently studied under the
title of 'computational complexity'.3 Complexity theory studies the practi-
cal feasibility or unfeasibility of algorithms. An algorithm's feasibility is
evaluated in terms of whether its execution time grows as a polynomial
function of the size of input instances of the problem. If it does, the algo-
rithm is generally treated as computationally feasible; if it increases faster,
usually as an exponential function, the algorithm is regarded as computa-
tionally intractable. Such intractability turns out to a large extent to be in-
dependent of how the problem is represented and of the computer model
involved.
The problem of determining whether a truth-functional argument is
valid belongs to a set of problems called ' ~ ~ - c o m ~ l e tWhile
e ' . ~ NP-com-
plete problems have not been proven inherently to require exponential
time, they are strongly conjectured to do so; and hence are thought to be
computationally intractable. So complexity theory suggests that even some
comparatively simple reasoning tasks cannot feasibly be performed in ways
guaranteed to be correct. Thus, it seems implausible to say that human
beings are 'irrational' or 'unintelligent' just because they fail to possess, or
utilize, algorithms for such tasks.
Reflections on the speed desideratum, and on the sketched results of
Deductive Reasoning 283
of cognitive process an 'intuition' is. Nor is it clear what would enable in-
tuitions to confer justifiedness. On my theory intuitions would have to be
very reliable. But what is the nature of an intuition that might give it relia-
bility? Lacking answers to these questions, appeal to intuitions is not at all
appealing.
The third possibility is that there are no basic processes by which belief in
logical principles can be arrived at justifiably. This possibility, of couFse,
would be savored by the skeptic.
So far I have mainly been conceiving logical endowments as inference
rules, procedures, or transformational routines. But there is another general
category in which our logical endowments might be described. This is the
set of logical concepts, or operators, that are native parts of our cognitive
equipment (assuming that some concepts must be native). Does our native
conceptual repertoire feature the concepts of implication and consistency?
Does it have the standard repertoire of operators familiar from formal logic:
'not', 'if-then', 'if and only if, 'or', 'there exist some', and the like?
Most of the research on the psychology of reasoning has concentrated on
the repertoire of inference procedures. But the category of logical concepts
is just as fundamental. It is often suggested, of course, that people are not
familiar with the material conditional. But it is also questionable whether
they have an antecedent hold on the logician's concept of validity or logical
implication. One possible explanation of why logic is difficult (for some stu-
dents) is that the concepts of logic do not have precise analogues in our na-
tive system of thought. Learning logic, on this scenario, is not just a matter
of translating new notation into the language of thought. Formal logic ben-
efits people because it introduces a new array of concepts and displays pro-
totypes of argumentation not already contained in thought.
In the section that follows, I shall examine some selected work in the psy-
chology of reasoning that bears on the issues I have just broached, especially
the question of justifiedness. Very little of this, however, will deal with na-
tive logical concepts, since less research has been done on this notoriously
difficult topic.' In fact, the entire topic of reasoning is very difficult, and the
research I shall sample will not manage to answer (to everyone's satisfac-
tion) all the questions I have raised. Nonetheless, this research gives illu-
minating hints at the sorts of answers that would be of interest.
proves its theorems from "outside in," working alternately backward from
the conclusion (in the subgoal tree) and forward from the premises (in the
assertion tree). The subgoal tree has no obvious counterpart in formal logic
proofs. Subgoals keep the proof procedure aimed in the direction of the ar-
gument's conclusion rather than allowing it to produce implications at ran-
dom from the premises. (This basic idea is due to A. Newel1 and H. A.
~imon.)'
ANDS' inference routines control its proof endeavors by placing new as-
sertions and subgoals in memory-these are the actions instigated by the
routines. Each routine consists of a set of conditions that must be checked
against the memory trees to determine whether the routine applies. If all
the conditions are met, then a series of actions are carried out that modify
the trees in specific ways. The routines are tested for applicability in a fixed
order, and the first routine whose conditions are fulfilled is then carried out.
Here are two examples of ANDS' inference routines, formulated in a
condition-action format. (This format resembles the format of production
systems, which I will explore in Chapter 17.) Corresponding to ordinary
modus ponens, ANDS has two inference routines, R1 and R2:
R1 (modus ponens, backward version)
Conditions: 1. Current subgoal = q
2. Assertion tree contains IF p, q
Actions: 1. Set up subgoal to deduce p
2. If Subgoal 1is achieved, add q to assertion tree.
R2 (modus ponens, forward version)
Conditions: 1. Assertion tree contains proposition x = IF p, q
2. x has not been used by H1 or R2
3. Assertion tree contains proposition p
Actions: 1. Add q to assertion tree.
they are wholly neglected by the present approach. I will return to this
issue in section 13.4.
We come now to the mental models theory. According to this approach,
reasoning is not a matter of applying abstract logical rules at all; it is a mat-
ter of creating and manipulating internal analogues. The internal analogues
'simulate' what is asserted in the premises and conclusion. By manipulating
these simulations, allegedly, one arrives at a judgment of validity or inval-
idity, or at a conclusion that appears to follow from the specified premises.
This approach has something in common with other developments in cog-
nitive science that stress inner representations of concrete instances, or ex-
emplars, such as the prototype theory of concepts.
The best developed version of the mental models approach to reasoning
is presented by Philip Johnson-Laird, in Mental ~ o d e k . As
' ~ he explains it,
the models people use to reason "are more likely to resemble a perception
or conception of the events (from a God's-eye view) than a string of symbols
directly corresponding to the linguistic form of the premises . . . The heart
of the process is interpreting premises as mental models that take general
knowledge into account, and searching for counterexamples to conclusions
by trying to construct alternative models of the premises."12
How, in particular, would this be done?13Suppose you are given a pair of
syllogistic premises and are asked to produce a conclusion that follows from
these premises. (This is the standard task that Johnson-Laird has used in his
experiments, rather than the task of deciding the validity of a full argument,
with conclusion specified.) For example, you may be given the following:
(1) Some of the artists are not beekeepers.
All of the clerks are beekeepers.
Therefore, ???
According to the theory, the first thing to do is to interpret the individual
premises, and you do this by configuring a set of mental exemplars that cor-
respond to particular artists, beekeepers, and clerks. The theory stipulates
that the model for the first premise is the one in (2), and the model of the
second premise is that of (3).
(2) artist
artist
(artist) = beekeeper
beekeeper
(3) clerk = beekeeper
clerk = beekeeper
(beekeeper)
(beekeeper)
In these diagrams the exemplars are indicated by words, though the words
themselves have no significance other than to differentiate exemplars of dif-
Deductive Reasoning 291
ferent types. To indicate identity among the exemplars, you place an equal-
ity (or identity) sign between them. So the equal sign between clerks and
beekeepers in (3) represents the fact that those clerks also are beekeepers.
In (2) the artists above the line are those that are definitely not beekeepers.
However, the parenthesized artist below the line symbolizes the possibility
that there may be an artist who is a beekeeper. In (3)both of the clerk ex-
emplars are beekeepers and there may or may not be beekeepers who are
not clerks.
After you form these representations of the premises, you have to com-
bine them into a unified model of the problem as a whole. One such model
might be:
(4) artist
artist
(artist) = beekeeper = clerk
beekeeper = clerk
(beekeeper)
(beekeeper)
In order to generate a conclusion to the syllogism, you then try to 'read'
from this model the relationships that hold between the 'outer' terms: 'art-
ists' and 'clerks'. Two conclusions that suggest themselves on the basis of the
unified model is 'Some artists are not clerks' and 'Some clerks are not art-
ists'. However, to be sure that these candidate conclusions really follow, you
need to check that there are no alternative models of the premises that
make these conclusions false.
Johnson-Laird claims that his theory can successfully predict the diffi-
culty people have in producing correct conclusions.
The major source of difficulty should be the construction and evaluation of
alternative models. This process places an additional load on working mem-
ory-the memory system that is used for holding the mental models while
they are manipulated. The greater the number of models to be considered,
the harder it should be to construct and to evaluate them. Hence, the task
should be harder when it is necessary to construct two models, and harder
still when it is necessary to construct three models.14
These predictions are borne out by experiments that Johnson-Laird did
with various colleagues. In one experiment, for example, twenty students at
Teachers College, Columbia University were given all sixty-four possible
pairs of syllogistic premises. For the premise pairs requiring one model to
be constructed to yield the correct conclusion, 92 percent of the correct
(valid) conclusions were drawn. Of the premise pairs requiring two models,
46 percent of the correct conclusions were drawn. And of the premise pairs
requiring three models, 27 percent of the correct conclusions were drawn.''
Johnson-Laird also believe-uite plausibly-that a good theory of rea-
292 Deductive Reasoning
sequence whose final stage incorporates a mere guess presumably would not
be very reliable and would not be licensed by a right J-rule system.
By 'guess', I here mean a process that fixes a belief. There is another sense
of 'guess' that simply involves the choice of a verbal response, unaccompa-
nied by a genuine belief. Many of the subjects' responses in Rips's experi-
ment may have been guesses in this second sense, because subjects were
forced to respond either 'valid' or 'invalid'. But since the present discussion
focuses on processes that yield justified or unjustified beliefs, I am only con-
cerned with guesses in the sense that triggers beliefs. (Notice that if many of
the responses were just guesses in the second sense-mere verbal responses
not registering genuine belief-then the number of subjects' false beliefs
was not as high as the data suggest.)
It might be argued that when a subject tries to construct a proof, fails,
and infers that the argument is invalid, this last inference is not a mere
guess. It is an inductive inference, perhaps of an 'explanatory' variety. It
might run: 'I tried a variety of ways of constructing a proof, but none of
them worked. The best explanation of this failure is that there is no proof.
Therefore, the argument is not valid'. This is indeed one way a subject
might reason. It is not obvious, though, whether this would be an approved
process of reasoning. What it shows, however, is that even in Rips's model
there might be justified beliefs in propositions concerning logical subject
matter that are justified by inductive rather than deductive inference pro-
cesses. Actually, it seems to be a consequence of Rips's model that the only
way in which a person can justifiably believe in an argument's invalidity is
by this sort of inductive inference. The ANDS model offers no way of prov-
ing invalidity.
I can now summarize the implications of Rips's model as follows. If this
model is right-and right about our basic reasoning processes-there
clearly are some operation sequences used to form beliefs in argument va-
lidity that would be licensed by a right J-rule system. Specifically, where a
person constructs a proof by means of any sequence of (officially designated) ,
routines incorporated in Rips's model, and infers validity of the target ar-
gument from this proof, then this belief is presumably justified. But it leaves
open the possibility that people have no available operations for judging
argument invalidity that would be permitted by a right J-rule system. Or, if
there are such operations, they must have a significant inductive compo-
nent. This last possibility is especially interesting. Could it be that for some
judgments of logic, people have to use inductive processes to arrive at justi-
fied beliefs? Shades of John Stuart Mill!
We come next to the plausibility of a Rips-style model. How, in particu-
lar, does it square with (apparent) evidence of fallacies and mistakes of vari-
ous kinds? Consider fallacies that are sufficientlyprevalent to have received
labels, such as 'denying the antecedent' and 'affirming the consequent'.
296 Deductive Rearoning
Some data on this matter are found in a study by Rips himself and a col-
league.16 Subjects were asked to evaluate eight types of "conditional syllo-
gisms," shown in Table 13.1. Subjects were asked to judge whether the
conclusion was "always" true, "sometimes" true, or "never" true, given the
premises. As the figure shows, subjects could apply modus ponens quite
successfully. But there was considerable evidence of tendencies to 'deny the
8. P > Q
-Q 57" 39 4
. :. -P
Source: L. J. Rips and S. L. Marcus, "Supposition and
the Analysis of Conditional Sentences," in M. A. Just
and P. A. Carpenter, eds., Cognitive Processes in
Comprehension (Hillsdale, N .J.: Erlbaum, 1977).
a. The correct response.
Deductive Reasoning 297
Figure 13.1
task."lQIn the central experiment from this research, four cards were shown
to subjects (see Figure 13.1).Subjects were told that a letter appeared on
one side of each card and a number on the other. They were asked to con-
sider the following rule, which referred only to these four cards: "If a card
has a vowel on one side, then it has an even number on the other side." The
task was to decide exactly which cards had to be turned over to determine
the correctness or incorrectness of the rule.
Forty-six percent of the subjects elected to turn over the E and the 4,
which is a wrong combination of choices. The E card does have to be turned
over, but the 4 card does not, since neither a vowel nor a consonant on the
other side would falsify the rule. Only 4 percent elected to turn over the E
and the 7, which is the correct answer. (An odd number behind the E or a
vowel behind the 7 would falsify the rule.) Another 33 percent of the sub-
jects elected to turn over the E only. The remaining 17 percent made other
incorrect choices.
There are two main types of error in this task. First, subjects often turn
over the 4, which is not necessary. Second, there is almost total failure to
appreciate the necessity of turning over the 7. Admittedly, not all of these
mistakes need be attributed to application of unsound inference routines;
they might be attributed to omission of sound routines. But they again cause
one to wonder whether all the manipulations subjects are performing can
be accounted for by the kind of routine repertoire that Rips postulates.
Let us turn briefly now to Johnson-Laird. It is hard to compare his theory
to that of Rips for several reasons. First, it deals primarily with syllogistic
reasoning, whereas Rips's model deals with propositional reasoning. Sec-
ond, the tasks Johnson-Laird gave his subjects were Wh-tasks ('What con-
clusions follows from these premises?), not Y/N tasks ('Is this argument
valid?). Third, Johnson-Laird does not really present any routines by which
subjects are alleged (a) to construct 'models', or 'interpretations', of the
premises and contemplated conclusion, and (b) to check whether they have
considered all relevant models. His description of the operation sequences
that subjects undergo is therefore quite sketchy.
Given this sketchiness, my treatment of the theory must be a bit specula-
tive. It seems, though, as if nothing corresponding to deduction-framed
routines are part of the process envisaged by the theory. In particular, first,
nothing of that sort seems to be at work when subjects attempt to construct
Deductiue Reasoning 299
Probability Judgments
What can psychologists do, or tell us, that concerns rationality or other
dimensions of intellectual appraisal? I have pursued this question, of course,
through much of this book. At this point it may be helpful to approach it
from a slightly different angle, more directly dovetailing with the work of
people like Tversky and Kahneman.
First of all, the psychologist can arrange experiments in which subjects
answer certain questions, and he can report these responses. Second, the
psychologist can often say what proportion of these responses are true or
false; and presumably he can say whether the responses are consistent or co-
herent. It is not obvious that this second sort of issue is always within the
psychologist's sphere of competence. In perceptual experiments the psy-
chologist tells us the true properties of the presented stimuli; he is pre-
sumed to be authoritative (enough) on that matter. In other kinds of
experimental tasks, though, the truth or falsity of the subjects' responses
may be more controversial. For example, whether specific probability judg-
ments are true or false is entangled in controversial issues about the inter-
pretation of probability sentences. So what a psychologist might say on this
score could be disputed by others.
Third, the psychologist can offer hypotheses about the cognitive pro-
cesses that underlie the observed responses. That is, he can theorize about
the procedures, mechanisms, strategies, or heuristics that must be leading
the subjects to respond as they do. Of course, any proffered explanatory hy-
pothesis is open to empirical dispute. How well does it explain the evi-
dence? What competing model might explain the evidence as well, or
better?
Fourth, there is the question of the normative status of the subjects' re-
sponses; or the normative status of their using certain procedures, and fail-
ing to use others, in arriving at their responses. This question, it seems to
me, falls outside the domain of psychology, narrowly construed. It is not the
job of empirical science to make normative judgments (whether ethical,
aesthetic, or epistemological). When it comes to epistemic normative judg-
ments, this is the task of epistemology.
However, if certain epistemic rules or principles are agreed upon, then
the psychologist may be in a position to tell us whether or not human beings
conform with these rules or violate them, and with what frequency or in
what circumstances. But if there is any dispute about which normative
principles are correct, this dispute goes outside the realm of psychology
per se.
Of course, I have been contending that principles of primary epistemol-
ogy are partly determined by psychological facts. So on my view the psy-
chologist can help epistemology ascertain correct normative principles.
And the very sorts of experiments in question may help do this. But this
flows from the criteria of rightness (at least J-rule rightness) that I have de-
Probability Judgments 307
fended, and the defense of such criteria falls outside psychology altogether.
In the present context, then, there are three points I wish to stress about
the psychologist's role. First, it is certainly possible that the psychologist
might tell us things about human cognitive performances which imply (to-
gether with correct normative principles) that people's beliefs or probabil-
ity estimates are frequently, even characteristically, unjustified, irrational,
or otherwise counternormative. (Here I disagree sharply with Cohen.) But
second, the psychologist is not necessarily authoritative on what the correct
normative principles are. So if he makes any claims about such principles,
those are open to dispute. Third, since I have argued that epistemic rules, or
norms, should specify acceptable (and unacceptable) cognitive processes,
the psychologist can help arrive at such rules by trying to uncover what
sorts of cognitive processes are available to the human organism.
versity of California, Berkeley and at Stanford University, all with credit for
several statistics courses, were given a transparent version of the Linda
problem. For the first time in this series of studies, less than half of the re-
spondents committed the fallacy-'only' 36 percent. But even this seems
fairly high, given the subjects' statistical sophistication.
The hypothesis proposed by Tversky and Kahneman to explain the find-
ings I have summarized is the 'representativeness' heuristic. Both in the
studies where subjects were insensitive to prior probabilities (base rates)
and in the conjunction fallacy studies, subjects' probability estimates were
apparently guided by the degree to which the person in question was repre-
sentative of, or similar to, the classes defined by the queried events or out-
comes. For example, the description of Linda presumably made her
representative of an active feminist and unrepresentative of a bank teller.
Presumably, she was also more representative of a feminist bank teller than
a bank teller. Indeed, in many experiments on the conjunction fallacy, sub-
jects were explicitly asked to rank various outcomes both for representa-
tiveness and for probability. The correlation between the mean ranks of
probability and representativeness were typically extremely high, often
greater than .95. Tversky and Kahneman also propose the representative-
ness heuristic as an explanation of misconceptions of chance. People's
(mis)estimates of coin toss likelihoods, they suggest, is due to their expecta-
tion that coin toss sequences will be representative of the random generat-
ing process.
What, exactly, is 'representativeness'? It turns out to be a rather multi-
faceted relation. In "Judgments of and by Representativeness" Tversky and
Kahneman discuss this relation in some detail.'' They say that representa-
tiveness is a relation between a process or model M and some instance or
event X, associated with that model. The model in question could be of a
person, a fair coin, or the world economy, and the respective outcomes
might be a comment, a sequence of heads and tails, or the present price of
gold. Normally, the sample or act is said to be representative of the popula-
tion or person. But these roles can be reversed, and we can say that the oc-
cupation of librarian is representative of a person.
They then distinguish four types of cases in which representativeness can
be invoked: (1)M is a class and X is a value of a variable defined in this class.
For example, the mean income of college professors might be representa-
tive of that class. (2) M is a class and X is an instance of that class. In this
case, John Updike might be regarded as a representative American writer.
(3) M is a class and X is a subset of M. The set of psychology students might
be moqe representative of the entire student body than the set of astronomy
students. (4) M is a (causal) system and X is a (possible) consequence. For
example, M can be the U.S.economy and X the rate of inflation, or M can
be a person and X an act performed by M (such as divorce, suicide, profes-
sional choice).
Probability Judgments 31 1
In early work Tversky and Kahneman suggested that people rely exclu-
sively on the similarity of a sample to its population parameters to make
probability judgments. Now they make a more moderate hypothesis: pre-
dictions and probability judgments are "highly sensitive" to representative-
ness, though not completely dominated by it. The magnitude of the appeal
to representativeness depends on the nature of the problem, the character-
istics of the design, the sophistication of the respondents, and the presence
of suggestive clues or other demand characteristics."
Given the multifacetedness of the representativeness relation, and the
multiple determinants of whether, or to what degree, people rely on it, it is
not clear that any very specific hypothesis concerning the reliance on repre-
sentativeness is currently proposed. Rather, it is a more general suggestion
that people 'often' rely on representativeness, or something of this sort. But
however frequently this reliance is made, there is the suggestion that any
appeal to representativeness in making probability judgments is counter-
normative. There is also the claim that failure to employ correct probabilis-
tic and statistical rules is counternormative. This normative claim is what I
address next.
set theory, that for every formulable condition, there is a set whose mem-
bers are just the things fulfilling that condtion. Russell's paradox showed
that this axiom implies contradictions; but were those who believed in the
axiom prior to Russell's discovery of the paradox irrational? Surely not. In-
consistencies are often extremely difficult to detect; and failure to detect
them need not betoken irrationality.
An example from another domain may help solidify this point. The artist
M. C. Escher is famous for drawings that appear, on initial perusal, to de-
pict a consistent scene. But close inspection gradually reveals that there is
no possible coherent interpretation. Is a viewer who fails (even for some
time) to recognize the scene's incoherence irrational? I think not. Doubt-
less, such a viewer fails to notice something of importance, but this hardly
warrants the charge of irrationality.
None of these comments is intended to deny that a right set of epistemic
rules-either at the primary or the secondary level-might require at-
tempts to discover and remove inconsistencies in some contexts, perhaps
many contexts. Shaping your belief corpus with an eye toward consistency
is doubtless an effective way of increasing reliability. If one's perceptual
beliefs are usually true, then forming further beliefs consistent with these is
presumably a helpful way of working toward new beliefs with a high truth
ratio. My point has only been to put the seriousness of inconsistency into a
better, more measured, perspective.
Bearing these points about inconsistency in mind, let us turn now to
questions about probabilistic incoherence. Here we must deal separately
with objectivist and subjectivist interpretations of probability judgments.
On an objectivist interpretation probabilistic incoherence is just a special
case of inconsistency. Suppose you judge that the probability of its being
rainy tomorrow is .60, and that the probability of its being rainy and windy
tomorrow is .75. On the objectivist interpretation this means that you be-
lieve both Q, 'The chance of its being rainy tomorrow is .6OY,and Q", 'The
chance of its being rainy and windy tomorrow is .75.' Since chances obey .
the probability calculus (all objectivists assume this much), these proposi-
tions are inconsistent; they cannot both be true. (At least they are mathe-
matically if not logically inconsistent.) So probabilistic incoherence in this
case is just inconsistency. As we have seen, however, inconsistency is not
necessarily the most flagrant or serious epistemic defect. It is not clearly
epistemically objectionable in every case. Similarly cautious normative
conclusions should be made for probabilistic incoherence.
When we move to the subjectivist interpretation of probability judg-
ments, the matter is somewhat different. Probabilistic incoherence is no
longer a case of believing logically inconsistent (or mathematically incon-
sistent) propositions. In the foregoing example the subjectivist would not
interpret you as believing propositions Q and Q". Rather, he would inter-
316 Probability Jtidgments
tenstein showed that subjects become sensitive to base rates when they en-
counter successive problems that vary in this critical variable.''
As far as secondary justifiedness goes, people are evidently capable of
using correct probabilistic methods. They often do not use them, even when
they have been trained; but sometimes they do. Hence no universally bleak
conclusion about secondary justifiedness of probabilistic judgments is war-
ranted. Thus far, there are no grounds for the eager skeptic to rejoice.
But how does this conclusion affect primary justifiedness? How adequate
are the basic processes that get deployed in connection with probability
tasks? Are they so poor as to be excluded by all right J-rule systems? Or are
some of them sanctionable by a right J-rule system? Much of the work in
this field has centered on the representativeness heuristic, which bears the
brunt of the attack on native processes in probability judgment. Recall from
section 14.2, however, that Tversky and Kahneman no longer claim that
people rely exclusively on representativeness in making probability judg-
ments. They now say that predictions and probability judgments are
"highly sensitive" to representativeness, not completely dominated by it.
There must, then, be other processes that get deployed in such tasks.
In "Extensional versus Intuitive Reasoning" Tversky and Kahneman
speak of people using a "holistic, singular and intuitive" manner of evaluat-
ing probabilities.18This phrase, though, does not identify any specific cog-
nitive process. In another article they explain their use of 'intuitive' as
follows. A judgment is called 'intuitive' if it is reached "by an informal and
unstructured mode of reasoning, without the use of analytic methods or de-
liberate calculation."19 But this is a largely negative characterization,
which still does not identify the specific processes at work. There is no as-
surance, moreover, that only one such process is used. Maybe a number of
processes can be used, some more reliable than others. In particular, some
reIiable processes may be employed by sophisticated cognizers when they
correctly identify problems as susceptible to probabilistic methods and then
apply those methods appropriately.
But what about the representativeness heuristic? Doesn't the evidence
suggest that it is frequently used? And doesn't this seem to be a basic pro-
cess that would be barred from any right J-rule system? First, a point of ter-
minology. The term 'heuristic' connotes a deliberately chosen shortcut
procedure. In the present case the process is not deliberately chosen and is
not seen as a shortcut. So I will speak instead of a representativeness 'rou-
t i n e ' a n R-routine.
I do not wish to deny that use of the R-routine leads to errors in the class
of cases-to which Tversky and Kahneman draw our attention. Specifically,
when subjects rank the probability of a conjunction as higher than the prob-
ability of a conjunct, this is clearly an error. (At least it is an error if this
ranking expresses a belief in a comparative objective probability statement,
Probability Judgments 321
and if each event has a genuine objective probability.) However, it does not
follow from this that the R-routine is not admissible in a right J-rule system.
Perhaps the R-routine is generally quite reliable, though it breeds errors in
this subclass of cases. A case can be made for just this thesis by contemplat-
ing the possibility that the R-routine is a facet of the general use of match-
ing operations, or similarity assessments, in cognition.
Matching operations may well be a fundamental part of cognitive pro-
cessing; and matching of an object X to a model M seems to be the core of
the R-routine, as Tversky and Kahneman view it. Another way of putting
this idea is in terms of similarity assessments. The use of similarity is a ubiq-
uitous facet of human cognition (as noted in Chapter l l ) , and is particularly
stressed in recent work on concepts and categorization. Let us explore this
latter research, as a prelude to the discussion of the R-routine in probability
judgments.
Cognitive psychologists investigate the psychological representations and
processes that underlie the use and comprehension of (natural-language)
words, especially common nouns for ordinary objects like 'bird' or 'chair'. A
popular theory is that people mentally represent the properties in question
by means of 'prototypes', and that they decide how an object should be ca-
tegorized by assessing the similarity of the object to the category's
prototype.
The chief rationale for this theory is the ubiquity of 'typicality' effects.
For example, subjects find it natural to rate the subsets or members of a cat-
egory with respect to how typical each is of the category. Such ratings were
first reported by L. J. Rips, E. J. Shoben, and E. E. Smith and by E. ~osch."
Rosch had subjects rate the typicality of members of a category on a 1-to-7
scale, where 1 meant very typical and 7 very atypical. Subjects were ex-
tremely consistent. In the bird category 'robin' got an average rating of 1.1,
and 'chicken' 3.8; 'football' was rated a very typical sport (1.2) but weight-
lifting was not (4.7); 'carrot' was rated a very typical vegetable (1.1),and
'parsley' a less typical one (3.8).Other typicality effects include these: (a)
categorization is faster with more typical test items; (b) typical members of
a concept are the first ones learned by children; and (c) typical members of
a concept are likely to be named first when subjects are asked to produce all
members of a category.
Typicality effects are to be expected if subjects represent a category in
terms of a prototype, and if categorization proceeds by judging similarity to
prototype. There are also several variants of the prototype theory that
would all invoke similarity assessments, or matching processes, in
categorization.'l
What is important for my purposes is that judgments of similarity, or fea-
ture matching, are fundamental ingredients of the categorization process. If
this general idea is correct, the use of representativeness is an extremely
pervasive facet of human cognition. Whenever objects or events are cate-
gorized as belonging (or not belonging) to this or that concept, representa-
tiveness is employed. Furthermore, as long as cognizers apply this routine
to perceptually detected features of the object, as long as the cognizer's
prototype of the category corresponds closely enough to the prototypes
used by other members of the linguistic community, and as long as the rou-
tine of matching to prototype is executed properly, virtually all the cate-
gorization judgments made in this manner will be true! Objects judged to be
members of the concept Y will really be members of Y-that is, will really
qualify under the lexical label, as it is used by the linguistic community. In
short, this use of the R-routine will be highly reliable.
If the R-routine is so reliable for this large class of cases, the general rou-
tine might be endorsed by a right system of J-rules, despite the fact that the
routine generates errors in some of its applications. For purposes of com-
parison, consider the case of perceptual processes that sometimes breed il-
lusion. (Tversky and Kahneman themselves draw the analogy between
judgmental illusions and perceptual illusions.) Presumably, the processes
responsible for well-known illusions like the moon illusion or the Ames-
room illusion are generally reliable processes. As such, they would presum-
ably be sanctioned by a right J-rule system, despite the fact that they un-
doubtedly breed false judgments in the designated contexts. Thus, even in
the illusion cases, a cognizer's belief is justified, though it is of course false.
Similarly, if the R-routine is simply part of a generally reliable judgmental
process, the beliefs it generates on probabilistic subject matter might still be
justified, even if they are false.
Actually, this needs to be qualified. The most we could conclude is that
these beliefs have primary justifiedness. They could still be denied the status
of secondary justifiedness, because of failure to deploy learned probabilistic
methods. The same point would hold for the perceptual cases. If an adult
believes that the moon hovering over the horizon is genuinely bigger this
evening than at other times, we would not accord this belief the status 'jus-
tified'. We would surely hold that the report of his visual system (his 'visual
belief) should be overridden by his central system, by appealing to facts or
methods he has learned from past experience. Similarly, even if the R-rou-
tine is permitted by a right system of J-rules (at the primary level), someone
who has studied probability theory might be held accountable for failure to
employ correct probabilistic methods to revise his initial, intuitive
judgment.
There is another, more important, reason why the foregoing analysis
might not suffice for justifiedness. We have been supposing that there is a
single process that employs matching, or the R-routine; but this is dubious.
Arguably, there are significant distinctions between the process used in ca-
tegorization and the process used in the probability judgment cases. Al-
Probability Judgments 323
ity of 99, with respect to each ticket, that the ticket will lose. On the pre-
sent proposal this would be interpreted as acceptance of the proposition
that the ticket will lose, for each such proposition. Yet he also accepts the
proposition that some ticket will win, so the current proposal forces us to
ascribe to him inconsistent acceptances. That seems unfair. Nothing in the
initial description of him warrants the conclusion that he has inconsistent
beliefs.
Thus far I have worried about finding mutually comfortable niches for
both subjective probabilities and acceptance. But now I want to turn to
problems for the subjective probability interpretation of credal uncertainty.
In other words, granted that there is some psychological reality to the no-
tion of tentative, hesitant, or incomplete conviction, is this correctly or best
described in terms of the probability calculus? I wish to question this de-
scriptive assumption, generally associated with (subjective) Bayesianism.
One problem is the empirical evidence against the notion that people
conform their beliefs to the probability calculus. One piece of evidence, in-
tensively discussed in Chapter 14, concerns the principle of probability the-
ory that the probability of a conjunction cannot exceed that of any of its
conjuncts, a rule people readily violate. Similarly, according to standard
models, the probability of any tautology, or necessary truth, is 1; and the
probability of a contradiction, or impossible proposition, is 0. But people
are not always maximally confident in every tautology or maximally confi-
dent in rejecting every contradiction. When they do not recognize a tautol-
ogy or a contradiction, they may have quite different degrees of confidence
in them.
Next, consider whether it is plausible that humans should have a psycho-
logical mechanism that fixes credal states in accord with the probability
calculus. Remember, I am here regarding subjective probability as a de-
scriptive theory, one that is intended to hold for all human beings, whatever
their age or educational background. Subjective credences must therefore
be fixed by native mechanisms. Are there any native mechanisms that
would make credal states behave like probabilities? To assume there are is
to assume there are native mechanisms for making arithmetic calculations
of the kind needed for determining probabilistic relations. But native mech-
anisms do not seem adequate for even the simplest arithmetic. Most arith-
metic needs to be culturally acquired by explicit instruction; it is not
programmed into the organism. Admittedly, there might be mechanisms
that simulate arithmetic computation at a preconscious level, just as visual
mechanisms apparently simulate mathematical operations in determining
distance or in computing the trajectory of an object flying toward you. But
there is no clear evidence of credence-fixing mechanisms that simulate the
probability calculus in this way.
Many theorists of subjective probability might reply that the subjective
probability model is descriptively tenable as long as there is a procedure for
Acceptance and Uncertainty 327
ascribing such states to the organism on the basis of overt choice behavior,
including expressed preferences or indifferences vis-a-vis gambles. A num-
ber of such procedures, of course, are proposed in the literature? This liter-
ature includes a number of 'representation theorems', which prove that if
certain preference axioms are satisfied, then there exists a set of subjective
probabilities that can be imputed to a decision maker. To be more precise,
if these axioms are satisfied, there exists a subjective probability function P
and a desirability function D such that, when 'subjective expected utility'
(SEU) is calculated in terms of them, an item X will be preferred to an item
Y if and only if SEU (X) > SEU (Y).
There are several problems with this approach. First, the mere fact that
subjective probabilities can be imputed to a cognizer does not establish any
psychological reality to those imputed states. Lots of different models of in-
ternal states and processes could be consistent with choice behavior. Just
because a given assignment of desirability and subjective probability is pos-
sible does not show that there is any genuine psychological reality to this
assignment. The epicycles of Ptolemaic astronomers kept their theory con-
sistent with observed planetary trajectories, but that did not establish the
astronomical reality of those postulations.
Second, in the present case, there are empirical doubts about some of the
preference axioms presupposed by the representation theorems. For exam-
ple, Tversky found systematic and predictable intransitivities of prefer-
ence,%nd K. R. MacCrimmon reported violations of most of Leonard
Savage's postulates.6 Moreover, many of the popular approaches to subjec-
tive probability assignments presuppose that choice is determined by
SEU; but this presupposition is empirically quite doubtful. For example,
in one of Tversky and Kahneman's experiments subjects were asked
to choose between gambles A and B (Choice I), and then between C and
D (Choice 11).
Choice I A = $1,000 with probability .5, or $0 with probability .5
B = $400 with probability 1
-
Choice I1 C = $1,000 with probability .l,or $0 with probability .9
D $400 with probability .2, or $0 with probability .8
Nearly all subjects chose B over A, and C over D, and analogous results
were obtained using different payoffs and probabilities. These results pre-
sent grave difficulties for the assumption that choice follows S E U . ~ To the
extent that subjective probability assignments are predicated on standard
decision theory, there are serious grounds for doubts about their empirical
worth.
There is a method for determining subjective probabilities that does not
explicitly presuppose SEU: the Dutch Book method. This determines a sub-
jective probability by taking the highest odds one is willing to offer for a bet
on p-let's say a to b. Then the degree of belief in p is fixed as a/a + b.
328 Acceptance and Uncertainty
Figure 15.1
Acceptance and Uncertainty 331
- -
6
5.5
5.5
6
6.5
7.5
9.5
Sot
Figure 15.2
To illustrate these ideas, let a unit be a neuron with dendrites that receive
inputs. Each dendrite can be thought of as an alternative enabling condi- ,
tion: if it receives enough inputs, it activates the neuron. The firing fre-
quency of the neuron might be the maximum of the firing rate at any one of
its dendritic sites. The formula for this unit's potential could be written as
follows:
pt+l+-- pt + Max (i, + i2; iJ + i4; i5 + i6 - i7)
This is shown in Figure 15.1. The minus sign associated with i7 indicates
that it is an inhibitory input.
Drawing on these ideas, I will illustrate a configuration of symmetric
units that mutually inhibit one another. First, consider Figure 15.2, con-
taining two units, A and B, each with two input sites. Suppose the initial
input to site Al is 6, after which that site receives 2 inputs per time step.
Suppose B1 receives an initial input of 5, and then 2 inputs per time step.
332 Acceptance and Uncertainty
Notice that each unit transmits an inhibitory input to its mate, at site 2.
Suppose further that the weights at the sites are as follows. The weight at
site 1 is 1, and the weight at site 2 is -.5. Then at each time step, each unit
changes its potential by adding the external input value and subtracting
half the output value of its mate. So we have the equation:
Finally, assume that the output variable v takes the discrete values: 0, . . .,
9; and that v is always the rounded value of the potential p.
The result, as indicated in the table in Figure 15.2, is that the potential
and output of unit B are gradually reduced to 0, while unit A gradually in-
t P( A ) P( B) P(C) P( D l
1 6 5 6 5
2 6.5 4.5 6.5 4.5
3 7.5 305 7.5 3.5
4 9.5 1.5 9.5 1.5
Sot 0 Sat 0
Figure 15.3
Acceptance and Uncertainty 333
creases its potential and output to a saturation point. There the system
stabilizes.
Another symmetric configuration involves coalitions of units, shown in
Figure 15.3.The idea here is that units A and C form a coalition with mutu-
ally reinforcing connections. This coalition competes with another coalition
comprised of B and D. Competition between the coalitions is determined
by their mutual inhibition. As in the previous example, such mutual inhibi-
tion results in convergence toward a stable state, but here the convergence
is faster: convergence occurs in only four time steps. This system is de-
scribed by the following equations and specifications:
W1 = 1, W2 = .5, W3 = -.5
v = round (p)
pt+= + +
~ pt i, .5(i, - i3)
A, C start at 6; B, D start at 5
A, B, C, D have no external input for t > 1
\
\ I \ I
Figure 15.4
In other words, each unit sets itself to zero if it knows of a higher input.12
This general idea is illustrated in Figure 15.5. The inhibitory connections
guarantee that each unit stops if it 'sees' a higher value. This is fast and
simple, though probably not terribly realistic. However, there are other
ways in which a WTA network might be realized.
The basic idea of acceptance, as I view it, is that a proposition is accepted
only when all rivals, or contraries, are rejected. This is well captured by the
idea that all rival hypotheses, or interpretations, have their activity levels
reduced to zero. This does not necessarily mean that these competing hy-
potheses cannot later be reactivated. But acceptance marks the emergence
of a decisive winner, not a hypothesis that merely has more strength than its
competitors. At the same time, belief must also be a reasonably stable, or
settled, state. If one's attitude is continually fluctuating, one does not yet
have 3 belief. Roughly this account of belief was endorsed by C. S. Peirce,
who viewed belief as a "calm" and "satisfactory" state, in the nature of a
"habit."13
This account of acceptance carries with it a plausible account of its com-
Acceptance and Uncertainty 335
Winner-Toke- All
E a c h unit stops if i t sees o higher volue
Figure 15.5
bilities: a coalition. Furthermore, the larger and more detailed such a coali-
tion, the greater the amount of reinforcement each hypothesis (unit) re-
ceives from the remaining members. Thus, a comprehensive and detailed
scientific theory would have a high degree of coherence. Of course, no
member of a coalition can raise the strength of others unless it gets some
strength from outside the coalition; presumably, the external sources of ex-
citation are ultimately traceable to perception. Thus, perception as well as
coherence will serve as a major determinant of strength.
How can coherence, or coalition formation, serve as a guide or indicator
of truth? First, since a necessary condition of a hypothesis' being true is its
compatibility with other truths, the absence of coherence may be a good
sign of falsity. 'Isolated' hypotheses cannot benefit from reinforcement by
other coalition members, so they are ultimately rejected. Furthermore, if
coalitions are formed under some previously established pattern*
schemas, frames, scripts, scenarios, and so on-instances of which have
been frequently encountered in the past, there is evidence that this pattern
has a non-negligible prior probability of occurrence. Thus, a coalition will
represent not merely an abstract conjoint possibility, but also an instance of
a type with a non-negligible chance of recurrence. So, on the present con-
ceptualization, coalitions register 'induction' from past observation-in the
form of inductively based schemas, scripts, and so on-as well as the other
ingredients previously cited.''
Although membership in a strong coalition is a useful test of truth, it ob-
viously is not a sufficient indicator of truth. Suppose a rival hypothesis also
belongs to a strong coalition. That ought to raise doubts about the truth of
the target hypothesis. Allowing groups of coalitions to compete may thus be
a useful device for selecting true hypotheses as winners.
Until now I have directed all attention to a single dimension of epistemic
appraisal, justifiedness, and its associated standard, reliability. But what of
the other standards of appraisal that interest us: speed and power? Any
facet of a WTA mechanism that delays acceptance obviously works against
speed. For example, the larger the d-parameter, the more it takes for the
mechanism to reject rivals. On average, it will take longer for the mecha-
nism to select a winner. Similarly, the more the cognitive system seeks and
creates competing coalitions, which might give an initially favored hypoth-
esis a run for its money, the slower the system will be in selecting a winner.
Of course, both of these traits would enhance reliability; but there would be
a loss in speed.
It seems likely that the human system has a speed-sensitive design, not
just a reliability-sensitive design. In language understanding, for example,
the hearer processes sentences very rapidly, normally selecting a unique
grammatical parsing and a unique semantic interpretation for each sen-
tence, even where there are (in principle) several alternative parsings and
Acceptance and Uncertainty 341
readings. (Of course, if there is a special language module, this fact may not
speak to doxastic decisions in other domains. But it is certainly suggestive of
a design that is oriented toward speed considerations.)
The core of a WTA device may also be seen as sensitive to considerations
of power. The fundamental impact of a WTA device is to remove clutter
from active memory. As long as rivals are competing with a given hypoth-
esis, active memory will be cluttered with items requiring processing re-
sources. This makes it more difficult to move on to other questions that
could be answered. Selection of a winner settles an issue and allows the
mind to proceed more quickly to the rest of its agenda, enabling it (in prin-
ciple) to solve a larger number of problems.
What I have said thus far shows that a WTA mechanism is cogenial to the
desiderata of speed and power, but not incompatible with achieving high
reliability. Other epistemologists, however, have not viewed the idea of
'acceptance' so favorably. Here is a kind of objection they might raise
against acceptance, against a process that selects one hypothesis as a winner
while reducing the strengths of rival hypotheses to zero.
As background, notice that many theorists are attracted by what may be
called a pragmatic criterion of evaluation (recall section 5.4). An example of
a pragmatic criterion is one that evaluates cognitive processes by their im-
pact on desire satisfaction. Decision theory stresses that a cognizer's subjec-
tive probabilities are among the factors that determine the choice of
behavior. And behavior, of course, may be more or less successful in obtain-
ing outcomes the agent desires. Now it may be suggested that an optimal
cognitive system is one that so selects behavior as to optimize (the agent's)
desire satisfaction, or payoffs. Is a WTA mechanism optimal according to
this standard?
One problematic feature of a WTA mechanism leaps out in this context.
It is in the nature of a WTA mechanism to reduce the strength of all but one
of a set of competitors to zero. Doesn't this entail a loss of critical informa-
tion, and won't this produce nonpragmatic, or nonprudential, choices? Con-
sider an example. You are driving on a two-lane road, and there is Iittle
traffic in the opposite direction. You are now caught on a hill behind a slow
truck, and you are pressed for time. Should you try passing the truck, even
though you cannot see over the hill? The sparsity of traffic traveling the op-
posite way suggests that the hypothesis 'A car is coming over the hill', is
much weaker in strength than its rival, 'No car is coming over the hill'. It
appears, then, that a WTA mechanism would reject the former hypothesis
and accept the latter. Once these doxastic decisions are plugged into your
behavioral decision matrix, it looks like a decision to pass the truck will be
made. Yet this seems likely to be a nonpragmatic, or nonprudential, choice.
In particular, if the objective probability that a car is coming is, say, .l5,
and if you act this way repeatedly in similar situations, the long-run per-
342 Acceptance and Uncertainty
formance of these mental mechanisms will probably lead to a very bad pay-
off. This kind of case can obviously be generalized. It then begins to look as
if the WTA process of acceptance has very bad properties, at least as judged
by a pragmatic standard.
There are at least two ways of countering this objection, granting (for ar-
gument's sake) the appropriateness of a pragmatic standard. First, an ac-
ceptance process does not necessarily preclude retention of probabilistic
information. As stressed in section 15.2, information about chances may be
accepted and therefore preserved for decision-making purposes. In the
driving example you could accept the proposition 'The chance that a car is
coming in the opposite lane is .IS'. The prevalance of WTA processes, then,
does not entail the suppression of all probabilistic information. Admittedly,
such preservation of information about chances involves belief contents
that invoke an objectiue probability notion, which is philosophically con-
troversial. But if sense cannot be made of objective probabilities, it is not
clear that the prudential criticism can be properly motivated in the first
place. Certainly the form of criticism advanced above relies on objective
probabilities.
Second, there is a possible mechanism, a complementary mechanism,
that would avert the apparent shortcoming of an acceptance process. Sup-
pose there is an allied cognitive mechanism that is sensitive to possible pay-
offs with extreme values: very rewarding or very threatening outcomes. This
mechanism increases the activation level of any contemplated hypothesis h
that is believed to make some extreme payoff likely. That is, h's activation
level is increased if the cognizer believes that some extreme payoff would
be likely if h were true. Such heightened activation would tend to prevent h
from being rejected, would keep it 'in contention7, and hence not ignored
during behavioral choice. In the foregoing example the threat of a fatal ac-
cident heightens the activation level associated with 'A car is coming7.This
keeps it from being rejected, and keeps its negation from being accepted.
Thus, the choice of the 'don't pass' option is promoted. (I do not offer any
specific model of how a choice would then be made. But such models could
readily be constructed.)
I suspect that human beings do have some such mechanism. Indeed, it
may be just the mechanism that accounts for wishful thinking and fearful
thinking. The present analysis shows how such a mechanism can have pru-
dential merit--even if it has attendant liabilities for reliability. If a cognizer
believes hypotheses largely because they would, if true, fulfill his fondest
dreams or deepest fears, these beliefs are unlikely to have a high truth ratio.
But if such hypotheses are kept from being rejected and therefore ignored,
because of their potentially extreme consequences, one's behavioral deci-
sions may be prudentially more effective. Thus, an acceptance mechanism
with built-in sensitivity to extreme values may facilitate prudentially effec-
342 Acceptance and Uncertainty
formance of these mental mechanisms will probably lead to a very bad pay-
off. This kind of case can obviously be generalized. It then begins to look as
if the WTA process of acceptance has very bad properties, at least as judged
by a pragmatic standard.
There are at least two ways of countering this objection, granting (for ar-
gument's sake) the appropriateness of a pragmatic standard. First, an ac-
ceptance process does not necessarily preclude retention of probabilistic
information. As stressed in section 15.2, information about chances may be
accepted and therefore preserved for decision-making purposes. In the
driving example you could accept the proposition 'The chance that a car is
coming in the opposite lane is .ISy.The prevalance of WTA processes, then,
does not entail the suppression of all probabilistic information. Admittedly,
such preservation of information about chances involves belief contents
that invoke an objective probability notion, which is philosophically con-
troversial. But if sense cannot be made of objective probabilities, it is not
clear that the prudential criticism can be properly motivated in the first
place. Certainly the form of criticism advanced above relies on objective
probabilities.
Second, there is a possible mechanism, a complementary mechanism,
that would avert the apparent shortcoming of an acceptance process. Sup-
pose there is an allied cognitive mechanism that is sensitive to possible pay-
offs with extreme values: very rewarding or very threatening outcomes. This
mechanism increases the activation level of any contemplated hypothesis h
that is believed to make some extreme payoff likely. That is, h's activation
level is increased if the cognizer believes that some extreme payoff would
be likely if h were true. Such heightened activation would tend to prevent h
from being rejected, would keep it 'in contention7, and hence not ignored
during behavioral choice. In the foregoing example the threat of a fatal ac-
cident heightens the activation level associated with 'A car is coming7.This
keeps it from being rejected, and keeps its negation from being accepted.
Thus, the choice of the 'don't pass' option is promoted. (I do not offer any
specific model of how a choice would then be made. But such models could
readily be constructed.)
I suspect that human beings do have some such mechanism. Indeed, it
may be just the mechanism that accounts for wishful thinking and fearful
thinking. The present analysis shows how such a mechanism can have pru-
dential merit--even if it has attendant liabilities for reliability. If a cognizer
believes hypotheses largely because they would, if true, fulfill his fondest
dreams or deepest fears, these beliefs are unlikely to have a high truth ratio.
But if such hypotheses are kept from being rejected and therefore ignored,
because of their potentially extreme consequences, one's behavioral deci-
sions may be prudentially more effective. Thus, an acceptance mechanism
with built-in sensitivity to extreme values may facilitate prudentially effec-
chapter 16
Belief Updating
It is assumed that people have some way of assessing whether new evi-
dence is positive or negative and quantifying the strength of that evidence
(assigning some degree of confirmatory or disconfirmatory import to the
evidence). Einhorn and Hogarth make no specific suggestions about this
process, and I too shall simply take it as a given.
To illustrate the process, consider the effect of the first piece of negative
evidence, al, on one's position. Assuming that
346 Belief Updating
Now consider the functional relation between the strength of the anchor
and the adjustment weight (the "adjustment weight function"). It is as-
sumed that bigger anchors have larger adjustment weights. This implies
that the adjustment weight function is a monotonically increasing function
of the size of the anchor. To illustrate this, Einhorn and Hogarth posit a
simple form often found in psychophysical judgments, that is,
Disconfirm
prone
. Disconfirm
neutral
, Disconfirm
avoiding
Sk-1
Figure 16.1
The discount model implies that the strength of a belief after the kth piece
of negative evidence is a function of three factors: (1) the size of the anchor,
Sk-l: (2) the strength of negative evidence, s(ak); and (3) an adjustment
weight, which is a function of the anchor and one's attitude toward discon-
firming evidence (a).
A model for adjustments in response to new positive evidence is exactly
parallel to the discounting model. Einhorn and Hogarth call it the "accre-
tion" model. The form of the model is given by
where
rk-l = the adjustment weight for positive evidence
and
s(bk)= the strength of the kth piece of positive evidence, bk (0< s(bk)
s 1).
348 Belief Updating
Equation (6) follows the same general form as the discount model except
that the final position results from an anchoring and upward adjustment
process. The basic assumption in the accretion model is that weak beliefs
are increased more by positive evidence than are strong beliefs. The same
positive evidence 'helps' a weaker position more than a stronger one. As in
the discounting model, Einhorn and Hogarth posit a simple form to capture
the monotonically decreasing relation between rk-l and s ~ - In
~ .particular,
they propose the following relation:
rk- * = (1 - (P 2 0). (7)
The adjustment weight function is shown in Figure 16.2. The parameter P is
interpreted as reflecting one's "attitude toward confirming evidence."
When P > 1,confirming evidence is given a relatively small weight. In fact,
as increases, the weight for confirming evidence approaches 0.When 0 G
/3 < 1, confirming evidence is given a large weight, reflecting a "confirma-
tion prone" attitude. When P = 1, the attitude is labeled as "confirmation
neutral." Obviously, the role of P in the accretion model is directly analo-
gous to the role of a in the discount model. The full accretion model is ob-
tained by substituting (7) into (6):
Confirm
prone
Confirm
neutral
Confirm
avoiding
Figure 16.2
Belief Updating 349
0 1 2 3 k
Figure 16.3
starting points and the same attitude toward disconfirming evidence, diver-
gence can occur when the a's are not equal. Consider person C, whose ini-
tial position is the same as A's but whose a is very high. Since C gives little
weight to negative evidence, Figure 16.3 shows his or her beliefs to be un-
changed (or very slightly changed) over the k pieces. This leads to diver-
gence of beliefs between A and C given the same evidence. However, as k
increases, the two positions will converge toward zero unless C gives zero
weight to negative evidence. Analogous points about convergence and non-
convergence hold for the receipt of positive evidence under the accretion
model.
Moving to the consequences of mixed evidence, a major implication of
the contrast approach is strong recency efiects in belief change. To see why,
350 Belief Updating
consider the strength of belief after receiving one piece each of positive and
negative information. Consider the difference between the belief strength
after receiving them in the positive-negative order, S(+,-), as compared
with the strength after receiving them in the negative-positive order,
S(-,+). Figure 16.4 shows the effects of these two orders at different start-
ing levels, So, and So,. When the initial belief is at So,, the two orders are
shown in the top half of the figure. Compare the effects of the (+,-) and the
(-,+) orders. Note that the slope of the line connecting Sk-Oand Sk-l in
the (-,+) order is less steep than when negative evidence occurs in the
(+,-) order. The reason is that the same negative evidence has a larger
discounting weight after the positive evidence because of the contrast ef-
fect. Similarly, the slope of Skll to Ska2in the (-,+) order is steeper than
the slope for positive evidence from the initial position. These differences
in slopes lead to crossing lines that resemble fish tails. The fish-tail pat-
tern implies recency effects since the final position after the (+,-) order
is lower than for the (-,+) order. The prediction of recency effects does
not depend on one's initial position, as illustrated in the lower half of the
figure.
Another implication of the model concerns the relative effect of present-
ing multiple pieces of evidence simultaneously instead of sequentially. Si-
multaneous presentation of two or more pieces of negative evidence, for
example, results in greater discounting than presentation of these same
pieces one after another. Thus, the contrast or surprise effect is 'diluted' by
sequential processing.
Experiments done by Einhorn and Hogarth, and by J. D. Shanteau, sup-
port these order-effect implications. In two of Einhorn and Hogarth's own
experiments, subjects were given four scenarios, each scenario consisting of
an initial description (the "stem") and four additional pieces of information
Figure 16.4
Belief Updating 351
C
z
W
5
C3
n
3
>.
k
-r
m
a
m
0
LT
a
z
a
W
5
0 1 2 3 4 0 1 2 3 4
k k
Figure 16.5
by N. H. ~ n d e r s o nEinhorn
.~ and Hogarth proceed to incorporate the at-
tention-decrement factor into their theory.
I will not pursue the details of this theoretical refinement. Suffice it to say
that attention decrement works in the opposite direction from the contrast
or surprise factor in that the latter leads to recency effects, whereas atten-
tion decrement leads to primacy effects. The net effect of these opposing
EXP. 1 EXR 2
Simul t.
fq.
Figure 16.6
Belief Updating 353
Still, the core idea of making an epistemic judgment rest on successive ap-
plications of a method is present here too.
The diachronic approach to epistemic evaluation has also been popular-
ized by Imre Lakatos. He proposed that we appraise scientific research
programmes by their historical progressiveness.8 On Larry Laudan's per-
mutation of this idea a research tradition may be evaluated by the "general
progress," or the "rate of progress," of a research tradition in solving prob-
lems over time.'
I do not wish to endorse any of these specific proposals in the philosophy
of science literature. I merely call attention to the fact that the kernel of my
idea-judging a method by its repeated use--has enjoyed significant popu-
larity among philosophers.
An emphasis on diachronic properties of a process or method is also
urged by Robin Hogarth.lo Hogarth argues that the biases of judgmental
processes revealed in "discrete," isolated incidents may be outweighed by
the merits of those processes when they are used "continuously" (repeat-
edly) over time, with feedback.
Although I shall not develop the calibrational convergence approach in
further detail, I regard it as significant for two reasons. First, it brings into
prominence an important way of evaluating judgmental processes and
methods, namely, evaluating their success over time, through repeated ap-
plication. Second, it shows the versatility of the general reliabilist and cali-
brationist approach to epistemic norms. If the criterion of rightness
proposed in Chapter 5 does not strike one as fully correct, the verific ap-
proach still allows for variations in many directions. Calibrational conver-
gence is one such promising direction.
chapter 17
From the start, production systems have had an ambiguous status, being in
part programming languages for computer science and in part psychologi-
cal theories. I am interested in them, of course, in the latter guise. I shall
concentrate on perhaps the best developed psychological theory within the
production system tradition: that of John Anderson in The Architecture of
Anderson started his work in this mode in an earlier book, Lan-
guage, Memory, and Thought, where he called his system ACT."^ The
more recent version, which I shall examine, is called "ACT'."
To illustrate the idea of a production system, Anderson provides a hypo-
thetical set of productions, usable for doing addition. Let's see how such a
set of productions might be deployed to solve the following addition prob-
lem:
Figure 17.1. A representation of the flow of control among the various goals. The
boxes correspond to goal states and the arrows to productions that can change these
states. The goal at the origin of the arrow corresponds to the goal that elicits the
production, and the goal at the terminus of the arrow corresponds to the goal that
is set after the application of the production. Control starts with the top goal.
Production Systems and Second-Order Processes 363
to 0.Then production P6 sets the new subgoal to add the top digit of the
row (4) to the running total. In terms of Figure 17.1 this sequence of three
productions has moved the system down from the top goal of doing the
problem to the bottom goal of performing a basic addition operation. The
system has four goals stacked, with attention focused on the bottom goal.
At this point production P10 applies, which calculates 4 as the new value
of the running total. In doing this it retrieves from the addition table (in
+
LTM) the fact that 4 0 = 4. Production PI0 also pops the goal of adding
the digit to the running total. 'Popping' a goal means shifting attention from
the current goal to the one above it in the hierarchy. In this situation atten-
tion will return to iterating through the rows of the column. Then P7 ap-
plies, which sets the new subgoal of adding 8 to the running total. PI0
applies again to change the running total to 12, then P7 applies to create
the subgoal of adding 3 to the running total, then P11 calculates the new
running total as 15. At this point the system is back to the goal of iterating
through the rows and has processed the bottom row of the column. Then
production P9 applies, which writes out the 5 in 15, sets the carry to the 1,
and pops back to the goal of iterating through the columns. At this point the
production system has processed one column of the problem. This should
suffice to illustrate the application of these productions.
In Anderson's ACT" theory three memories are postulated: working
memory, declarative memory, and production memory. Working memory
contains the information that the system can currently access, consisting of
information retrieved from long-term declarative memory as well as tem-
porary structures deposited by encoding processes and the actions of pro-
ductions. Working memory is declarative knowledge, permanent or
temporary, that is in an active state. Declarative memory contains declara-
tive knowledge that may not be in an active state. Production memory con-
tains the set of productions, or procedures, available to the system. In a
mature human cognizer, Anderson puts the number of productions some-
where between the tens of thousands and the tens of millions.
The interactions between these memories is illustrated in Figure 17.2.
Most of the processes depicted here involve working memory. Encoding
processes deposit information about the outside world into working mem-
ory; performance processes convert commands in working memory into be-
havior. The storage process can create permanent records in declarative
memory of the contents of working memory and can increase the strength
of existing records in declarative memory. The retrieval process retrieves
information from declarative memory. In the rnatch process data in work-
ing memory are put into correspondence with the conditions of produc-
tions. The execution process deposits the actions of matched productions
into working memory. The whole process of production matching followed
by execution is referred to as production application. Note that the arrow
364 Production Systems and Second-Order Processes
APPL'CATloN
DECLARATIVE PRODUCTION
MEMORY
WORKING
MEMORY
ENCODING PERFORMANCES
T
OUTSIDE WORUI
Figure 17.2
marked 'application' cycles back into the production memory box, reflect-
ing the fact that new productions are learned from studying the history of
application of existing productions. This involves the idea of procedural
learning, which I shall explore in sections 17.3 and 17.4.
I will not summarize all the basic principles of ACT*, but let me mention
a few salient ones. Since all activities of productions require activation-
only activated data can be matched with production condition+an im-
portant question concerns the sources of activation. What Anderson calls
"source nodes" are the "springs" from which activation flows throughout
the system. Nodes can become source nodes in several ways. First, they can
be created by the perception of objects in the environment. Second, ele-
ments deposited in working memory by a production's action become tran-
sient source nodes. These are "internal" sources. They stay active for a brief
period of time and then cease to be sources..A special goal element, how-
ever, permits one internal source that is not so transient, and this enables
the system to maintain focus on a current goal.
A critical issue concerning the flow of cognitive activity is the selection of
productions that get matched. Given what has been said thus far, it would
seem possible for incompatible productions to match data in working mem-
ory, and for all to get applied. But this is assumed to be psychologically un-
realistic. Rival productions are not allowed to apply simultaneously. What,
Production Systems and Second-Order Processes 365
then, are the principles for conflict resolution, for choosing the production
that does get applied? ACT0 involves several such principles.
First, ACT0 allows the conditions of a production to be applied when its
condition is matched only partially, not fully. However, a full match is pre-
ferred to a partial match, and among partial matches, more complete
matches are preferred to less complete matches.
Any number of phenomena, Anderson says, point to the fact that produc-
tions can be evoked by partial matches. Faces change but still can be recog-
nized. We can recognize partial lexical patterns, as in Figure 17.3, and this
seems best explained by the concept of partial matching. Many errors de-
scribed by Donald or man' also can be explained by partial matching. For
instance, there is the case of a person who threw a dirty shirt by mistake
into the toilet rather than into the hamper. Apparently a 'hamper-match-
ing' pattern partially matched the toilet.
A second principle for production selection concerns the relative
strength of rival productions. Stronger productions are preferred to weaker
productions. The determinants of production strength will be examined in
section 17.3.
A third principle for production selection is the specificity principle. It
states that when two productions match the same data, preference is given
to the production with the more specific condition. Condition A is more
specific than condition B if A matches in a proper subset of the situations
where condition B would match. This principle handles cases of exceptions
to rules, as in productions P1 and P2 below.
P1 IF the goal is to generate the plural of a noun
THEN say the noun s.
P2 IF the goal is to generate the plural of man
THEN say men.
Production P1 gives the general rule for pluralizing a noun, while P2 gives
the exception for man. The specificity principle enables the exception to
take precedence over the general rule. There are two ways in which one
production can be more specific than another. One possibility is that the
more specific one contains additional clauses in its condition. The second is
that the production contains additional tests about a data structure, as in
the foregoing example.
A fourth principle concerning production selection involves goal domi-
nance. Many productions contain in their conditions tests for the current
Figure 17.3
366 Production Systems and Second-Order Processes
goal. There can be only one current goal active in the system at a time. Pro-
ductions that refer to the current goal take precedence over productions
that might otherwise apply. A goal-directed production that matches the
current goal applies more rapidly and reliably, but if it does not match the
current goal, it will not apply no matter how well the rest of its condition
matches. If a number of goal-directed productions match the current goal,
the principles of degree of match, strength, and specificity select the one
that will apply.
These are some of the main features of ACT*. More of its features will be
presented in section 17.3. Let me now turn (briefly) to some evaluative re-
flections on the features of ACT* introduced thus far.
affect when the productions are applied). Clearly, all of these processes are
of interest to primary epistemology. In this section I shall reflect on (some
of) the processes in ACT" for using productions.
Processes for using methods are included under the category of 'first-
order' processes. Processes for acquiring new methods, and for deleting or
modifying them in various ways, would fall under the category of 'second-
order' processes (see sections 5.3 and 5.7). Thus far I have only presented
first-order processes of ACT". So in this section my evaluative comments on
production use are confined to such first-order processes. An exposition of
ACT"'s second-order processes will be given in the next section; an evalua-
tion of them will appear in section 17.4.
Let us turn now to the reliability of the processes for using a repertoire of
productions. This includes the processes for selecting which productions to
apply on given occasions as well as the process of application itself. Of
course, the reliability of selection and application processes cannot be com-
pletely isolated from the reliability of the productions on which they oper-
ate. In the interest of simplicity, though, let us assume that all productions
in a person's repertoire are perfectly reliable. How, then, would selection
and application processes affect reliability?
One important point to note here is the partial matching feature of
ACT". Even if all the productions are perfectly reliable, error can creep in
through partial matching. A new belief can be formed in working memory
by application of a production that does not really fulfill a preexisting con-
dition. This amounts roughly to taking a sound inference procedure and
misapplying it.
A partial matching process will typically reduce overall reliability. Does
this mean that any cognitive architecture that realizes ACT0 has a feature
that precludes its approval by a right J-rule system? Not necessarily. This
will depend on two factors: the degree of reliability required for justified-
ness, and the degree of match that constitutes the minimal threshold for
production application. In section 17.1 I stressed the comparative principle
for production selection: ceteris paribus, productions with a higher degree
of match are preferred to productions with a lower degree of match, But
ACT" also includes a cutoff feature: below a certain level of match, a pro-
duction will not be selected at all. Obviously, this cutoff level is important
for reliability.
However, even if the cutoff level of a given architecture allows too much
error, as judged by a rightness criterion with a fixed truth-ratio require-
ment, this does not mean that no belief formed under this architecture is
justified. When productions are selected in virtue of a sufficiently high de-
gree of match-high enough to satisfy the specific rightness criterion-then
beliefs so formed may be justified. In those cases the minimal cutoff point is
not the causally operative degree of match. (Recall my discussion of this
sort of case in Chapter 3.)
368 Production Systems and Second-Order Processes
ing how subjects plan a series of errands around town, these researchers
found evidence of the subjects' proceeding unsystematically and multi-
directionally in their planning space, seeking to develop the most promising
aspects of their plan as the occasion permits. Subjects thereby mix low-level
and high-level decision making, instead of proceeding in a strictly hierar-
chical fashion.'Anderson tries to capture such possibilities in ACT*. If the
resulting procedures are indeed psychologically correct, it is an open ques-
tion how they affect the general power of the ACT0 architecture. Clearly,
however, the goal-linked properties of this architecture are importantly
connected to the epistemically significant dimension of question-answering,
or problem-solving, power.
tion, the following example assumes that the person has compiled some
special-case production rules from experience. Two such rules might be:
P1 IF I am playing no trump
and my dummy has a long suit
THEN try to establish that suit
and then run that suit.
P2 IF I am playing spades
and my dummy has a long suit
THEN try to establish that suit
and then run that suit.
ACT* will create a generalization of these productions:
P3 IF my dummy has a long suit
THEN try to establish that suit
and then run that suit.
It is certainly useful to generalize overly specific procedures. But it is
even more necessary to restrict the range of overly general procedures. The
discrimination process tries to restrict the range of application of a produc-
tion-actually, construct a new production-to just the appropriate cir-
cumstances. This requires that the cognizer have examples of (ostensibly)
correct and incorrect applications of the production. The discrimination
process remembers and compares the values of the variables in the correct
and incorrect applications. It chooses a variable for discrimination among
those that have different values in the two applications. Having selected a
variable, it looks for some attribute that the variable has in only one of the
situations. A test is added to the condition of the production for the pres-
ence of this attribute.'
The feature selected for discrimination is determined by comparing the
variable bindings in the successful and unsuccessful production applica-
tions. A variable is selected on which the applications differ, and features
are selected to restrict the bindings. Anderson says that the probability of
choosing the right feature to discriminate upon depends on the similarity of
the successful and unsuccessful situations, The more similar they are, the
fewer the distracting possibilities, and the easier it is to identify the critical
feature. This is the idea of near misses, discussed for example by P. H. Win-
ston.''
A prerequisite for discrimination is that the system have feedback indi-
cating that a particular production has misapplied. In principle, Anderson
says, aproduction application could be characterized as being in one of
three states-known to be incorrect, known to be correct, or correctness
unknown. However, the mechanisms implemented in ACT* do not distin-
guish between the second and third states. If a production applies and there
Production Systems and Second-Order Processes 373
fewer demands are made on working-memory space. This allows that space
to be used for other, concurrent cognitiye tasks, thereby increasing overall
power.
The impact on reliability is somewhat mixed, although in the long run it
appears to be positive. Anderson notes that the transition phase from de-
clarative to procedural embodiment of a skill often breeds the most errors.
But these are eventually corrected. Moreover, he emphasizes that heavy use
of working-memory space produces errors, so the bypassing of working
memory effected by proceduralization should increase reliability.
The fact that new productions can only be compiled when their com-
ponents have already been applied successfully is another feature of the
ACTo processes conducive to metareliability. It means that merely arbi-
trary will not be constructed. Nor will they be constructed
just from verbal instructions, which might possibly be misunderstood or
mistaken.
I turn now to the production-building processes of generalization and dis-
crimination. Both of these processes can sometimes produce incorrect,
meaning unreliable, productions. Overgeneralization is a well documented
phenomenon in the language acquisition literature. It occurs with both
syntactic rules and natural-language concepts. Discrimination can also fail
by introducing inappropriate restrictive conditions, as discussed in section
17.3.
Strength-adjustment processes are ones by which such errors can be miti-
gated. If overgeneralizations or misdiscriminations are detected, that is, if
negative feedback follows their application, then they are reduced in
strength by 25 percent. Also, in many such cases, new productions are in-
troduced to handle (more correctly, it is hoped) the same tasks. With these
rival productions in place, the old mistaken productions tend to lose out and
not get applied. Such disuse leads to gradual decay.
Is it clear, though, that strength-adjustment processes can adequately
compensate for the errors of production-creation processes? This depends
partly on just how meta-unreliable the production-creation processes are,
and partly on the effectiveness of feedback processes.
Mfiththese questions in mind, consider the process of discrimination. Ac-
cording to Anderson's description, the discrimination algorithm randomly
chooses a variable for discrimination from among those that have different
values in successful and unsuccessf+ applications. It then looks for some at-
tribute that the variable has in only one of the situations, a successful situa-
tion, and adds this condition to the production. But it appears that correct
identification of such an attribute is largely a chance affair. True, a discri-
minated production yields the same belief ('action') as the original produc-
tion in the restricted situation; so the discrimination process does not
impose an unreliable production over a reliable one. But it is problematic
whether improvements will be secured very often.
Production Systems and Second-Order Processes 3 77
I HAVE NOW completed the task of Part 11: to provide some detailed speci-
mens of primary epistemology or primary epistemics. Needless to say, these
chapters do not purport to be a final embodiment of primary epistemics.
They are, as billed in section 9.1, only first steps in the subject. The enter-
prise should of course continue with new developments in a number of di-
rections.
It is a bit misleading to call the present chapters 'first' steps in primary
epistemics. As indicated in the Introduction, there are earlier conceptions
and traditions with which primary epistemics has close affiliations, in-
cluding the psychologistic epistemologies of the seventeenth through the
nineteenth centuries and W. V. Quine's conception of "naturalistic episte-
mology." However, only the accomplishments and prospects of contempo-
rary cognitive science make feasible a serious execution of primary
epistemics. Although the general conception of primary epistemics could
have been advanced in previous centuries (and was indeed adumbrated by
Locke and Hume), comparable opportunities for scientifically grounded
analysis of mental operations were not then available.
What further developments should be expected in primary epistemics?
First and most obvious, developments are to be expected in substantive
cognitive sciencwhanges in theoretical posture as well as refinements in
empirical detail. The 1980s are witnessing new approaches to memory, and
these can be expected to burgeon. Linkages to neural constructs are bound
to be strengthened. Whether such a schema as the production system
framework continues to be fruitful only time will tell. Whether the mind is
importantly modular is equally a matter for future cognitive science to de-
termine. New developments in cognitive theory mean changes in the po-
sited 'iepertoire of fundamental operations, or changes in the precise
properties of these operations. Such changes could require significant revi-
sions in empirical material of the kind discussed in Part 11.
Second, new developments should be encouraged in primary epistemics
Conclusion 379
even with a completely static cognitive science. After all, epistemics un-
dertakes to draw evaluative conclusions about cognitive processes. Even if
cognitive theory is held fixed, considerable work must be done in assessing
the reliability, power, and speed of the components of the cognitive reper-
toire. There is the task of determining what right J-rule systems, if any, can
be devised for this repertoire. Can some (or all) of the processes belonging
to this repertoire be deployed in a way that satisfies a specified truth ratio
criterion? Are there more than one set of processes that can do this? If so,
what are they? Even holding a posited cognitive repertoire constant, an-
swers to these questions are not so easy to give; and revisions in putative an-
swers are naturally to be expected.
Third, developments are needed in the formulation of evaluative criteria.
My own proposals, which have been deliberately flexible, would profit from
some tightening. As an example, let me review my treatment of justified-
ness. I began in Chapter 5 by proposing a family of alternative rightness
criteria for belief-forming processes. My calibrational approach to degrees
of credence, in Chapter 5, was also rather exploratory. Further proposals for
a calibrational approach to degree of credence were advanced in Chapter
16, but these too were quite tentative. Finally, Chapter 17 surveyed alter-
native criteria for second-order processes--all involving metareliability-
but none was definitely endorsed. The criteria for J-rule rightness, then,
clearly warrant further attention.
Looking beyond criteria of justifiedness, more detailed standards of eval-
uation may be proposed for other dimensions of evaluation, such as intelli-
gence or rationality. I have made a few suggestions on these matters, but
more remains to be done. Again, while I have expressed doubts about a
unifid criterion of epistemic appraisal, perhaps other theorists could in-
vent and defend such a measure, integrating such standards as reliability,
power, and speed. This would certainly be within the spirit of the general
enterprise.
Proposals for revising criteria of evaluation can come from different .
sources. One is general reflection on the ordinary use of the relevant evalua-
tive terms. Another might be progress in empirical psychology, which
highlights factors that could be missed in purely abstract reflection. An ex-
ample was given in Chapter 17, in the discussion of rightness criteria for
second-order processes. We had earlier supposed that rightness criteria
would concern only the acquisition and retention of methods. But consider-
ation of ACT* drew our attention to the strength variable. It became plau-
sible to suppose that a satisfactory J-rule criterion for second-order
processes might invoke some factor like strength, or probability of employ-
ment.
Another avenue of development lies in possible applications of primary
epistemics. I have presented the evaluative enterprise as a 'theoretical'
380 Conclusion
rather than a 'regulative' one. But there are some prospects for deliberate
control of cognitive operations, should this prove advisable. Habits in de-
ployment of the cognitive repertoire may be amenable to inculcation and
training. There may be techniques for promoting the use of certain se-
quences or patterns of operations over others. If primary epistemics distin-
guishes superior from inferior processes, it is natural to try to promote the
better over the worse. A challenge is then extended to educational theory to
devise techniques for achieving this end.
There is, then, a substantial agenda for the future of primary epistemics,
and for offshoots from it. Exciting prospects lie ahead for a collaboration
between philosophy and a variety of cognitive disciplines.
In recent years philosophers have laid the groundwork for a working rela-
tionship between cognitive science and the philosophy of mind. The argu-
ment here is that cognitive science is equally important (though in different
ways) to epistemology. However, my conception of epistemology is even
broader than this. At the social level it would intersect with yet other disci-
plines outside of philosophy proper. A full argument for this thesis, how-
ever, must await a future volume.
Notes
Illustration Credits
Author Index
Subject Index
Notes
Introduction
1. Ludwig Wittgenstein, Philosophical Investigations, trans. C . E. M. An-
scombe (New York: Macmillan, 1953), para. 309.
2. W. V. Quine, "Epistemology Naturalized," in Ontological Relativity and
Other Essays (New York: Columbia University Press, 1969), p. 82.
3. Donald Campbell, "Evolutionary Epistemology," in Paul Schilpp, ed., The
Philosophy of Karl Popper, vol. 1 (La Salle, Ill.: Open Court, 1974), p. 413.
4. Two other recent works that treat epistemology in some sort of psycholo-
gistic vein are Gilbert Harman's Thought (Princeton: Princeton University Press,
1973) and Fred Dretske's Knowledge and the Flow of Information (Cambridge,
Mass.: MIT Press, 1981). But the approaches of these authors are quite different
from mine. For example, instead of drawing on psychology to reach epistemolo-
gical conclusions, Harman's main strategy is to use epistemological judgments to
reach psychological conclusions. Specifically, he seeks to decide what inferences
people make by seeing what inferences must be postulated to account for our
judgments of whether or not they 'know' (see Thought, pp. 20-23.) This seems to
me to get the relation between psychology and epistemology backward. Dretske
refines a popular term of cognitive theory-'informationJ-and puts it to episte-
mological use. But his theory assigns no systematic role to empirical psychology;
nor does he conceive of epistemology as an evaluative discipline. '
5. Although it is not clear that Quine intends to neglect the evaluative di-
mension, his actual characterizationsof naturalistic epistemology do not expressly
introduce this dimension. See, for example, "The Nature of Natural Knowledge,"
in Samuel Guttenplan, ed., Mind and Language (Oxford: Clarendon Press, 1975).
6. Otto Neurath, "Protokollsatze," Erkenntnis, 3 (1932):206.
7. Francis Bacon, N m m Organum (1620), bk. 1, sect. 2.
8. John Locke, An Essay concerning Human Understanding (1690), intro-
duction, sect. 4.
9. David Hume, A Treatise of Human Nature (1739), introduction.
10. On Carnap's sharp separation, see The Logical Syntax of language, trans.
Amethe Smeaton (London: Routledge and Kegan Paul, 1937), sect. 72, esp. pp.
278-279. Reichenbach's distinction appears in Experience and Prediction (Chi-
cago: University of Chicago Press, 1938), pp. 6-8.
384 Notes to Pages 7-28
2. Skepticism
1. A classification of types of skepticism is given by George Pappas in "Some
Forms of Epistemological Scepticism," in George Pappas and Marshall Swain,
eds., Essays on Knowledge and Justification (Ithaca: Cornell University Press,
1978). For a historical overview of skepticism, see Richard Popkin, "Skepticism,"
in Paul Edwards, ed., The Encyclopedia of Philosophy, vol. 7 (New York: Macmil-
lan, 1967). A critical survey of skeptical arguments is given by Nicholas Rescher
in Sceptic- (Totowa, N.J.: Rowman and Littlefield, 1980).
2. Sextus Empiricus, Outlines of Pywhonism, trans. R. G. Bury (London:
Heinemann, 1933).
Notes to Pages 30-40 385
21. It should also be noted that certain questions about knowledge and justifi-
cation can be studied with little or no concern for skepticism. For example, how
should knowledge be analyzed in a way that meets the Gettier problems? This
question about knowledge has little connection to skeptical concerns. Similarly,
there is the question: what is the structure of epistemic justification? Is it foun-
dationalist or coherentist? This problem can also be investigated with little con-
cern for skepticism. It may be assumed that we do have justified beliefs in some
fashion or other; it just remains to be shown how that justification arises. I
have not stressed these points because they will not be central to the continuing
discussion.
3. Knowledge
1. Some people object to using 'cause' in the analysis of knowledge because
that notion is itself philosophically problematic. But we can make progress in un-
derstanding one concept by using others, even if the latter also invite analysis. In
particular, it is useful to analyze epistemological notions in terms of nonepiste-
mological ones, as in the present case. One comment on my view of causation is in
order. I do not presume that causation implies determinism. Hence, while knowl-
edge cannot be attained if beliefs are wholly uncaused, knowledge is possible in a
nondeterministic universe.
2. D. M. Armstrong, Belief, Truth and Knowledge (Cambridge: Cambridge
University Press, 1973). Also see Marshall Swain, Reasons and Knowledge (Ithaca:
Cornell University Press, 1981).
3. I will not reproduce all the details of Armstrong's account, which gets
rather intricate. (See Belief, Truth and Knowledge, pp. 192ff.) But I do not believe
that any of his qualifications ultimately save his approach.
4. Armstrong gives a somewhat similar case, attributed to Ken Waller. (Cf.
Belief, Truth and Knowledge, pp. 178-179.) But my case is a bit different, and is
not (I believe) open to the same rejoinder as he gives to Waller's. For more on the
reliable-indicator versus reliable-process approach--applied, however, to the the-
ory of iustification-see Frederick Schmitt, "Justification as Reliable Indication
or Reliable Process," Philosophical Studies, 40 (1981):409-417.
5. Frank Ramsey, The Foundations of Mathematics and Other Logical Essays
(London: Routledge and Kegan Paul, 1931).
6. These terms are borrowed from Colin McGinn, "The Concept of Knowl-
edge," in Peter French, Theodore Uehling, Jr., and Howard Wettstein, eds., Mid-
west Studies in Philosophy, vol. 9, Causation and Causal Theories (Minneapolis:
University of Minnesota Press, 1984). In earlier drafts of this chapter I used the
phrases 'generic' and 'situation-specific' reliability. But I prefer the 'global' and
'local' terminology.
7. Robert Nozick, Philosophical Explanations (Cambridge, Mass.: Harvard
Universip Press, 1981), chap. 3.
8. Fred Dretske, "Conclusive Reasons," The Australasian Journal of Philoso-
phy, 49 (1971):l-22. Also see his Knowledge and the Flow of Information (Cam-
bridge, Mass.: MIT Press, 1981). It is not clear that Dretske's theory should be
viewed as a specimen of the reliable-processapproach since it places little impor-
tance on the causal production of belief.
Notes to Pages 46-52 387
!24. For more on the psychology of procedure acquisition, see Chapter 17.
25. Further discussions of this topic appear in section 5.7 and Chapter 17.
26. Cf. Gettier, "Is Justified True Belief Knowledge?'
27. Keith Lehrer and Thomas Paxson, Jr., "Knowledge: Undefeated Justified
True Belief," The Journal of Philosophy, 66 (1969):225-237.
28. Fred Dretske, "Epistemic Operators," The Journal of Philosophy, 67
(1970):1007- 1023, and Nozick, Philosophical Explanations, pp. 204-21 1, 227-
229.
29. "The Concept of Knowledge," pp. 542ff.
30. See William Alston, "Level-Confusions in Epistemology," in Peter
French, Theodore Uehling, Jr., and Howard Wettstein, eds., Midwest Studies in
Philosophy, vol. 5, Studies in Epistemology (Minneapolis: University of Minnesota
Press, 1980).
31. W. V. Quine, "The Nature of Natural Knowledge," in Samuel Guttenplan,
ed., Mind and Language (Oxford: Oxford University Press, 1975), p. 68.
in Joel Feinberg, ed., Reason and Responsibility, 5th ed. (Belmont, Calif.: Wads-
worth, 1981); and Walter Stace, "Ethical Relativism," in ibid.
For an incisive critique of epistemological cultural relativism, see Hilary Put-
nam's discussion of Rorty's brand of relativism, in "Why Reason Can't Be Natu-
ralized," in Realism and Reason (Cambridge: Cambridge University Press, 1983).
24. See John Pollock, "A Plethora of Epistemological Theories," in George
Pappas, ed., Justijication and Knowledge (Dordrecht: D. Reidel, 1979).
25. Admittedly, I disclaim any general association between judgments of jus-
tifiedness and praise or blame (see Chapter 1, note 6). But we needn't exclude
praise and blame entirely. At any rate, casting the issue in these terms is a useful
device for crystallizing intuitions about the kind of case cited.
26. Philosophy and the Mirror of Nature, p. 170.
27. The distinction between primary and seconcZury justification will be ex-
pounded in section 5.3.
28. See Gilbert Harman, Thought (Princeton: Princeton Upiversity Press,
1973), pp. 27-29.
29. The only exceptions are some proposals on 'secondary' justifiedness.
30. A good exposition of the essentials of foundationalism (plus one kind of
variant on the main theory) can be found in William Alston, "Two Types of
Foundationalism," The Journal of Philosophy, 73 (1976):165-185.
31. See Roderick Firth, "Coherence, Certainty, and Epistemic Priority," The
Journal of Philosophy, 61 (1964):545-557, and Mark Pastin, "Modest Foundation-
alism and Self-warrant," American Philosophical Quarterly, monogr. ser. no. 4
(1975):141-149.
32. Keith Lehrer, Knowledge (Oxford: Oxford University Press, 1974), pp.
78-79.
33. Philosophy and the Mirror of Nature, for example, pp. 163,211-212.
34. This possibility has not been raised in the literature, to my knowledge.
35. I will examine some empirical evidence on this score in section 9.4.
13. On this point see Roderick Firth, "Epistemic Merit, Intrinsic and Instru-
mental," in P~oceedingsand Addresses of the Anerican Philosophical Association,
55 (1981):s-23.
14. Keith Lehrer's coherence theory of justification, in Knowledge (Oxford:
Oxford University Press, 1974), seems to be a specimen of weak coherentism.
With one qualification, to be explained below, Lehrer's theory says that a belief is
justified if the proposition believed is viewed by the cognizer as having a better
chance of truth than its competitors (see p. 212). But when a person believes any
proposition, he views it as having a better chance at truth than its competitors (at
least the competitors he thinks of). So even an idle dreamer's beliefs will satisfy
this condition. Lehrer's theory does contain the qualification that the belief
should be viewed from the cognizer's 'corrected' doxastic system, which is ob-
tained by deleting beliefs the cognizer would not retain as a "veracious inquirer,"
that is, as an "impartial and disinterested truth-seeker." But suppose a fantasizer
honestly believes--again, through fantasy-that fantasizing is the optimal
method for getting the truth. Then he is genuinely seeking truth (impartially and
disinterestedly); so his actual belief system is not subject to any correction. But
then the "veracious-inquirer" condition does no real work, and we are left with
the counterintuitive result obtained earlier.
Perhaps Lehrer would say that what counts as veracious belief formation must
not be judged by the inquirer's own lights, but by the facts of the matter. In that
case dreaming and fantasizing would not be methods of a veracious inquirer, and
this man's doxastic system would require drastic correction. But if this is the
proper interpretation of the view, it comes out more as closet reliabilism than as
subjective coherentism. Reference to the 'corrected doxastic system' really masks
a thoroughly reliabilist viewpoint.
Lehrer no longer subscribes to the view expressed in Knowledge. His more re-
cent position comes somewhat closer to reliabilism. This position is formulated in
"A Self Profile," in Radu Bogdan, ed., Keith Lehrer (Dordrecht: D. Reidel, 1979),
pp. 84-85. There he characterizes 'complete' justification as the conjunction of
'personal justifications-his old notion-plus 'verific justification'. Roughly, a be-
lief in p is 'verifically' justified just in case it coheres with the cognizer's true be-
liefs. But this notion of verific justification-and hence the notion of complete
justification-is excessively strong. A belief inferred from false premises is in dan-
ger of being unjustified solely because the premises are false. But, surely, the mere
falsity of previous beliefs should not endanger justifiedness. If believing those
premises was justified, and if the inferences were appropriate, then the resulting
beliefs should also be justified. Falsity of the premises endangers knowledge, but
not justifiedness. Hence, this strategy for introducing truth into the justification
picture does not work.
15. See Richard Jeffrey, The Logic of Decision, 2nd ed. (Chicago: University of
Chicago Press, 1983), chap. 11.
16. Lewis's term for this stronger notion of coherence is "congruence"; see An
Analysis of Knowledge and Valuation (La Salle, Ill.: Open Court, 1946), p. 338.
Chisholm's term for the same concept is "concurrence"; cf. Theory of Knowledge,
1st ed. (Englewood Cliffs, N.J.: Prentice-Hall, 1966),chap. 3; 2nd ed. (Englewood
Cliffs, N.J.: Prentice-Hall, 1977), chap. 4.
Notes to Pages 100-1 04 393
17. John Pollock also speaks of negative and positive coherence, in "A Pleth-
ora of Epistemological Theories," in George Pappas, ed., Jtcstijicationand Knowl-
edge (Dordrecht: D. Reidel, 1979). But his terminology marks a somewhat
different distinction.
18. See Carl Hempel, "Deductive-Nomologicalvs. Statistical Explanation," in
Herbert Feigl and Grover Maxwell, eds., Minnesota Studies in the Philosophy of
Science, vol. 3 (Minneapolis: University of Minnesota Press, 1962); Jaakko Hin-
tikka and J. Pietarinen, "Semantic Information and Inductive Logic," in Jaakko
Hintikka and Patrick Suppes, eds., Aspects of Inductive Logic (Amsterdam:
North-Holland, 1966);Isaac Levi, Gambling with Truth (New York: Knopf, 1967);
Risto Hilpinen, Rules of Acceptance and Inductive Logic (Amsterdam: North-
Holland, 1968); and Lehrer, Knowledge.
19. See Levi, Gambling with Truth, chap. 4 .
20. My framework principle, (P3*), employs a categorical notion of justifica-
tion, so I cannot easily incorporate this point into the current formulation of the
theory. But this feature could probably be revised to accommodate degrees of
justifiedness.
21. Apart from accounting for intuitions in all these cases, the truth ratio, or
reliabilist, theory of justifiedness holds out the prospect of coping satisfactorily
with the Humean riddle of induction. Although it does not fit naturally at this
juncture, let me sketch an approach to this problem, predicated on the correct-
ness of the reliabilist theory.
I propose to divide the traditional problem of induction into two problems.
The first is: Can we have justified beliefs based on inductive inference? The sec-
ond is the problem of second-order justifiedness: Can we have a justified belief
that we can have justified beliefs based on inductive inference? I think it quite
possible that Hume himself, and many followers of Hume, confused the first
problem with the second. At least they have held that a positive answer to the
first depends upon a positive answer to the second. I disagree with this conten-
tion. Nonetheless, the second problem is significant; and there is no reason why
one could not take the conjunction of these problems as the problem of induction.
If the reliability theory is correct, the first problem is not terribly serious. At
least it is 'logically' possible to have justified beliefs based on inductive inference,
whether or not it is humanly possible. Suppose we have a (basic) inductive process ,
with a sufficiently high truth ratio, at least a process whose output beliefs would
be true often enough if its input beliefs were true. Then this process could be
permitted by a right system of J-rules. Hence, beliefs generated by this process
could be justified, as long as the input beliefs to the process were justified. A criti-
cal point here is that our framework principle imposes no iterative requirement.
A person's belief is justified if it is permitted by a right rule system (I ignore the
no-undermining clause for simplicity); the person does not also have to be justi-
fied in believing that the belief is so permitted.
This takes care of the first problem of induction. But what about the second? If
we stick to the framework principle, there is no reason why a belief in the justifi-
cation-conferring power of induction could not be justified as a result of the self-
same inductive process! If the indicated inductive process is permitted by a right
rule system, then one might apply that same permitted process to beliefs in the ,
394 Notes to Page 104
process's past successes, and draw the conclusion that the process is successful
(reliable) in general. From this one could permissibly infer that the process is per-
mitted by right rules.
Of course, this inductive justification of induction is staunchly resisted by de-
fenders of Hume. The familiar charge is one of circularity. But how would this
charge be formulated in terms of my general theory? It would have to be couched
as a further restriction on justifiedness, to be incorporated into a new framework
principle. It might run: 'No belief about the permissibility of a process is justified
if the belief results from that selfsame process'. But does this restriction have any
plausibility? Surely not. It seems quite arbitrary. Would one say that a person
could not be justified in believing in the validity of modus ponens if he used
modus ponens to amve at this belief? (Here I waive the point that modus ponens
isn't really a process.) Acceptance of this stricture leads to the dilemma posed by
Lewis Carroll, in "What the Tortoise Said to Achilles," in Mind, vol. 4 (1895), pp.
278-280.But there is no good reason to accept it. If a process deserves to be per-
mitted, then its permission should extend to all subject matter, including its own
performance and its own permissibility.
Another attempt to address the problem of induction in a reliabilist vein is
made by James Van Cleve, in "Reliability, Justification, and Induction," in Peter
French, Theodore Uehling, Jr., and Howard Wettstein, eds., Midwest Studies in
Philosophy, vol. 9, Causation and Causal Theories (Minneapolis: University of
Minnesota Press, 1984). Van Cleve has an instructive discussion of circularity. He
says that an argument is viciously circular only if it is epistemically circular. He
then distinguishes premise circularity and rule circularity, as possible sources of
epistemic circularity. He indicates that the critical argument in the case of in-
duction does not suffer from premise circularity, and that rule circularity is not
really a source of epistemic circularity. I will not rehearse these arguments but
commend them to the interested reader.
22. This phrase is used by Hilary Kornblith, in "Justified Belief and Epistemi-
cally Responsible Action," The Philosophical Review, 92 (1983):33-48. However,
Kornblith uses it to express a different theory than the one intended in the text.
Kornblith's own theory is that a belief is justified just in case it is guided by a de-
sire to believe the truth. Such a theory is extremely weak, though. If a cognizer is
sufficiently benighted, such a desire may lead to very weird processes or 'evi-
dencecollecting' activities. It is questionable whether beliefs formed from such
processes or activities are justified. Furthermore, the theory may also be too
strong. Perceptual beliefs may presumably be justified; yet such beliefs are not
controlled or guided by a desire to believe the truth:
Let me return to the theory intended in the text: a truth-ratio maximization
theory, relativized to human cognitive endowments. There are various possible
refinements of this approach. Suppose, for example, that the best human percep-
tual processes can secure a truth ratio of .99, but the best human inductive pro-
cesses can only secure a truth ratio of .65. The simple maximizing criterion might
then disallow rule systems that sanction inductive processes, on the ground that
inductive processes would reduce the truth ratio below an otherwise attainable
maximum. This is (arguably) counterintuitive. However, the simple maximizing
criterion might be replaced by a more sophisticated criterion, which maximizes
within a category. For example, a criterion could license the best perceptual be-
Notes to Pages 104-1 07 395
lief-forming processes, the best inductive belief-forming processes, the best me-
morial belief-forming processes, and so on. This approach would encounter some
difficulties--how, exactly, should the categories be selected?-but it is still at-
tractive.
23. As formulated, (ARI) has a problem. It presupposes that each J-rule system
has a unique truth ratio. This is plausible if we consider only actual instantiations
of the permitted processes (by all cognizers, past, present, and future), assuming
these are finite. But presumably a truth ratio should be fixed by possible as well as
actual instantiations. For one thing, we want to allow a system to have a nonzero
truth-ratio propensity even though its authorized processes are never actually in-
stantiated. However, this makes the assumption of a unique truth ratio extremely
doubtful. There are many possible ways of instantiating a permitted set of pro-
cesses. Because of the permissive nature of the rules, one pattern of instantiation
might use a given subset of permitted processes (for example, inductive processes)
very often, while another pattern of instantiation might use this subset sparingly.
Let us call any pattern of permitted instantiations of a given system a compliance
profile of that system. The problem is that it is most unlikely that all compliance
profiles of a system have the same truth ratio. But if a rule system's truth ratio is
not inuuriant under diverse compliance profiles, the system does not have a
unique truth ratio.
One way to handle this problem is to require, for rightness, that all possible
compliance profiles of the target system meet the threshold truth ratio. (An ex-
ception could be made for the 'null' profile, which does not use any of the per-
mitted processes. This is a compliance profile in the sense that it does not uiolate
the rule system. But the truth ratio of this profile is 0.)This is a very stringent re-
quirement. A milder way of handling the problem is to require only that the me-
dian truth ratio of all the compliance profiles meet the threshold. But this seems
too weak. The worst truth ratio might be too low.
A third solution is to adopt the previous condition as necessary, but add a fur-
ther requirement that the amplitude of departures from the median value must
not be too great. To borrow a term from Brian Skyrms, the rule system should be
quite resilient. (Cf. Causal Necessity, New Haven: Yale University Press, 1980, p.
11.However, Skyrms's notion of resiliency is defined for probabilities.) Obviously,
many other measures of these sorts could be proposed, so the problem appears sol-
uble. But I will not try to pick a particular solution; I will simply assume that
(ARI) can be reformulated with a satisfactory solution.
Notice that similar problems would arise for resource-relative criteria, such as a
maximizing criterion. In fact, here the problems would be trickier because one
would have to compare various systems, each with multiple compliance profiles.
One possible solution is a 'maximax' solution: the system with the highest maxi-
mum compliance profile is best. Another possible solution is a 'maximin' condi-
tion: the system with the highest minimum compliance profile is best. Again, I
will not try to pick any particular solution from these candidates.
24. I assume that a system's truth ratio is relatively invariant across different
normal worlds.
25. Saul Kripke, Naming and Necessity (Cambridge, Mass.: Harvard Univer-
sity Press, 1980), and Hilary Putnam, "The Meaning of 'Meaning'," in Mind,
Languuge and Reality (Cambridge: Cambridge University Press, 1975).
396 Notes to Pages 107-1 14
v(m) = 1,
v(x) = 0 for x # m.
Notes to Pages 114-1 26 397
In other words, where the relative frequency of rain is -20,the probability of this .
forecaster making a 20 percent chance of rain prediction on a given day is 1. This
man is a least refined, though well-calibrated, forecaster. Contrast him with an-
other well-calibrated forecaster, characterized by the following probability func-
tion:
v(1) = m,
v(0) = 1-m,
v(x) = O f o r x Z 0 , 1.
The only probabilities of rain that this forecaster ever specifies are 0 and 1, and
since he is well calibrated, his predictions are always correct. This man is referred
to as a mst-refined forecaster. (In meteorology, the first forecaster is said to ex-
hibit zero sharpness and the second is said to exhibit perfect sharpness.)
Now it clearly seems too strong to require, for justificational rightness, that a
rule system yield mariml refinement as well as good calibration. But some
threshold of refinement might be set, as a further condition for rightness beyond
good calibration. This seems feasible since DeGroot and Fienberg develop a con-
dition for determining whether one well-calibrated forecaster is more refined
than another.
For another attempt to use calibration to explicate epistemic standards, see Bas
van Fraassen, "Calibration: A Frequency Justification for Personal Probability,"
in R. Cohen and L. Laudan, eds., Physics, Philosophy and Psychoanalysis (Dor-
drecht: D. Reidel, 1983).
35. Hilary Putnam, "Why Reason Can't Be Naturalized," in Realism and
Reason (Cambridge: Cambridge University Press, 1983), p. 231.
36. Roderick Firth, "Epistemic Merit, Intrinsic and Instrumental," p. 19.
37. Ibid., p. 17.
8. Nuel Belnap, Jr., and Thomas Steel, Jr., The Logic of Questions and An-
swers (New Haven: Yale University Press, 1976).
9. For further treatment of the semantics of questions and answers, see
Jaakko Hintikka, "The Semantics of Questions and the Questions of Semantics,"
Acta Philosophical Fennica, 28 (1976), issue 4.
10. Discussions of problem solving in artificial intelligence portray it as get-
ting to an end state from some initial state. For surveys of this topic, see Patrick
Winston, ArtificMl Intelligence (Reading, Mass.: Addison-Wesley, 1977); Nils
Nilsson, Principles of Artijicial Intelligence (Palo Alto, Calif.: Tioga, 1980); and
Elaine Rich, ArtificMl Intelligence (New York: McGraw-Hill, 1983).
11. See Belnap and Steel, The Logic of Questions and Answers, pp. 5,109-121.
12. Progress and Its Problems, p. 24.
13. Cf. Lakatos, "Falsification and the Methodology of Scientific Research
Programmes"; Laudan, Progress and Its Problems; and Thomas Nickles, "Scien-
tific Discovery and the Future of Philosophy of Science," in Thomas Nickles, ed.,
Scientific Discovery, Logic, and Rationality (Dordrecht: D. Reidel, 1980).
14. Actually, pursuit may not be wholly different from the components men-
tioned above. A decision to pursue a specific research program is arguably predi-
cated on the cognizer's judgment of which program is most promising (in my
interpretation, most likely to lead to truth), which is a species of doxastic decision.
Having selected a research program, the scientist usually needs to generate new
hypotheses under the aegis of that program and then devise tests of these hypoth-
eses. So pursuit can itself be broken down into subcomponents, perhaps of the
same variety as previously mentioned.
15. Paraphrased from Michael Posner, Cognition: An Introduction (Glenview,
Ill.: Scott, Foresman, 1973), pp. 150-151.
16. Alternative modes of framing outcomes also have significant ramifications
for choice behavior, as shown in Amos Tversky and Daniel Kahneman, "The
Framing of Decisions and the Psychology of Choice," Science, 211
(1981):453-458, and Daniel Kahneman and Amos Tversky, "Choices, Values, and
Frames," American Psychologist, 39 (1984):341-350. However, representations
for behavioral choice do not concern me here.
17. G. Polya, How to Solve It, 2nd ed. (Princeton: Princeton University Press,
1971), pp. 10, 209.
18. Cognition, pp. 151-152. The example is drawn from C. N. Cofer, "The
Role of Language in Human Problem Solving," paper presented at a conference
on human problem solving, New York University, 1954.
19. Cf. A. de Groot, Thought and Choice in Chess (The Hague: Mouton, 1965),
and William Chase and Herbert Simon, "The Mind's Eye in Chess," in William
Chase, ed., Visual Infomution Processing (New York: Academic Press, 1973).
20. Walter Kintsch and T. A. van Dijk, "Toward a Model of Text Compre-
hension and Production," ,PsychologicalReoiew, 85 (1978):363-394.
21. Mary Gick and Keith Holyoak, "Analogical Problem Solving," Cognitive
Psychobgy, 12 (1980):306355.
22. For reviews of artificial intelligence approaches to strategies, see the
works cited in note 10.
23. The relevance of real-time constraints is discussed further in Chapter 13.
Notes to Pages 136-1 49 399
9. Perception
1. Irvin Rock, The Logic of Perception (Cambridge, Mass.: MIT Press, 1983);
see chap. 1, especially pp. 17-19. David Marr is even better known for positing
multiple perceptual stages and outputs (see his Vision, San Francisco: W. H.
Freeman, 1982). But Marr's perceptual stages do not exemplify the kind of dis-
tinction I wish to emphasize here.
2. For example, see Marr, Vision, pp. 339-343.
3. D. H. Hubel and T. N. Weisel, "Receptive Fields, Binocular Interaction,
and Functional Architecture in the Cat's Visual Cortex," Journal of Physiology,
166 (1962):106-154; J. Y. Lettvin, H. R. Maturana, W. S. McCulloch, and W. H.
Pitts, "What the Frog's Eye Tells the Frog's Brain," Proceedings of the IRE, 47
(1959):1940-51.
4. G. C. Kinney, M. Marsetta, and D. J. Showman, Studies in Display Symbol
Legibility, pt. 21, "The Legibility of Alphanumeric Symbols for Digitalized Tele-
vision" (Bedford, Mass.: Mitre Corp., 1966), no. ESD-TR-66-117.
5. My discussion here and below draws on John R. Anderson, Cognitive Psy-
chology and Its Zmplicatbns (San Francisco: W. H. Freeman, 1980).
6. R. M. Warren, "Perceptual Restorations of Missing Speech Sounds," Sci-
ence, 167 (1970):392-393.
7. G. A. Miller and S. Isard, "Some Perceptual Consequences of Linguistic
Rules," Journal of Verbal Learning and Verbal Behavior, 2 (1963):217-228.
8. Jerry Fodor, The Modularity of Mind (Cambridge, Mass.: MIT Press,
1983).
404 Notes to Pages 187-1 95
25. On this topic see Roderick Firth, "Coherence, Certainty, and Epistemic
Priority," The Journal of Philosophy, 61 (1964):545-557, and Mark Pastin, "Mod-
est Foundationalism and Self-warrant," in George Pappas and Marshall Swain,
eds., Essays on Knowledge and Justification (Ithaca, N.Y .: Cornell University
Press, 1979).
26. See John Pollock, "A Plethora of Epistemological Theories," in George
Pappas, ed., Justification and Knowledge (Dordrecht: D. Reidel, 1979).
27. Cf. Roderick Chisholm's notion of "concurrence," in Theory of Knowl-
edge, 2nd ed. (Englewood Cliffs, N.J.: Prentice-Hall, 1977), p. 83. I referred to
other versions of positive coherence in Chapter 5.
28. See Pollock, "A Plethora of Epistemological Theories," p. 102.
29. Since there can be multiple right J-rule systems, on my theory, one right
system might be F-ist and another right system might be C-ist. This suggests the
commonly ignored possibility that both F-ism and C-ism might be right. But it is
also possible that F-ism could be viewed as the doctrine that all right J-rule sys-
tems have a certain F-ist property; and C-ism could say that all right J-rule sys-
tems have a certain C-ist property. On this approach the possibility of multiple
right systems would not provide for the cotenability of F-ism and C-ism.
30. See Fodor, Modularity of Mind, p. 107.
10. Memory
1. For example, see my A Theory of Human Action (Englewood Cliffs, N.J.:
Prentice-Hall, 1970; Princeton: Princeton University Press, 1976), chap. 4.
2. See Robert Crowder, "The Demise of Short-Term Memory," Acta Psycho-
bgica, 50 (1982):291-323.
3. John Anderson, The Architecture of Cognition (Cambridge, Mass.: Harvard
University Press, 1983), chaps. 1, 3.
4. I argued for this sortof distinction in an earlier paper, "Epistemology and
the Psychology of Belief," The Monist, 61 (1978):525-535.
5. See Daniel Dennett, "Brain Writing and Mind Reading," in Brainstom
(Montgomery, Vt.: Bradford Books, 1978).
6. Ibid., p. 45.
7. For example, see Robert Cummins, "Inexplicit Representations," forth-
coming in Myles Brand and Robert Harnish, eds., The Representation of Knowl-
edge and Belief (Tucson: University of Arizona Press).
8. This case is borrowed from Christopher Cherniak, "Rationality and the
Structure of Human Memory," Synthese, 57 (1983):163-186.
9. Rudolf Carnap, Logical Foundationsof Probability, 2nd ed. (Chicago: Uni-
versity of Chicago Press, 1962), and Carl Hempel, "Inductive Inconsistencies," in
Aspects of Scientific Explanation (New York: Free Press, 1965).
10. "Inductive Inconsistencies," p. 64; my emphases.
11. Cf. Cherniak, "Rationality and the Structure of Human Memory," pp.
164-165.
12. Gilbert Harman, Thought (Princeton: Princeton University Press, 1973),
pp. 158-159.
13. The complexity of the time course of culpability is discussed by Holly
406 Notes to Pages 207-215
and beliefs survive the total discrediting of their evidence base" (p. 179). This
summary statement does not do justice to the facts of the experiment I have just
reviewed. After all, most of the initial impressions in this experiment were suc-
cessfully eliminated by process debriefings, and some of the effects of the initial
impressions were eliminated by outcome debriefings. Admittedly, there are many
other studies in their chapter on which Nisbett and Ross's summary statement
draws. But, all in all, the evidence they marshal does not support the very strong
statement quoted from Bacon. And their next sentence suggests that conflicting
evidence is always treated as if it were supportive of beliefs, that beliefs always
survive the discrediting of their evidence, and so on; whereas, in fact, these things
only sometimes happen.
29. Ibid., p. 181.
30. P. C. Wason and P. N. Johnson-Laird, Psychology of Reasoning: Structure
and Content (London: Batsford, 1965); M. Snyder and W. B. Swann, "Behavioral
Confirmation in Social Interaction: From Social Perception to Social Reality,"
Journal of Experimental SocMl Psychology, 14 (1978):148-162; and Snyder and N.
Cantor, "Testing Theories about Other People: Remembering All the History
that Fits," unpublished manuscript, University of Minnesota, 1979.
31. L. Ross, M. R. Lepper, F. Strack, and J. L. Steinmetz, "Social Explanation
and Social Expectation: The Effects of Real and Hypothetical Explanations upon
Subjective Likelihood," Journal of Personality and Social Psychology, 35
(1977):817-829.
32. R. Hastie and P. A. Kumar, "Person Memory: Personality Traits as Organ-
izing Principles in Memory for Behavior," Journal of Personality and Social Psy-
chology, 37 (1979):25-38.
33. Human Inference, p. 167.
34. The literature on artificial intelligence contains a good bit of discussion of
belief revision and dependency-directed backtracking. See Jon Doyle, "A Model
for Deliberation, Action, and Introspection," (Cambridge, Mass.: MIT Artificial
Intelligence Laboratory, 1980), no. TR-581; Doyle, "A Truth Maintenance Sys-
tem," Artificial Intelligence, 12 (1979):231-272; and R. M. Stallman and. G. J.
Sussman, "Forward Reasoning and Dependency-Directed Backtracking in a Sys-
tem for Computer-Aided Circuit Analysis," Artificial Intelligence, 9
(1977):135-196. I cannot pursue this technical literature in detail. But the com-
plications involved in a system of backtracking show that it may be excessive to
require ubiquitous backtracking. At a minimum this may prove excessive from a
resource-relative perspective.
35. Similar ideas are expressed by Gilbert Harman, in "Positive versus Nega-
tive Undermining in Belief Revision," Nous, 18 (1984):39-49. My own thoughts
on this were arrived at independently.
36. S. A. Madigan, "Intraserial Repetition and Coding Processes in Free Re-
call," Journal of Verbal Learning and Verbal Behavior, 8 (1969):828-835.
37. See Anderson, Cognitive Psychology and Its Implications, chap. 7.
38. W. Penfield, "The Interpretive Cortex," Science, 129 (1959):1719-25.
39. T. 0.Nelson, "Savings and Forgetting from Long-Term Memory," Journal
of Verbal Learning and Verbal Behavior, 10 (1971):568-576.
40. D. E. Rumelhart, P. Lindsay, and D. A. Norman, "A Process Model for
408 Notes to Pages 223-233
39. Richard Nisbett and Lee Ross, Human Inference: Strategies and Shortcorn-
ings of Social Judgment (Englewood Cliffs, N.J.: Prentice-Hall, 1980), chap. 3.
40. See Janellen Huttenlocher, "Constructing Spatial Images: A Strategy in
Reasoning," in P. N. Johnson-Laird and P. C. Wason, eds., Thinking (Cambridge:
Cambridge University Press, 1977).
41. Ibid., pp. 89-90.
42. H. H. Clark, "Linguistic Processes in Deductive Reasoning," in Johnson-
Laird and Wason, eds., Thinking.
43. David Hilbert, "On the Infinite," in Paul Benacerraf and Hilary Putnam,
eds., Philosophy of Mathematics (Englewood Cliffs, N.J.: Prentice-Hall, 1964).
44. Michael Resnik, "Mathematics as a Science of Patterns: Epistemology,"
Nous, 16 (1982):95-105. A similar view appears to have been maintained by the
Bourbaki mathematicians.
45. See Philip Kitcher, The Nature of Mathematical Knowledge (New York:
Oxford University Press, 1984), chap. 6.
46. Penelope Maddy, "Perception and Mathematical Intuition," The Philo-
sophical Reuiew, 89 (1980):163-196.
47. G. H. Hardy, A Mathmautician's Apology (Cambridge: Cambridge Univer-
sity Press, 1941), p. 24.
48. Human Inference, pp. 44-45.
49. W. C. Thompson, R. M. Reyes, and G. H. Bower, "Delayed Effects of
Availability on Judgment," unpublished manuscript, Stanford University, re-
ported in Nisbett and Ross, Human Inference, p. 52.
50. R. Hamill, T. D. Wilson, and R. E. Nisbett, "Ignoring Sample Bias: Infer-
ences about Collectivities from Atypical Cases," unpublished manuscript, Uni-
versity of Michigan, reported in Nisbett and Ross, Human Inference, p. 57.
51. R. E. Nisbett, E. Borgida, R. Crandall, and H. Reed, "Popular Induction:
Information is Not Always Informative," in J. S. Carroll and J. W. Payne, eds.,
Cognition and Social Behuuior (Hillsdale, N.J.: Erlbaum, 1976).
52. Human Inference, pp. 51-52.
53. Shelley Taylor and Suzanne Thompson, "Stalking the Elusive 'Vividness*
Effect," Psychological Review, 89 (1982):155-181.
nisms elsewhere in nature. Consider, for example, the WTA mechanisms of mat-
ing found in certain species. Elk and bighorn sheep have battles among their
males to determine dominance within the herd. The females of the herd all mate
exclusively with the single victorious male. This mechanism has some obvious ge-
netic advantages. It allows the genes of only the strongest or most battleworthy
male to be transmitted to subsequent generations. It also has some apparent dis-
advantages: it does not make use of the entire gene pool of the male population.
The viability of such species, however, suggests that this is a tenable mating
mechanism. Analogous advantages of WTA mechanisms in cognition may also be
identified.
Figure 9.1: redrawn from 0.G. Selfridge, "Pattern Recognition and Modem
Computers," in Proceedings of the Western Joint Computer Conference (New
York: Institute of Electrical and Electronics Engineers, 1955), p. 92. Copyright
o by the American Federation of Information Processing Societies.
Figure 9.2: redrawn from Irvin Rock, % Logic of Perception (Cambridge,
Mass.: MIT Press, 1983), p. 309.
Figure 9.3: redrawn from Stephen Palmer, "Symmetry, Transformation, and
the Structure of Perceptual Systems," in J. Beck, ed., Organization and
Representation in Perception (Hillsdale, N.J.: Erlbaum, 1982),p. 122.
Figure 10.1: redrawn from L.Ross, M. Lepper, and M. Hubbard,
"Perseverance in Self Perception and Social Perception: Biased Attributional
Processes in the Debriefing Paradigm," Journal of Personality and Social
Psychology, 32 (1975):887. Copyright Q 1975 by the American Psychological
Association. Adapted by permission of the publisher and author.
Figure 10.2: redrawn from Stephen Madigan, "Intraserial Repetition and
Coding Processes in Free Recall," Journal of Verbal Learning and Verbal
B e h a w , 8 (1969):829.
Figure 11.1: redrawn from John Anderson, Cognitive Psychology and Its
Implications (San Francisco: W . H. Freeman, copyright o 1980),p. 53.
Figure 11.2: redrawn from G. H. Bower and A. L. Glass, "Structural Units
and the Redintegrative Power of Picture Fragments," Journal of Experimental
Psychology: Human Learning and Memory,2 (1976):459,460. Copyright o 1976 .
by the American Psychological Association. Adapted by permission of the
publisher and author.
Figure 11.3: redrawn from Stephen Palmer, "HierarchicaI Structure in
Perceptual Representation," Cognitive Psychology, 9 (1977):452.
Figures 11.4 and 11.5: redrawn from Diana Deutsch and John Feroe, "The
Internal Representation of Pitch Sequences in Tonal Music," Psychological
Reoiew, 88 (1981):504, 505. Copyright @ 1981 by the American Psychological
Association. Adapted by permission of the publisher and author.
Figures 11.6 and 11.7: redrawn from Donald Hoffman, "The Interpretation
of Visual Motion," Scientific American, 249 (1983):157, 159. Copyright o 1983
by Scientific American, Inc. All rights reserved.
Figure 12.1: redrawn from Roger Shepard and Lynn Cooper, Mental Images
and Their Transformations (Cambridge, Mass.: MIT Press, 1982), p. 21.
422 Illustration Credits
Derivability, of right J-rules from logic, Error, 26; and skepticism, 30-32. See also
81-89 Skepticism
Descriptional representations. See Repre- Evaluation, 20, 137-138, 247; as episte-
sentations mology's mission, 2 3 ; objects of, 3-6; of
Detachment, rules of, 73, 115 psychological processes, 4-5; factual
Dichotomy in epistemology, 299-304 basis of, 22-25, 63, 69, 78; regulative vs.
Dimensions of evaluation. See Epistemic nonregulative, 25-26, 59. See also Epis-
evaluation temic evaluation; Value
Direct skepticism. See Skepticism Evidence, 17, 145-146, 147, 339; eviden-
Discrimination, 47, 371-373 tial evaluation, 21, 22; evidential gaps,
Discount model, 347, 348 32; and best explanation, 37; theories of,
Dispositional belief. See Belief 37; evidential strength and J-rule right-
Doing the best one can, 104,3941122 ness, 66, 69; weight of, 89; proportion-
Domain of evaluation. See Epistemic eval- alism, 89-93, 97; retrieval of, from
uation memory, 133-134, 204-207; availability
Doxastic attitudes, 14: and J-rules, 77-78 of, 204; total evidence principle, 204;
Dreams, 239 and inference, 204-207; and belief up-
Duplex theory of memory, 200 dating, 344-358. See also Justification;
Dutch Book argument, 316,327428 Skepticism
Evolutionary epistemology, 2
Elaboration, 209-214, 217, 218-219. See Explanation: and theory of evidence, 37,
also Memory 249-250; as theory of justificational
Elenchus, 8 value, 98, 100-101, 140; and question
Empirical science, 2, 299,300. See also answering, 125; psychological, 172-173;
Science and perseverance, 217-218
Empiricism, 39-40 Externalism, 24, 72
Epistemic approach to truth. See External world, problem of, 31. See also
Truth Skepticism
Epistemic boundedness, 228,245
Epistemic decision theory, 101-103 Fallacy, 295-298,308410,311. See also
Epistemic evaluation, 20, 74; domain of, Conjunction fallacy
21; terms and dimensions of, 21-22: Fallibility: of cognitive faculties, 31; of
standards of, 22-25.26-27; styles reason, 31; of the senses, 31. See also
of, 25-26; multiple standards of, 122- Error; Skepticism
125. See also Evaluation, factual Fantasizer, 99, 100, 101
basis of First-order processes, 94, 369, 373
Epistemics, 9,28, 181-184,378-380. See Fittingness, 152- 156
also Primary epistemology Folk psychology, 14, 164-169
Epistemic utility, 101-102 Forgetting, 222-223,225
Epistemologically Original Position. See Fonnalism: mathematical, 34; in theory of
Original Position rationality, 353
Epistemology: as multidisciplinary affair, Formality condition, 170
1-3,8-9; and psychology, 1-5,6-7,37, Foundationalism: and basic beliefs, 79,
53,57,66, 163, 181-184,249,378480, 196; core doctrine, 79; and truth, 79;
383n4; history of, 6-8,299300,304. and coherentism, 80, 194-198; and relia-
378; and skepticism, 28-41; belief vs. bilism, 80; strong vs. weak, 196, 198
product, 138. See also Epistemics; Evo- Foundational questions, 8-9
lutionary epistemology; Individual Framework principle (for justification),
epistemology; Naturalistic epistemology; 59-63; rejection of iterative, 61; no-un-
Primary epistemology; Secondary dermining clause in, 62-63: historical
epistemology; Skepticism; Social episte- reformulation of, 83
mology Functionalism, 164
432 Subject Index
Virtues and vices, epistemic, 25. See also Weight, of evidence. See Evidence
Epistemic evaluation Winner-take-all (WTA) network, 329,
Visual buffer, 262, 264,265, 267 333-334,338-343,417n19
Visual representation, 229-233, 235-237
Vividness, 268, 273-277