Academia.eduAcademia.edu

Are Newcomb problems really decisions?

2007, Synthese

Richard Jeffrey long held that decision theory should be formulated without recourse to explicitly causal notions. Newcomb problems stand out as putative counterexamples to this 'evidential' decision theory. Jeffrey initially sought to defuse Newcomb problems via recourse to the doctrine of ratificationism, but later came to see this as problematic. We will see that Jeffrey's worries about ratificationism were not compelling, but that valid ratificationist arguments implicitly presuppose causal decision theory. In later work, Jeffrey argued that Newcomb problems are not decisions at all because agents who face them possess so much evidence about correlations between their actions and states of the world that they are unable to regard their deliberate choices as causes of outcomes, and so cannot see themselves as making free choices. Jeffrey's reasoning goes wrong because it fails to recognize that an agent's beliefs about her immediately available acts are so closely tied to the immediate causes of these actions that she can create evidence that outweighs any antecedent correlations between acts and states. Once we recognize that deliberating agents are free to believe what they want about their own actions, it will be clear that Newcomb problems are indeed counterexamples to evidential decision theory. Keywords Newcomb Problem • Richard Jeffrey • Causal Decision Theory • Evidential Decision Theory • Ratifiability • Freedom Richard Jeffrey long held that decision theory should be formulated without invoking explicitly causal notions. The version of expected utility theory he championed in The Logic of Decision, Jeffrey (1983), did not use counterfactual conditionals, imaging functions, causally homogenous partitions, directed causal graphs, causal dependency hypotheses, or any of the other vehicles of causal information that "causal" decision

Synthese (2007) 156:537–562 DOI 10.1007/s11229-006-9137-6 O R I G I NA L PA P E R Are Newcomb problems really decisions? James M. Joyce Published online: 20 April 2007 © Springer Science+Business Media B.V. 2007 Abstract Richard Jeffrey long held that decision theory should be formulated without recourse to explicitly causal notions. Newcomb problems stand out as putative counterexamples to this ‘evidential’ decision theory. Jeffrey initially sought to defuse Newcomb problems via recourse to the doctrine of ratificationism, but later came to see this as problematic. We will see that Jeffrey’s worries about ratificationism were not compelling, but that valid ratificationist arguments implicitly presuppose causal decision theory. In later work, Jeffrey argued that Newcomb problems are not decisions at all because agents who face them possess so much evidence about correlations between their actions and states of the world that they are unable to regard their deliberate choices as causes of outcomes, and so cannot see themselves as making free choices. Jeffrey’s reasoning goes wrong because it fails to recognize that an agent’s beliefs about her immediately available acts are so closely tied to the immediate causes of these actions that she can create evidence that outweighs any antecedent correlations between acts and states. Once we recognize that deliberating agents are free to believe what they want about their own actions, it will be clear that Newcomb problems are indeed counterexamples to evidential decision theory. Keywords Newcomb Problem · Richard Jeffrey · Causal Decision Theory · Evidential Decision Theory · Ratifiability · Freedom Richard Jeffrey long held that decision theory should be formulated without invoking explicitly causal notions. The version of expected utility theory he championed in The Logic of Decision, Jeffrey (1983), did not use counterfactual conditionals, imaging functions, causally homogenous partitions, directed causal graphs, causal dependency hypotheses, or any of the other vehicles of causal information that “causal” decision J. M. Joyce (B) Department of Philosophy, University of Michigan, 2215 Angel Hall, 435 South State Street, Ann Arbor, MI 48109-1003, USA e-mail: jjoyce@umich.edu 538 Synthese (2007) 156:537–562 theories employ. Jeffrey did not, however, think that causal information is irrelevant to decision making. Rather, he maintained that, insofar as rational choice theory is concerned, an agent’s beliefs about causal relationships can be adequately characterized in terms of her subjective conditional probabilities for non-causal propositions. Since these conditional probabilities characterize the agent’s views about evidential relationships, Jeffrey’s formalism is often referred to as evidential decision theory because it asks agents to choose auspicious acts that provide them with evidence for thinking that desirable outcomes will ensue. Newcomb problems stand out as putative counterexamples to Jeffrey’s theory. In these problems evidential decision theory asks agents to choose auspicious but inefficacious acts that provide evidence for desirable outcomes without doing anything to causally promote these outcomes. This has struck many decision theorists, including Jeffrey himself, as wrongheaded. The whole point of rational agency is to make choices that change things for the better, and this is something one can only do by performing acts that causally promote desirable consequences. When an agent chooses an auspicious but inefficacious act she gets “good news,” but does nothing to improve her prospects. Jeffrey initially sought to retain the “evidential” character of his theory in the face of Newcomb problems via recourse to the doctrine of ratificationism. This doctrine, which owes much to Eells (1982), advises you to “choose for the person you expect to be once you have chosen.” (1983, p. 16) If an agent makes only ratifiable choices, Jeffrey argued, then the purely evidential import of her acts will be nullified. After she has made her choice she will know which action she plans to perform, and this knowledge will “screen off” any purely evidentiary relationships that might hold between her acts and states of the world. For reasons to be discussed below, Jeffrey came to see this response as problematic, and so gave up on ratificationism as a solution to Newcomb problems. Indeed, we shall see that the situation for ratificationism is worse than even Jeffrey imagined: it will be shown that the ratificationist argument presupposes the truth of causal decision theory. In his final treatment of the topic, Jeffrey (1993, 2004), Jeffrey argued that Newcomb problems present no obstacle to evidential decision theory because they are not really decision problems at all! Agents who face them, he claimed, possess so much evidence about correlations between their actions and states of the world that they are unable to regard their deliberate choices as causes of outcomes. Such agents do not see themselves as making free choices. Jeffrey’s reasoning goes wrong, I shall argue, because it fails to recognize that an agent’s beliefs about her own (immediately available) actions are so closely tied to the immediate causes of these actions that she is in a position to create evidence that will outweigh any antecedent correlations might have existed between acts and the states. Once we recognize that that freedom consists, in part, in being free to believe what one wants about one’s own actions, we will see that Newcomb problems are genuine decision problems, and that they do indeed serve as counterexamples to evidential decision theory. 1 Evidential versus causal decision theory The central difference between the evidential and causal theories concerns, not what an agent should do, but whether we can adequately characterize her rationale for doing what she should do without making explicit reference to her beliefs about what Synthese (2007) 156:537–562 539 her choices are likely to cause. Causal decision theorists maintain that there is no avoiding causality.1 To know what an agent should do in a given situation we must know what she believes about the effects of her actions, and this is only possible if we make reference either to her beliefs about propositions with explicitly causal content or to forms of belief revision that are subject to explicitly causal constraints.2 Evidential theorists dispute this. They claim that, for purposes of decision theory, we can capture the relevant beliefs about causes and effects by appealing to nothing more than the agent’s ordinary subjective probabilities for non-causal propositions. In the simplest sorts of cases, the issue boils down to a dispute about the probabilities that figure into calculations of expected utility. Imagine a 2 × 2 decision matrix E ∼E A ∼A des(A & E) des(∼A & E) des(A & ∼E) des(∼A & ∼E) in which A and ∼A are propositions that describe possible acts and E and ∼E are propositions that describe the “states of the world” on which the outcomes of these acts depend. In Jeffrey’s framework outcomes are act/state conjunctions, and each has a utility (or “desirability”) as listed in the matrix. Early decision theorists, following Savage (1972), computed the expected utility of an act by weighting the utilities of its outcomes by the probabilities of the events that bring them about, so that EU V(A) = prob(E)des(A & E) + prob(∼E)des(A & ∼E) V(∼A) = prob(E)des(∼A & E) + prob(∼E)des(∼A & ∼E) Among other things, this entails the unrestricted “sure-thing” principle of Savage (1972, pp. 21–22). Sure-thing Principle (STP). For any event E, if an agent prefers A to ∼A both on the supposition that E obtains and on the supposition that E does not obtain, then she should prefer A to ∼A unconditionally. This rule, which relates an agent’s conditional preferences to her unconditional preferences, endorses a strong kind of dominance reasoning: if A is the better act whatever E’s truth-value, then A is the better act simpliciter. Jeffrey recognized that EU and STP are only valid if the agent believes that her choice will not causally affect E’s occurrence. Suppose the agent reasoned thus: “I can pay $10 for an influenza vaccination or I can keep my $10. Either I will get the flu this winter or I won’t. If I am to get the flu, I’d rather have the extra $10 to buy medicine. If I do not get the flu, I’d rather have the $10 to buy hot cocoa. So, whether I get the flu this winter or not, I prefer having the extra $10 to not having it. Thus, by STP, I should forgo the vaccination.” This is clearly absurd: if the agent believes that the vaccination will inhibit influenza, then choosing to forgo it is choosing to make the undesirable outcome more probable and the desirable outcome less probable. 1 For defenses of causal decision theory see Gibbard and Harper (1978), Cartwright (1979), Skyrms (1980), Lewis (1981), Sobel (1985), Armendt (1986), and Joyce (1999). 2 These two options mark an important distinction among causal decision theorists. Some, like Lewis, Skyrms, and Gibbard and Harper, represent the relevant causal beliefs using unconditional subjective probabilities of propositions with causal content. Others, like Sobel and Joyce, prefer to incorporate the causal elements into the process of belief revision. 540 Synthese (2007) 156:537–562 The moral Jeffrey drew from cases like this was that if the agent suspects that A or ∼A might influence E, then the correct expected utilities are given not by EU but by3 EDT V(A) = prob(E/A)des(A & E) + prob(∼E/A)des(A & ∼E) V(∼A) = prob(E/∼A)des(∼A & E) + prob(∼E/∼A)des(∼A & ∼E) Here the utility of each outcome is multiplied by the agent’s subjective probability for its occurrence conditioned on the evidence that the act in question is in fact performed. This requires the deliberating agent to treat information about which action she will ultimately perform as “news items,” on a par with information about any other aspect of the world. On this reading, the value of choosing an action is the same as the value of being told by a reliable soothsayer that one will perform it. Actions are thus evaluated on the basis of the evidence they provide for thinking that desirable results will ensue. EDT sanctions only a restricted form of the sure-thing principle. STPE . For any event E that is evidentially independent of both A and ∼A, so that prob(E/A) = prob(E/∼A), if the agent prefers A to ∼A both on the supposition that E obtains and on the supposition that E does not obtain, then she should prefer A to ∼A unconditionally. The requirement that E be evidentially independent of A and ∼A is meant to prevent the application of STP in cases like the influenza example, where the agent thinks that she can causally influence the state of the world. Jeffrey initially thought that the evidential relationships encoded in ordinary conditional probabilities would provide as much information about causal relations as decision theory requires. In any case where it mattered, he hoped, an agent’s views about the differing causal powers of A and ∼A vis-à-vis E would be revealed in the disparity between prob(E/A) and prob(E/∼A). Since the difference prob(E/A) – prob(E/∼A) is one way to measure of the amount of evidence that A provides for E, the substance of Jeffrey’s initial proposal can be rendered as follows: J For purposes of decision theory, an agent regards A as more efficacious than ∼A as a cause of E when prob(E/A) > prob(E/∼A). Moreover, the quantity prob(E/A)− prob(E/∼A) is her quantitative estimate of the degree of A’s causal efficacy as a promoter of E.4 Causal decision theorists deny this, arguing that no function of prob(E/A) and prob(E/∼A) can ever adequately encode the agent’s beliefs about what her acts are likely to cause. While the values of these conditional probabilities depend, partly, on the agent’s judgments about causal linkages between E and A, they can also reflect 3 Throughout I will use prob(X/Y) to denote the conditional probability of X given Y. It is defined as prob(X/Y) = prob(X & Y)/prob(Y) when prob(Y) > 0. 4 While Jeffrey does not make the second point explicitly, it does follow from his theory. To see why, consider the special case in which A and ∼A produce equally desirable outcomes under both E and ∼E, with the better outcome being produced by E. We can then set des(A & E) = des(A & ∼E) = 1 and des(∼A & E) = des(∼A & ∼E) = 0, and write the difference in the expected utilities of A and ∼A as V(A) – V(∼A) prob(E/A) – prob(E/∼A). Since there is no reason for the agent to prefer A to ∼ A in this context unless she thinks that A’s performance promotes E’s occurrence, it follows that, as far as decision making is concerned, the difference prob(E/A) – prob(E/∼A) may be interpreted as a reflection of the degree to which the agent regards A as a cause of E. Synthese (2007) 156:537–562 541 the agent’s knowledge of non-causal correlations between E and A. For example, prob(E/A) – prob(E/∼A) can be large when A and E are correlative effects of a common cause, but are otherwise causally disconnected. The contrast between the causal and evidential approach to decision theory is brought out most clearly in Newcomb problems. The initial Newcomb problem was a bizarre science fiction story involving a master psychologist who could almost infallibly predict an agent’s actions. Unfortunately, focus on this example has lead many to think that Newcomb problems are so recherché as to bear little relevance to actual decision making. In fact, they are fairly common. Consider, for example the Twin’s Dilemma, in which two agents, Row and Column, must decide whether or not to take some cooperative action as described in the following decision table: R ∼R C ∼C 9, 9 10, 0 0, 10 1, 1 Row chooses a row, Column chooses a column, and the listed pairs give their respective utilities for each outcome. Three conditions must hold for Row to be facing a true Newcomb problem: • • • Row must know that Column’s action is causally independent of her own, i.e., she must know that she has no way of causally influencing what Column will choose or do. Row must believe she and Column are somewhat like-minded. Specifically, her subjective probabilities should meet the (rather mild) condition that prob(C/R) – prob(C/∼R) > 1/9. Row must take herself to be a free agent; she must believe that she has the ability to freely choose either R or ∼R. The first of these conditions is easily met. Indeed, we can ensure it by having the players make simultaneous choices at distant locations so that no causal signal can pass between them. To satisfy the second condition we need only imagine that the players acquired their decision making skills in a common setting: perhaps they grew up in the same neighborhood, or had the same decision theory teacher, or have played the game many times before, or whatever. Given the symmetry of the situation, and their common background, Row might well believe that, whatever she chooses, Column is a bit more likely (probability > 1/9) to choose the same way. On the face of it, the third condition seems trivial as well. We can imagine that Row is sitting before two buttons marked R and ∼R, and that all she has to do is to reach out and push one. What could be freer than that? We will return to this question shortly. When the above three conditions are met, the evidential theory encourages Row to reason thus: “Since Column is significantly more likely to perform C if I do R than if I do ∼R (in the sense that prob(C/R) – prob(C/∼R) > 1/9), and since I am better off by nine utiles if he chooses C, it follows that I should choose R.” This reasoning recommends choosing the auspicious but inefficacious R, which provides Row with evidence for thinking that Column will perform C, over the efficacious but inauspicious ∼R, which will earn her an extra utile whatever Column does (despite being correlated with the highly undesirable ∼C). Causal theorists see this as a fallacy. Row is passing up a sure utile to secure the “good news” that outcome R & C is probable. This would make sense if Row were, in 542 Synthese (2007) 156:537–562 Jeffrey’s phrase, “making the news” that R & C, but in this situation the only “news” Row makes concerns the row of the outcome to be received, Column independently fixes its column. In effect, Row’s control over the news extends only to the question of getting the extra utile. On the causal view, Row is obliged to take the course of action that is most likely to cause desirable results, and this is remains true even when this means getting some “bad news” on the side. EDT goes wrong by weighting the utilities of outcomes by the conditional probabilities of states given acts. Instead, the relevant utilities should be computed according to the formulas CDT U(A) = prob(E\A)des(A & E) + prob(∼E\A)des(A & ∼E) U(∼A) = prob(E\∼A)des(∼A & E) + prob(∼E\∼A)des(∼A & ∼E) where prob(E\A) and prob(E\∼A) are causal probabilities that captures the agent’s views about the relative causal powers of A and ∼A vis-à-vis E. Causal theorists interpret these quantities in a variety of ways,5 but all agree that (a) the agent regards A as a promoting cause of E exactly if prob(E\A) – prob(E\∼A) > 0, and (b) these are not ordinary conditional probabilities, i.e., prob(E\A)  = prob(E/A) and prob(E\∼A)  = prob(E/∼A). Since CDT reduces to EU when acts have no causal influence over states, the causal theory entails EU when E is causally independent of A.6 Thus, the causal decision theorist endorses the following version of the sure-thing principle STPC . For any event E that is causally independent of A and ∼A, so that prob(E\A) =prob(E\∼A), if the agent prefers A to ∼A both on the supposition that E obtains and on the supposition that E does not obtain, then she should prefer A to ∼A unconditionally. Accordingly, in Newcomb problems, causal decision theory always recommends that Row choose the efficacious act ∼R, even though this provides her with evidence for thinking that the undesirable state ∼C obtains. Most philosophers, Jeffrey included, now agree that choosing the merely auspicious R is a mistake. The vexed question is whether one needs the entire apparatus of causal decision theory, with its invocation of mysterious causal probabilities or causal independence conditions, to get this answer. 2 Ratificationism Some evidentialists have sought to defuse Newcomb Problems by arguing that EDT does not recommend the merely auspicious choice because the agent will make her decision from an epistemic perspective in which the R and C are evidentially independent. In his “tickle defense,” Eells (1982) argued that a fully rational agent will come to know what she intends to do before she acts, and that this information will 5 See (Joyce, 1999, pp. 161–180) for a discussion of the options. 6 Unlike evidential independence, the notion of causal independence, as it is being used here, is non-symmetric. When I say that E is causally independent of A I mean that changes in A’s truth-value do not cause changes in E’s truth-value. It does not follow that changes in E’s truth-value do not cause changes in A’s truth-value. Thus, E might be causally independent of A even though A is causally dependent on E. Indeed, this will happen whenever E is a cause of A. Synthese (2007) 156:537–562 543 Fig. 1 “screen off’ any evidential connections there might have been between her acts and states of the world. In a similar vein, Jeffrey’s doctrine of ratificationism suggests that a rational agent should choose for the person she expects to be once she has chosen by evaluating each action on the hypothesis that she will ultimately decide to perform it. On either view, a rational agent’s ability to anticipate her own decisions nullifies any purely evidential correlations that might exist between states and acts. To see how this works, let dA denote the decision to perform A. This is a different proposition from A itself; it is at least logically possible for the agent to decide on an act and yet not perform it, so that ∼A & dA and A & d ∼ A are non-contradictory. We assume that all conditional probabilities of the form prob(±C/ ± R& d ± R) make sense.7 Jeffrey’s ratificationist solution to the Newcomb problem is based on the following principles: Screening. The decision to perform an act screens off any evidential correlations that might exist between that act and states of the world , so that in Twin’s Dilemma one has prob(C/R & dR) = prob(C/∼R & dR) = prob(C/dR) prob(C/R & d∼R) = prob(C/∼R & d∼R) = prob(C/d∼R) Maxim of Ratifiability . An agent can rationally perform act A only if A is ratifiable in the sense that there is no alternative B whose expected utility exceeds that of A on the supposition that A is decided upon. Screening is an instance of the Causal Markov Condition8 that figures so prominently in contemporary approaches to Bayesian causal modeling. (See Pearl, 2000; Spirtes, Glymour, & Scheines, 2000) Roughly, the Causal Markov Condition says that, given a sufficiently complete description of the causal structure of a situation, knowledge of all an event’s direct causes (its “causal parents”) makes any other information statistically irrelevant to the event’s occurrence (except for information about the event’s own “causal descendents”). For the Twin’s Dilemma, the picture Jeffrey has in mind is as described in (Fig. 1). Letters denote causal variables, and each arrow indicates that the variable at its tail is a causal parent of the one at its head. An assignment of values to all parents of a given variable causally determines the variable’s value. The graph is assumed to be complete in the sense that variables, like B and B∗ , with no common causal ancestors are both causally and statistically independent. dR is the variable of Row’s decision, 7 To ensure this Jeffrey introduced a “trembling hand” condition that requires the agent to believe that, for any two actions, there is some positive probability that she will decide to do one of them but end up doing the other. If one thinks conditional probabilities can be well-defined even when their conditions have probability zero (Joyce, 1999, pp. 201–214), then such a trembling hand condition is unnecessary. 8 This misnamed principle is a generalization of Reichenbach’s “common cause” principle (1956, Ch. 3). 544 Synthese (2007) 156:537–562 while dC is the column player’s choice. These decisions, it is assumed, are correlated in virtue of having a common cause (or set of causes) CC. Both decisions, as well as the player’s subsequent actions, might depend on other causal variables, X and B or X ∗ and B∗ as the case may be. Crucially though, these other variables are not correlated, so that, e.g., prob(B/B∗ ) = prob(B). This ensures that the only causal or evidential relationships that obtain between R and C hold in virtue of the path through CC that connects them. In Twin’s Dilemma, the existence of this path provides the basis for a “backtracking” inference from R to C. Since effects provide information about their causes, Row can reason “up the graph” from a truth-value for R to some conclusion about CC. The sort of inference she is able to make will depend on the exact relationship between B, X, X ∗ , dR and CC, but in a genuine Newcomb problem she will learn enough about CC’s value to reason “down” to a conclusion about C’s value (using the principle that causes provide evidence for their effects). Again, the sort of inference she can make will depend on the precise relationship between B∗ , X, X ∗ , dC and CC, but in a Newcomb problem she will be able to assign subjective probabilities such that prob(C/R) – prob(C/∼R) > 1/9. The Markov condition blocks such “backtracking” inferences by ensuring that every variable in the graph, other than B, is evidentially independent of R conditional on a value for dR. Consequently, Row will not see any evidential connection between her acts and Column’s acts after she comes to know her decision. Moreover, she will prefer news of ∼R to news of R conditional on whatever decision she makes. If she decides on R, and so learns dR, then her evidential utilities are V(R/dR) = 9prob(C/dR) < V(∼R/dR) = 9 prob(C/dR) + 1. If she decides on ∼R, and so learns d∼R, then V(R/d∼R) = 9prob(C/d∼R) < V(∼R/d∼R) = 9prob(C/d∼R) + 1. Since ∼R has an additional unit of “news value” either way, Row can be certain that she will prefer ∼R to R once she knows her decision, no matter what that decision turns out to be. This highlights a clear asymmetry between the two acts. Row knows in advance that if she decides on R she will regret her choice because, from the perspective of her new epistemic state, ∼R will be better news. She also knows that if she decides on ∼R, she will be pleased with what she has done and so will “ratify” her decision. But why should the fact that ∼R is ratifiable in this way make a difference to what Row chooses? Even though ∼R’s news value exceeds that of R when Row is certain of her decision, the reverse is true while she is still trying to make up her mind. The Maxim of Ratifiability supplies the last, crucial step in Jeffrey’s argument. In the context of evidential decision theory, it requires the agent to subordinate her current estimates of the auspiciousness of acts to her future estimates of their auspiciousness, at least to the extent that she can determine what her future estimates are. Thus, the combination of the Maxim of Ratifiability and evidential decision theory imposes the following necessary condition on rational choices: An agent can rationally perform the act A only if A is evidentially ratifiable in the sense that V(A/dA) > V(B/dA) for every alternative act B. Jeffrey initially maintained that, in the presence of Screening, this solves the Newcomb without leaving the confines of the evidential theory. Even though performing R on the basis of a decision to perform R provides Row with better news than does performing ∼R on the basis of a decision to perform ∼R, only the efficacious act ∼R is ratifiable. Thus, as Jeffrey put it, Synthese (2007) 156:537–562 545 the [Twin’s dilemma] exemplifies a class of problems in which the agent foresees during deliberation that the choice will change [her] subjective probabilities so as to make the dominant act be the one whose performance will have the highest estimated desirability once a decision (either decision) is made, even though it did not have the highest estimated desirability when the agent’s uncertainty about his decision was taken into account, before the choice. It seems clear that one should choose an act whose performance one expects to have highest expected desirability if and when it is chosen. (1983, p. 17) Jeffrey soon came to reject this response to Newcomb problems because, as Bas van Fraassen pointed out, Screening is implausible in some versions of Twin’s Dilemma. Consider an example given in (Joyce, 1999, p. 159). Suppose Row is a bumbler who recognizes that she sometimes fails to carry out her decisions. One can think of the variable B in Fig. 1 as controlling Row’s bumbling. Column might also be a bumbler, with his bumbling controlled by B∗ . If we suppose that both players tend to bumble in a similar fashion, and that both know this, then it can happen that R rather than ∼R is the uniquely ratifiable act. If the probabilities are prob(R & C/dR) = 0.6 prob(R & ∼C/dR) = 0.1 prob(∼R & C/dR) = 0.1 prob(∼R & ∼C/dR) = 0.2 prob(R & C/d∼R) = 0.1 prob(R & ∼C/d∼R) = 0.2 prob(∼R & C/d∼R) = 0.1 prob(∼R & ∼C/d∼R) = 0.5 then R and ∼R provide better evidence about C than do dR and d∼R alone. The news values work out so that V(R/dR) = 7.71 > V(∼R/dR) = 4 and V(R/d∼R) = 3 > V(∼R/d∼R) = 2.5, which makes R uniquely ratifiable. While this sort of argument led Jeffrey to give up on ratificationism as a solution to Newcomb’s problem, considerations introduced in Eells (2000) suggest that Jeffrey might have been too hasty. Eells argues that van Fraassen’s example should be “disqualified” because “rational deliberation loses its point” when there is a correlation among acts that cannot be screened off by decisions. (p. 896) The crux of the matter is that Row, who perceives a correlation between her bumbling tendencies and those of Column, will regard R’s truth-value as lying outside of the sphere of things that she can directly control. Moreover, she will seek to explain the correlations between the R and C that remain after conditioning on dR or d∼R by postulating the existence of a common cause, which lies outside her control (Fig. 2). For her, the correct picture is not Fig. 1, in which there is an unexplained correlation between B and B* (thus violating the Causal Markov Condition). Here the players’ bumbling has a common cause CC∗ that jointly influences the truth-values of R and C via pathways that do not pass through their decisions. Thus, each player believes that her action depends on factors that causally influence the other player’s action without affecting his decision. In such a situation, Row cannot regard herself as truly free in the matter of R since its truth-value is not determined solely by her deliberate choices. As Eells observes, “to the extent to which you believe that your actual act is influenced by factors other than your decision (the result of rational deliberation), it would seem that, to you rational deliberation will lose its point.” (p. 896) In effect, Eells is arguing that van Fraassen’s version of the Newcomb problem is not a genuine decision. I have come to think Eells is right. Row will see herself as fully free in the matter of R only if she believes that external factors influence R’s truth-value solely 546 Synthese (2007) 156:537–562 Fig. 2 Fig. 3 by influencing her decision. She must, that is, see herself as facing a problem like this (Fig. 3). If this is Row’s situation, then knowledge of her decision will screen off any purely evidential correlations that may hold between her acts and those of the other player. So, to concede Eells’s point, Screening is plausible in any Twin’s Dilemma played by agents who see themselves as making a genuinely free choice. Accordingly, I grant that “the ratificationist defense is stronger if one makes the (perhaps idealized) assumption that the [agent’s] act is not correlated with decision relevant causes that bypass the process of rational deliberation.” (p. 896) Still, the evidential ratificationist is not home free. Since the objections to Screening have seemed so decisive (even to Jeffery), critics have largely neglected the role of the Ratifiability Maxim in this argument. This is regrettable because, as we will see, its use is not nearly so straightforward as ratificationists make it seem. Despite appearances, the combination of evidential decision theory and the Maxim of Ratifiability does not provide a sound rationale for choosing efficaciously in Twin’s Dilemma and other Newcomb problems (even granting Screening). To justify the efficacious choice one must augment the Maxim with explicitly causal principles. In reality, then, the evidential ratificationist’s rationale for choosing the dominant act in Twin’s Dilemma and other Newcomb problems relies on a hidden appeal to causal decision theory! To see the lacuna in the ratificationist argument note first that, as Jeffrey stressed, the Maxim of Ratifiability is a only necessary condition; it merely gives agents a way of ruling out acts that cannot be rationally performed. As a necessary condition it has a great deal of plausibility. Unratifiable acts do seem genuinely defective: if one cannot choose to perform A without thereby giving oneself a compelling reason not to perform A, then one should not choose A. This necessary condition, however, is too weak to justify the efficacious choice in Twin’s Dilemma. Ruling out R as a rational choice is not the same as showing that ∼R is the rational choice. First, it could be that Synthese (2007) 156:537–562 547 there is no rational choice in Twin’s Dilemma.9 Second, if ∼R really is the unique rational choice, then it should be possible to explain why by showing that ∼R is better than R is at achieving the agent’s ends. To put the point in a more general way, evidential ratificationists seem to be imagining rational choice as a two-stage process. First, the Maxim of Ratifiability separates those acts that can be ratified from those that cannot. Second, the agent chooses from among her ratifiable options on the basis of expected utility considerations. The fact that some act outside the ratifiable set may have a higher expected utility than any act within the set is deemed irrelevant. Unratifiable acts are simply left out of the picture; they are treated like actions that the agent has lost the power to perform (though, it is crucial to keep in mind, she has not lost this power). Thus, the evidential ratificationist’s argument requires the following strengthening of the Maxim of Ratifiability: Strong Ratificationism. If an agent faces a decision that contains at least one ratifiable act, then she should choose a ratifiable act whose expected utility is not exceeded by that of any other ratifiable act. The chosen act’s expected utility is not required to meet or exceed that of all the unratifiable acts. This ensures that there will be a rational choice in any decision with at least one ratifiable act. Moreover, when there is exactly one ratifiable act, as in Twin’s Dilemma, then that act is the unique rational choice no matter what its expected utility might be. The challenge for evidentialists is to show why Strong Ratificationism is reasonable by explaining why the fact that a given act has maximal expected utility among the ratifiable options makes that act choiceworthy even though it does not maximize utility tout court. The core of the problem is that an act like ∼R in Twin’s Dilemma, though ratifiable, is still lousy news, which means that the agent faces a choice between an unratifiable act of high news value and a ratifiable act of low news value. It looks as though an evidentialist should read this as an indictment of both acts: R is a bad choice in virtue of being unratifiable, while ∼R is a bad choice in virtue of failing to maximize news value. To rebut this charge, evidentialists must explain how, in light of R’s unratifiability, the fact that ∼R is ratifiable provides the agent with a reason for doing ∼R that is sufficient to override or outweigh its low news value. ∼R’s ratifiability must somehow be portrayed as a reason in favor of performing it. Twin’s Dilemma would seem to afford evidentialists with their best chance of doing this because its one ratifiable act ∼R is also uniformly ratifiable: it uniquely maximizes news value conditional on every decision that she might make, so that V(A/dB) > V(A∗ /dB) for all acts A∗  = A and B. In this special case, Strong Ratifiability reduces to SR∗ . If the set of an agent’s possible options contains an option A whose performance is preferable to all others conditional on any decision that she might make, then A is her uniquely rational choice. Since ∼R is uniformly ratifiable, evidentialists can close the hole in their argument by supplying a compelling justification for SR*. At first glance, it seems easy to justify SR*. Indeed, there is a plausible rationale for the Maxim of Ratifiability that extends to SR* almost without modification. Why 9 For an example, see the “Death in Damascus” example in Gibbard and Harper (1978). 548 Synthese (2007) 156:537–562 should “one choose for the person one expects to be once one has chosen?” The answer seems obvious: it is because one will be better informed at that time. The main difference between a decision maker who has yet to settle on a definite course of action and her “post-choice self” is that the latter knows her choice, and so possesses more relevant information about the consequences of her actions. Generally, a person with more relevant information about a decision is better positioned to make a wise choice than a person with less information. Thus, it seems reasonable to introduce the following general principle: Preference Reflection.10 A rational agent should conform her preferences among acts to those of her better-informed post-decision self, at least insofar as she can determine what those preferences are. Now, if an act is unratifiable, the agent knows she will rue her decision to perform it. It would therefore be unwise for her to perform an unratifiable act since this would be to do something that her better-informed post-choice self is certain to regard as ill advised. So, it is irrational to choose unratifiable acts. In this way, the Maxim of Ratifiability is justified on the basis of the principle that rational agents should be guided by the preferences of their better-informed post-decision selves. To extend this justification to cover SR*, note first if an action is uniformly ratifiable then the agent is sure to prefer it after making her decision, no matter what decision she makes. Since the agent should be guided by her better-informed post-choice self, and since her post-choice self is sure to prefer the uniformly ratifiable act, it follows that she should perform this act. It seems, then, that we have found a justification for SR* based, again, on Preference Reflection. To evaluate this justification we must clarify the import and status of Preference Reflection. To this end, imagine two agents, “Doer” and “Knower,” who are considering the same decision. Doer is charged with choosing an act, but both agents will receive whatever outcome results from her choice. We assume that (a) (b) (c) Both agents have identical interests—their utilities for outcomes are the same— and Doer knows this.11 Each agent evaluates prospects on the basis of the same sort of subjective expected utility (evidential or causal)—and Doer knows this. Knower is better situated epistemically—he knows everything Doer knows, but has reliable information Doer lacks—and Doer knows this. When (c) holds Doer thinks that Knower’s beliefs are likely to be more accurate than her own. In the terminology of Gaifman (1988), she regards Knower as an epistemic expert in the sense that, given their respective subjective probabilities, prob for Doer 10 This name should remind readers of the Bas van Fraassen’s Reflection Principle, which is defended in his (1984). 11 Note that in a well-formed decision problem, the utilities of outcomes do not change when an agent acquires new information (though the expected utilities of less specific prospects might well change). In the current context, this means that neither Doer nor Knower will change their utilities for outcomes when they acquire information about what the other believes or prefers. Synthese (2007) 156:537–562 549 and PROB for Knower,12 the following holds • For any proposition X (relevant to the decision) and any number 0 ≤ x ≤ 1, prob(X/PROB∗ (X) = x) = x whenever prob(PROB∗ (X) = x) > 0.13 This says that Doer aligns her subjective probabilities with Knower’s to the extent that she can determine what Knower’s subjective probabilities are. In addition to seeing Knower as an epistemic expert, Doer might also regard him as a decisionmaking expert in the sense that: • For any acts A and B, Doer prefers A to B conditional on the hypothesis that Knower prefers A to B. This entails that Doer unconditionally prefers A to B whenever she is convinced that Knower prefers A to B.14 It is easy to find situations in which it would be reasonable for Doer to view Knower as an epistemic expert. Suppose, for example, that Doer knows that Knower’s epistemic state differs from her own only insofar as he has undergone a learning experience15 that has revealed the truth-value of some proposition Y about which she remains uncertain. Since Doer is then certain both that PROB*(•) = prob(•/Y) if Y and that PROB*(•) = prob(•/∼Y) if ∼Y, it follows that she regards Knower as an epistemic expert.16 When (a) and (b) hold, Doer should view Knower as a decision-making expert, and so emulate his preferences, whenever she regards Knower as an epistemic expert. For Doer can reason as follows: “Since Knower and I have the same interests and are equally rational, and since he knows more than I do about the decision I face, he is clearly better positioned than I am to see which act is best for me. Hence, I should want to do whatever he would want me to do (to the extent that I can determine what this is).” The upshot of this reasoning is encapsulated in the following principle, which is a more precise version of Preference Reflection: K When (a) and (b) hold, if Doer regards Knower as an epistemic expert, then she should also regard him as a decision-making expert. 12 Hereafter the “*” indicates non-rigid designation. Whereas “PROB(X) = x” expresses a (necessary) identity between two numbers, “PROB*(X) = x” expresses the contingent identity that the probability (whatever it is) that Knower assigns to X is the number x. One needs to be careful about rigid and non-rigid designation here to avoid Miller’s fallacy. See van Fraassen (1984). 13 For the sake of simplicity, I am leaving a few things out of this definition. For instance, Doer must have perfect knowledge of her own subjective probabilities, and must be convinced that Knower has perfect knowledge of his subjective probabilities. This requirement is important in contexts where experts might be self-undermining in the sense that learning the probabilities that they assign to propositions will alter those very probabilities. These cases need not concern us here. 14 Doer can regard Knower as an epistemic or a decision-making expert even though she does not know what he believes or prefers. In these cases, Doer’s subjective probability for a proposition X will be her expectation of Knower’s subjective probability for X, and her expected utility for any act A will be her expectation of Knower’s expected utility for A. 15 Doer must see this as an experience that increases the accuracy of Knower’s beliefs. Belief changes that are not the result of such learning experiences will not lead Doer to regard Knower as an epistemic expert. For a discussion of learning experiences see Skyrms (1987). 16 Proof: Since Doer knows her own subjective probabilities, prob(PROB*(X) = x) > 0 only when x is either prob(X/Y) or prob(X/¬Y). Since prob(PROB*(X/Y) = prob(X/Y) iff Y) = 1 it follows that prob(X/ PROB*(X/Y) = prob(X/Y) iff Y) = prob(X/Y). The same holds true with Y replaced by ∼Y. 550 Synthese (2007) 156:537–562 This captures the core of the idea that, all else equal, agents who know more are in a better position to make wise choices. In the special case where Doer regards Knower as an epistemic expert because he undergoes a learning experience that reveals the truth-value of some proposition Y, K becomes K+ When (a) and (b) hold, if Doer knows that Knower’s epistemic state is identical to her own except insofar as he learned Y’s truth-value by a reliable process, then Doer should prefer A to act B if she can deduce that Knower will prefer A to B no matter what he learns about Y’s truth-value. To see K+ in action, consider the following (in my view unassailable) rationale for choosing efficaciously in Twin’s Dilemma.17 Imagine that Row has an identical twin Rosie who is every bit as rational and who stands to gain exactly what Row gains from her decision. Unlike Row, who has no direct knowledge of Column’s choices, Rosie is observing Column through a one-way mirror. She is also receiving continuous updates about the state of Row’s deliberations, so Rosie knows every relevant fact that Row knows. Moreover, if Column acts before Row completes her deliberations, then Rosie will come to know something Row does not know: C’s truth-value. Finally, suppose that Rosie has the power to send Row a signal that tells her that Column has acted, but without indicating which act he has performed. What should Row do if she receives this signal? The obvious answer is that Row should perform the efficacious act ∼R because she can deduce that this is the act that her better-informed, equally rational, identically interested twin favors. Row can deduce this by reasoning as follows: “Either Rosie will have received the good news that Column chose C or Rosie will have received the bad news that he chose ∼C. In the first case, Rosie will hope that I put the icing on the cake by choosing ∼R. In the second, she will hope that I minimize the damage by choosing ∼R. So, whether she learns C or ∼C, Rosie will want me to choose ∼R. By K+, it follows that I should choose ∼R.” Of course, Rosie is a merely prop in this little drama. Row can easily see that she would prefer ∼R to R if she were to have a learning experience that revealed C or ∼C. Since her “better informed self” would want her to act efficaciously, she should act efficaciously too, just as K+ recommends. Evidential ratificationists offer a superficially similar rationale for choosing ∼R, but in their argument Rosie learns Row’s choice rather than Column’s action. Imagine that Rosie has somehow come to know either dR or d∼R. If Row knows this then, even though she might not know what her decision will be, she can reason as follows: “Either Rosie knows that I will decide on R or Rosie knows that I will decide on ∼R. In the first case, she will have strong evidence for believing C, and will therefore hope I put the icing on the cake by performing ∼R (contrary to my decision). In the second case, she will have strong evidence for believing ∼C, and will hope I mitigate the damage by carrying through on my decision to do ∼R. So, whether she learns dR or d∼R, Rosie will want me to choose ∼R. By K+ , I should choose ∼R.” If we replace Rosie in this scenario by Row’s post-decision self, then this becomes an argument for SP* on the basis of K+. The post-decision Row is a rational agent (we assume) who shares Row’s basic desires, and who has come to learn either dR or d∼R 17 I first heard about this justification from Don Hubin some years ago. He attributed it to J. H. Sobel. Synthese (2007) 156:537–562 551 via deliberative processes.18 Since ∼R is uniformly ratifiable the pre-decision Row can deduce that her post-decision self will prefer it to R whether she learns dR or d∼R. Thus, K+ entails that the pre-decision Row should also prefer ∼R to R, just as SR* says. Thus, SR* is justified on the grounds that an agent’s choices should be guided by the judgments of her post-decision self – an equally rational, identically interested self who is better informed about her decision. In this way, the ratificationist justification for choosing the efficacious act in Twin’s Dilemma comes to rest squarely on K+. Unfortunately for friends of Ratificationism, K+ can fail when the agent is in a position to influence Y’s truth-value by her actions. Indeed, K+ is nothing more than a disguised version of Savage’s sure-thing principle. If (a) and (b) hold, and if Knower’s epistemic state is identical to Doer’s except insofar as Knower has learned Y’s truth-value, then Knower’s beliefs and preferences are identical to those of Doer conditional on either Y or ∼Y. Accordingly, K+ entails that Doer should unconditionally prefer A to B if she can deduce that she will prefer A to B both if she learns Y and if she learns ∼Y. This is just the sure-thing principle! As we have seen, however, the sure-thing principle can fail when the process of learning Y’s truth-value involves making choices and taking acts that have undesirable causal consequences. No one should be tempted by the following reasoning: “I can pay $10 for an influenza vaccination or not. At the end of the flu season I will have learned whether or not I contracted the flu. If learn that I did contract it, then I will prefer having $10 in my pocket to having spent it on the vaccine. If I learn that I did not contract it, then I will also prefer having $10 in my pocket to having spent it on the vaccine. So, by K+, I should forgo the vaccination since my identically interested, equally rational self, who will know more than I do about my prospects for getting the flu, will prefer that I forgo the vaccination.” All decision theorists, be they of causal or evidential persuasion, will see this as a fallacy. The agent’s future self will, in all probability, only come to know that she contracted the flu as a direct causal consequence of the fact that her past self did not take a precaution that would have causally inhibited the illness. In cases like this, where the pre-decision self can manipulate the information that the post-decision self learns, K+ and the sure-thing principle are unsound. So, if the ratificationist rationale for choosing efficaciously in Newcomb problems is to succeed, then K+, and Preference Reflection, must somehow be restricted. What restriction is appropriate? Here, evidentialists and causalists must part ways since consistency requires that each treat K+ the same way she treats the sure-thing principle. Evidentialists must restrict K+ to cases where the agent’s acts provide no evidence about Y’s truth-value, while causalists must restrict it to cases where these acts do not causally influence Y’s truth-value. Both correctly prohibit K+’s use when the agent has the power to manipulate her post-decision knowledge about Y. In these cases, the post-decision self’s extra knowledge is not a useful indicator of the desirability of acts because the content of this knowledge is causally influenced by which action is chosen. Evidentialists are forced to go even farther: they must bar K+’s use in any context where the agent can act to provide her future self with evidence about Y, even contexts where she cannot causally influence Y’s truth-value. In particular, evidentialists must enjoin the use of K+ in any decision problem whose acts are (i) evidentially correlated with Y’s truth-value, but (ii) do not causally influence Y’s truth-value. 18 We are assuming that the deliberative processes that generate the decision constitute a genuine learning experience by which Row ultimately comes to know dR or d∼R. This is a substantive assumption, but we will not question it here. 552 Synthese (2007) 156:537–562 In such cases, they must also reject the strong form of ratifiability reasoning based on SR*. This makes it impossible for evidentialists to consistently deploy SR* in Twin’s Dilemma or other Newcomb problems. In any genuine decision, facts about which act is performed are highly correlated with facts about which decision is made because, with high probability, the decision causes the act. Once we know what the agent did, we have strong evidence about what she decided. In Twin’s Dilemma, for example, R and ∼R provide strong evidence for dR and d∼R in just this way: prob(dR/R) >> prob(dR/∼R). Of course, these disparities in probability are due to the fact that decisions cause the acts, but appeals to such explicitly causal facts are anathema to evidentialists, who are restricted to the use of the evidential version of K+. As we have just seen, however, this version of K+ explicitly prohibits ratifiability reasoning when the propositions that the post-decision self learns are evidentially tied to her acts. Thus, there is no ratificationist argument for the efficacious choice in Twin’s Dilemma that appeals exclusively to evidentialist principles. The causal version of K+ does sanction ratifiability reasoning in Twin’s Dilemma. Since Row’s acts do not cause her decisions—it’s the other way around—the “backtracking” evidential correlations between the acts and decisions do not prevent her from using her post-decision desires as guides to current decisions. So it is the causal decision theorist, not the evidentialist, who is in a position to employ SR* to rationalize the efficacious choice.19 Evidentialists who seek to do so are subtly begging the question by invoking a strengthened version of the Maxim of Ratifiability that can only be justified within the confines of a causal decision theory. 3 Are Newcomb problems genuine decisions? As already noted, Jeffrey ultimately rejected ratificationism as a solution to Newcomb problems (albeit for the wrong reasons). In his final word on the topic, (1993) and (2004), he alleged that Newcomb problems are not really decisions at all because those who face them know too much about their own actions to see themselves as making free choices. In any genuine decision an agent must be able to see her acts as causes of outcomes. This cannot happen in Newcomb problems, Jeffrey maintained, because in any plausible Newcomb problem the agent will be able to explain the correlation between acts and states in terms of the existence of a “deep state” that serves as a common cause of both. In Twin’s Dilemma, the picture is precisely as given in Fig. 1 with the variable CC serving as the “deep state.” This is, at best, a “quasi decision problem” because, in the presence of the deep state, the agent cannot regard her acts as causes of outcomes. Newcomb problems thus do not refute evidential decision theory, which is only meant to apply to “real” decisions in which an agent sees her acts as potential causes of outcomes. In making his argument Jeffrey appeals to a specific proposal, due to Arntzenius (1992), about what it means for an agent to see one event as a cause of another. To say what it is for an agent to see her actions as causes of outcomes we are not required, Jeffrey thinks, to represent her as having explicitly causal beliefs. We need only suppose that her (non-causal) beliefs about her own acts evolve in a specific way during 19 Note, however, that the causal decision theorist does not need to introduce ratifiability considerations to justify ∼R since it is the act that maximizes causal expected utility. Synthese (2007) 156:537–562 553 deliberation. “Imputations of causal influence,” he writes, “are not shown simply by momentary features of probabilistic states of mind, but by intended or expected features of their evolution.” (1993, p. 139) The picture is one in which the decision maker’s subjective probabilities and expected utilities change as a result of deliberation. We can think of our agent’s mental state at time t as being represented by a probability probt and an expected utility dest that jointly obey EDT. During deliberation, the agent’s mental state changes in accordance with a belief/desire revision process that maps an initial probability/utility pair (prob0 , des0 ) through a sequence of temporal stages (probt , dest ), 0  t 1, to a final state (prob1 , des1 ). At t = 0 the agent is undecided about what she ought to do; at t = 1 she has made up her mind. To focus on essentials, let us assume that the process of revision is purely epistemic, so that the agent’s basic desires remain fixed and all changes in (probt , dest ) are ultimately traceable to changes in probt , so that (probt , dest ) = (probt , des0 ). Within this framework Jeffrey’s proposes to analyze “imputations of causation” using Arntzenius’s Test: During the time interval 0  t  1, an agent regards one event C as a promoting (or inhibiting) cause of another event E only if  Correlation. prob0 (E/C) – prob0 (E/∼C) > 0 (or < 0).  Rigidity. probt (E/C) and probt (E/∼C) remain fixed as probt (C) varies.  Variability. prob0 (C) starts out at some intermediate value, and the belief revision process takes prob1 (C) either to zero or to one. Correlation ensures that a promoting cause always provides evidence in favor of its effect. Rigidity requires the belief revision process that alters C’s probability to behave like a Jeffrey shift on {C, ∼C} with respect to the propositions in the set {C & E, C & ∼E, ∼C & E, ∼C & ∼E}. Variability requires that C’s probability change over time. Rigidity then forces E’s probability to change over time as well. The manner in which these changes occur is supposed to determine “imputations of causal influence.” We can start to see why by noting that Arntzenius’s Test introduces a partial asymmetry between causes and their effects. As long as the agent’s probabilities for C and E change over time, as required by Variability, then Rigidity holds only if probt (C/E) and probt (C/∼E) vary with probt (E). This is the result of the following FACT: If probt (E/C), probt (E/∼C), and either probt (C/E) or probt (C/∼E) remain fixed over time, then so must probt (C) and probt (E) because probt (C) = probt (C/E)probt (E/∼C) probt (∼C/E)Probt (E/C) + Probt (C/E)Probt (E/∼C) Probt (E) = Probt (C)Probt (E/C) + probt (C)Probt (E/∼C) In other words, if you fix any three of the four conditional probabilities, then you also fix the unconditional probabilities of E and C as well. Thus, if the conditions of Arntzenius’s Test are satisfied, then it is impossible to both regard C as a cause of E and to regard E as a cause of C. While this does introduce a kind of asymmetry between causes and effects, it does not fully distinguish them since Arntzenius’s Test can hold relative to one sort of 554 Synthese (2007) 156:537–562 belief revision process with C and E as they are, and yet hold relative to another belief revision process with C and E reversed. For example, any time the agent has an E-learning experience modeled by a Jeffrey shift probt (•) = probt (E)probt (•/E) + probt (∼E)probt (•/∼E) the conditional probabilities probt (C/E) and probt (C/∼E) remain fixed as probt (E), probt (C), probt (E/C) and probt (E/∼C) all vary. Suppose, for example, that I know that C, an excess of bilirubin in the blood, causes E, a characteristic yellow hue of the skin (when viewed in sunlight). When looking at you in poor light I might have a learning experience in which my level of confidence in E increases from 0.2 to 0.4 while my probabilities for you having excess bilirubin conditional on having or not having yellow skin remain the same. If the light slowly improves and I go through a series of such Jeffrey shifts until I become certain that you have yellow skin, then my subjective probabilities satisfy the conditions of Arntzenius’s Test with E in for the putative cause and C in for the putative effect. This is no refutation of the Arntzenius’s Test, which is only intended to be a necessary condition anyhow, but it does show that one needs to be careful when applying it. The test can only reliably distinguish causes from effects when it is restricted to specific sorts of belief revisions. In the bilirubin case the changes in probabilities for C and E were driven by learning experiences involving their effects, specifically the effects of E. Arntzenius’s Test works best when the changes in probabilities for C and E are driven by learning experiences involving their causes. I do not think that Jeffrey should find anything objectionable in this reading. Indeed, he is quite clear about the fact that in decision-making contexts there is a “driving mechanism.” In decision-making it is deliberation, not observation, that changes your probabilities. To think that you face a decision problem rather than a question of fact about the rest of nature is to expect whatever changes arise in your probabilities. . . during your deliberations to stem from changes in your probabilities for choosing options. . . . As a decision maker you regard probabilities of options as inputs driving the mechanism, not driven by it. (Jeffrey, 1993, p. 8) Jeffrey’s claim, then, is that Arntzenius’s Test tells us when the agent genuinely regards her acts as causes of outcomes in cases where changes in her subjective probabilities for actions are directly induced by processes of rational deliberation. Thus, we have Jeffrey’s Test: During the time interval 0  t  1, an agent regards an act A as a promoting (or inhibiting) cause of an outcome A & E only if  Correlation. prob0 (E/A) – prob0 (E/∼A) > 0 (or < 0).  Rigidity. probt (E/A) and probt (E/∼A) remain fixed as probt (A) varies.  Variability. prob0 (A) starts out at some intermediate value, and is driven by the process of rational deliberation to a state in which prob1 (A) is either to zero or to one. Since an agent facing a genuine decision must see her acts as causes of outcomes, Jeffrey requires that the agent’s subjective probabilities satisfy this test for each act A and state E in the decision under consideration. Synthese (2007) 156:537–562 555 It is crucial to Jeffrey’s view that the deliberative process can alter an agent’s act probabilities. Her initial state of indecision is modeled by a time-0 probability that does not assign any act a probability close to one. As a result of deliberation, this indecisive state is eventually replaced by a time-1 state in which the probability of the chosen act is one or nearly one. In effect, deliberation is a process by which the agent learns what she will ultimately do. Skyrms (1990) provides a formal model of deliberation in which the agent iteratively revises her beliefs in light of information about the expected utilities of acts, and then recalibrates these utilities in light of her revised beliefs. She uses her t = 0 probabilities to calculate expected utilities for A, ∼A, and the “status quo” A∨∼A, and raises her degree of confidence in any act whose expected utility exceeds that of the status quo. The procedure is repeated, using the revised probabilities for A and ∼A as inputs, until some stable equilibrium is reached. Changes in act probabilities seek the good in the sense that probt+1 (A) > probt (A) whenever the time-t expected utility of A exceeds that of A ∨ ∼A.20 The character of these direct changes in probability will, of course, depend on how expected utilities are computed. For Jeffrey, A’s probability increases between t and t + 1 just in case A is good news at t. For a causal decision theorist, such an increase will occur only if A is judged to be more effective than ∼A at causing desirable outcomes. No matter what sort of expected utility calculation is driving these changes in act probabilities, they ramify through the agent’s beliefs via Jeffrey conditioning, so that probt+1 (•) = probt+1 (A)probt (•/A) + probt+1 (∼A)probt (•/∼A).21 Rigidity is then automatically satisfied. This has a number of implications. First, any correlations that exist between states and acts are preserved, so that probt (E/A) = prob0 (E/A) for all states E, acts A, and times t. Second, since we are assuming that the utilities of outcomes do not change over time (though the utilities of other propositions might), the news values A and ∼A remain fixed at their time zero values as t varies. Thus, deliberation, in the sense captured by the model, is not a process by which an agent revises her views about news values of actions: she merely revises her views about acts’ probabilities while keeping their news values fixed. These changes induce modifications in her estimates of the news value of the “status quo”, Vt (A ∨ ∼A), which varies until t = 1, at which point it coincides with V0 (A) or V0 (∼A) depending upon whether the former or latter is greater. The agent will then have made up her mind: she will have decided to perform the act that maximizes her time-0 news value, and will have come to assign this act a probability near one. The upshot is that in any genuine decision in which changes in act probabilities are driven by considerations about news values a rational agent will settle on the act sanctioned by evidential decision theory. 20 Skyrms and others have discussed a number of belief revision rules with this property. One is Nash rule, which, in the case of evidential decision theory, sets probt+1 (A) = (probt (A) + k)/(1 + k) where k = (Vt (A) – Vt (A ∨ ∼A)) > 0. Another is the Darwin rule, which sets probt+1 (A) = k probt (A)/Vt (A ∨ ∼A). 21 Admittedly, this is only (an idealization of) a small part of the deliberative process. When speaking of deliberation we often have in mind the whole raft of psychological processes by which agents come to arrive at decisions. Parts of this process involve the decision maker coming to have settled views about desirabilities of outcomes and probabilities of states of the world. Deliberation, in this broader sense, thus includes what fixes, or makes us aware of, the utilities of outcomes and the probabilities of states of the world that are relevant to our choices. This is not what the Jeffrey/Skyrms model is meant to capture. It presents us with an account of only the end of the deliberative process by explaining how an agent who already possesses fixed views about the desirabilities of outcomes and the probabilities of states is able to move from being undecided about her best option to having chosen an action. 556 Synthese (2007) 156:537–562 It follows that Newcomb Problems cannot be genuine decisions. These problems require the agent to know a great deal about the conditional probabilities of her acts given the world’s state, e.g., Row must knows that she is likely to do R if Column does C. If Row knows enough to fix sharp values for prob0 (R/C) and prob0 (R/∼C), and if these values are, as Jeffrey claims, “fixed once given” (1993, p. 141), then the FACT mentioned above entails that she cannot satisfy both Rigidity and Variability. If Rigidity holds, so that probt (C/R) and probt (C/∼R) remains fixed for all t, then probt (A) = prob0 (A). In a symmetric Twin’s Dilemma, for example, Row’s initial subjective probabilities might be prob0 (C/R) = prob0 (R/C) = 0.9 and prob0 (C/∼R) = prob0 (R/∼C) = 0.3, which mandates prob0 (A) = prob0 (C) = 0.875. Thus, enforcing Rigidity in Twin’s Dilemma entails that Row’s act probabilities cannot vary during her deliberations; her beliefs about her own acts are effectively hemmed in by the evidence that she has about the correlations between her acts and states of the world. According to Jeffrey, this shows that Row is not facing a real decision. Her choice is not free because her act probabilities are not free to vary; she knows too much about the correlations between her actions and states of the world to see her acts as causes of outcomes. If this is right, then evidential decision theory has nothing to fear. According to Jeffrey’s test, in any bona fide decision problem both probabilities of outcomes conditional on acts and news values of acts must remain constant throughout deliberation. Accordingly, an agent can confidently use an act’s news value as a “figure of its merit” (1993, p. 140) because she can be sure that its present news value will coincide with the news value she will accord to it after she deliberates herself into a state of certainty about what she will do. Newcomb problems fail as counterexamples to evidentialism because they are not really decisions at all! While there is some plausibility to this argument, it contains a subtle flaw. To ferret it out, we need to understand how an agent’s beliefs about her own acts are related to her decisions. Jeffrey’s argument rests on the claim that the conditional probabilities probt (R/C) and probt (R/∼C) are “fixed once given” because they are constrained by Row’s antecedent evidence concerning the correlations between her actions and those of Column. But many decision theorists (both evidential and causal) have suggested that free agents can legitimately ignore evidence about their own acts. Judea Pearl (a causalist) has written that while “evidential decision theory preaches that one should never ignore genuine statistical evidence. . . [but] actions—by their very definition—render such evidence irrelevant to the decision at hand, for actions change the probabilities that acts normally obey.” (2000, p. 109) Pearl took this point to be so important that he rendered it in verse: Whatever evidence an act might provide On facts that precede the act, Should never be used to help one decide On whether to choose that same act. (2000, p. 109) Huw Price (an evidentialist) has expressed similar sentiments: “From the agent’s point of view contemplated actions are always considered to be sui generis, uncaused by external factors. . . This amounts to the view that free actions are treated as probabilistically independent of everything except their effects.” (1993, p. 261) A view somewhat similar to Price’s can be found in Hitchcock (1996). These claims are basically right: a rational agent, while in the midst of her deliberations, is in a position to legitimately ignore any evidence she might possess about what Synthese (2007) 156:537–562 557 she is likely to do. She can readjust her probabilities for her currently available acts22 at will, including her probabilities for acts conditional on states of the world. Row can, for instance, say to herself, “though I have overwhelming evidence for thinking that Column and I are going to make the same choice, I am nevertheless free to believe anything I want about whether or not I am going to do R in the event that he does C.” If this is right, then there is then no conflict between Rigidity, Variability and the idea that Row has enough evidence to fix values for prob0 (R/C) and prob0 (R/∼C). Jeffrey’s mistake was to suppose that these latter probabilities are “fixed once given.” He thought this, I suspect, because he felt that the strength of the agent’s opinions about her own acts should, like her opinions about other matters, be proportioned to her antecedent evidence for them. Pearl, Price, Hitchcock and I all deny this. A deliberating agent who regards herself as free need not proportion her beliefs about her own acts to the antecedent evidence that she has for thinking that she will perform them. Let’s call this the evidential autonomy thesis. It is clear that Jeffrey’s attempt to exclude Newcomb problems from the realm of decision theory fails if we accept the autonomy thesis. One might, however, wonder why is it legitimate for deliberating agents to ignore evidence about their own actions when it is not legitimate for them to ignore evidence about other matters? While Row may ignore evidence about the causes of her acts, she may not ignore the evidence that her acts provide for their effects. What makes the difference? On the face of things, it seems unreasonable for an agent to ignore evidence for any belief, even a belief about what she will do. So, it seems as if the evidential autonomy thesis runs afoul of the requirement of Proportionism. An person’s level of confidence in a proposition X should always be proportioned to her balance of evidence for X’s truth, in particular it should be responsive to evidence about the causes of X. To refute Jeffrey one must either explain why a deliberating agent’s beliefs about her own currently available acts are not bound by this requirement, or to show that, contrary to appearances, they do not violate it. Neither Pearl, Price nor Hitchcock says as much as one would like on this topic. I shall try to do better by showing how a deliberating agent’s beliefs about her own actions can be justified on the basis of the totality of her evidence at the time of her decision. One initially tempting strategy for doing this has affinities with Jeffrey’s ratificationism and with Eells’s “tickle defense.” The idea would be to argue that, at the moment of choice, Row can ignore all evidence about correlations between her acts and Column’s acts because at that point she will have learned her decision, so that either prob1 (dR) = 1 or prob1 (d∼R) = 1, and this new knowledge will screen off all evidence about prior causes or correlations. Unfortunately, this strategy merely replaces the question of how changes in Row’s beliefs about A and ∼A can be justified with the question of how changes in her beliefs about dR and d∼R can be justified. Jeffrey will insist that, insofar as Row sees her decisions as causes of her acts, the conditional probabilities x = probt (R/dR) > y = probt (R/d∼R) will remain constant during deliberation while probt (dR) and probt (d∼R) vary. But, since probt (dR) = (probt (R) – y)/(x – y) this prevents probt (dR) from varying during deliberation unless probt (R) varies 22 It is important to understand that this freedom only extends to propositions that describe actions about which the agent is currently deliberating, and whose performance she sees as being exclusively a matter of the outcome of her decision. It does not, for example, apply to acts that will be the result of future deliberations. 558 Synthese (2007) 156:537–562 Fig. 4 as well. So, just as in the case of A, we are left with the task of explaining why it is legitimate for Row to change her opinions about dA during deliberation even though she has enough evidence to determinately fix a value for probt (dR). The general point is that, insofar as she regards her decision a cause of her act, Row’s beliefs about the former will be evidentially tied to her beliefs about the latter. Thus, with respect to autonomy from evidence, Row’s beliefs about A and her beliefs about dA are in the same boat. We can reconcile the autonomy thesis with the requirement that beliefs be proportioned to the evidence by recognizing that an agent’s beliefs about her own free decisions and actions provide evidence for their own truth. Normally, the fact that a person believes a proposition X says nothing about X’s truth-value that is not indicated by evidence she already possesses. This is because, in normal cases, (a) the person’s level of confidence in X is, at least partly, a result of her having the evidence she does, and (b) X’s truth or falsity either causes her to have this evidence or is tied to it as joint effects of a common cause. So, normal cases look like this (Fig. 4). Situations depicted on the left are quite familiar, as when the fact that the sun is shining, together with my standing outside, causes me to see my shadow, and thereby to believe that the sun is shining. Those on the right are common as well. For example, the fact that a coin has a certain physical constitution, combined with my observing it being tossed 100 times, might cause me to have the evidence that eighty of the last 100 tosses have been heads, and thus to be highly confident that the coin will come up heads between 65 and 95 times in the next 100 tosses. Neither (a) nor (b) hold in the situations that interest us. In these cases there is a causal path running between the belief and the proposition believed that is not mediated by the believer’s evidence, and the existence of this path makes it possible for the person to use the belief as evidence for its own truth. There are two relevant cases of this type.23 In the first, the person’s belief about X may not be fully responsive to her evidence for X, e.g., it might also depend on her desire for X’s truth. While this typically indicates an epistemic failing, it need not be problematic when the belief in question causally contributes to its own truth. For example, as William James argued, wishful thinking can be legitimate for beliefs that are self-fulfilling prophecies, i.e., when the fact that a person believes X to a sufficiently high degree causally promotes X’s truth. When there is a “power to positive thinking” in this way, a believer can acquire evidence in favor of a belief solely in virtue of holding it. This added evidence must be balanced off against the evidence she already has, but its general effect will be to justify a higher level of confidence. For example, it might be that those who stutter 23 A third sort of case, in which X directly causes the agent to believe that X, seems irrelevant to decision theory. If the agent learns that she will perform A as a causal consequence of A performance, then it is hard to interpret A as being the result of a free decision. Synthese (2007) 156:537–562 559 have fewer problems speaking when they believe they will not stutter than when they believe they will stutter. If so, then a stutterer who has strong inductive evidence for thinking that he will stutter the next time he speaks might nevertheless be in a position to reasonably believe that he will speak fluently. For, it might be that in the past he has generally believed that he will stutter. So, if he can somehow convince himself that he will not stutter this time, which he may or may not be able to do, then the fact that he has this belief will actually make stuttering less likely. If this “confidence effect” is great enough, then it can be permissible, from a purely epistemic perspective, for the person to believe that he will not stutter. Even thought all the evidence points the other way before he comes to hold the belief, the evidence he comes to have after adopting the belief might well tip the balance. Conditions (a) and (b) can also fail when X’s truth-value and the belief that X are joint effects of a common cause whose occurrence is only know to the person in virtue of being aware of the fact that she believes X. To illustrate, suppose that, in contrast with the story of the last paragraph, each incident of stuttering is caused by a specific brain event that also causes the stutterer to lose confidence in his ability to speak fluently, a confidence that is otherwise present. The stutterer’s belief that he will/will not stutter can then provide strong evidence for itself since its presence is a reliable indicator of the presence/absence of the underlying mental event. The following diagram depicts the “confidence effect” on the left, and the “hidden common cause” on the right. (The broken arrow indicates a weak or non-existent causal connection (Fig. 5)). In each scenario, it is reasonable for a believer who knows the causal structure of the situation to invoke her high degree of confidence in X as evidence for X’s truth. Her ability to do this does not depend on the character of her background evidence, except insofar as this evidence relates to her knowledge of the causal structure. Call beliefs that arise in these situations self-supporting. When the possibility of self-support is taken into account, apparent violations of Proportionism can be benign. If the “confidence effect” is strong enough, then the wishful thinker’s belief might be properly proportioned to her total evidence for it, which includes the fact that she holds the belief, even though it is not properly proportioned to the evidence she had before coming to hold the belief. The same thing can happen when a believer recognizes that her high degree of confidence in a proposition indicates the occurrence of a common cause of both it and the truth of the proposition believed. On the model of deliberation considered here, there is a deep connection between decision-making and self-supporting beliefs. The model has the agent becoming more confident in both A and dA upon learning that she prefers A to the status Fig. 5 560 Synthese (2007) 156:537–562 quo. Why isn’t this just wishful thinking? The answer is that it is wishful thinking, just not wishful thinking of the fallacious sort because the beliefs involved are selfsupporting. There are a number of “metaphysical” accounts of decision-making that make A and/or dA self-supporting. They key to all of them is the idea that there is a causal pathway, not mediated by the agent’s evidence, that runs between agent’s beliefs about A and dA and her performance of the act A. Rather than trying to decide which of these views is correct, I will simply sketch out some of the possibilities so as to make clear what the options are. One might hold a view on which an agent’s beliefs about her own acts are selfsupporting in virtue of being self-fulfilling prophecies. David Velleman defends such a position in his (1989). For our purposes, Velleman’s crucial insights are these:     An agent’s beliefs about her own potential acts can, under conditions of deliberation, play a direct causal role in her behavior, and so become self-fulfilling prophecies. An agent who sees herself as acting freely will recognize that her beliefs about her own actions have this “self-fulfilling” character. This creates a kind of “epistemic freedom” that can make it reasonable for the agent to adopt beliefs about her own actions that run contrary to other evidence that she might have about what she is likely to do. The basis of this epistemic freedom is the fact that by believing that she will do a certain act A an agent creates the evidence that justifies this belief. If all this is true, then an agent’s beliefs about her own free actions can be both evidentially autonomous and correctly proportioned to the evidence. For example, even though Row’s prior evidence requires that prob0 (C/R) = prob0 (R/C) = 0.9 and prob0 (C/∼R) = prob0 (R/∼C) = 0.3, and so mandates that prob0 (R) = prob0 (C) = 0.875, she can still become justifiably certain of ∼R during deliberation because ∼R’s objective probability grows with increases in her confidence in ∼R. Even if one does not agree with Velleman that an agent’s beliefs about her own acts directly cause those acts, one might think this is true for beliefs about decisions. One might think, that is, that the direct cause of A is the agent’s decision to A, and that the direct cause of her decision to A is her belief that she has decided to A. The agent’s belief that she will A will not then be a cause of A, but, since the agent’s beliefs about A are directly tied to her beliefs about dA, it will be connected to A by a causal chain that is not mediated by background evidence. The last bit of the causal chain will then look like this: A ← dA ← prob1 (dA) → prob1 (A). Here, the agent’s belief that A is not itself self-fulfilling, but it is justified by a belief that is both self-fulfilling and a cause of A. Again, the requirement of evidential autonomy is met without violating Proportionism.24 An alternative way of reconciling evidential autonomy and Proportionism is to maintain that an agent’s beliefs about her own acts are reliable indicators of some hidden, immediate common cause of both the act and the belief. On this view, Row’s changing degrees of confidence in R and ∼R (or dR and d∼R) track vacillations in some aspect of her underlying mental state, we might call it her state of indecision, that has direct causal influence over what she will ultimately do. If she eventually becomes certain that she will do ∼R, this indicates that she has come to rest in a 24 Even more generally, this will be true if there is a proposition X such that the agent rightly believes that causal relationships are such that A ← X ← prob1 (X) → prob1 (A). Synthese (2007) 156:537–562 561 decisive state that causes ∼R. She can then use the fact that she is certain about ∼R as evidence for the conclusion that she is in such a state, thereby justifying that very belief. Again, the belief is autonomous of believer’s background evidence, but it does not run afoul of Proportionism because it is self-supporting. This brief discussion has indicated a few ways in which an agent’s beliefs about her own current actions can be self-supporting. For our purposes, it is less important to know which one of these accounts is correct than it is to recognize that the selfsupporting nature of act beliefs makes it legitimate for agent to ignore any antecedent evidence she might possess about what she is likely to choose or to do. Appreciating this makes it clear that, contrary to what Jeffrey suggests, the values of probt (R/C) and probt (R/∼C) are not “fixed once given.” No matter how much evidence Row might have for the correlation between her acts and Column’s acts, if she regards herself as entirely free in the matter of R, then her own beliefs about the causal structure of her decision situation put in her in a position to legitimately disregard this evidence. Jeffrey’s attempt to disqualify Newcomb problems thus fails. If there is to be a successful argument for the conclusion that these problems are not real decisions it will have but made on grounds other than epistemic ones. The beliefs of Newcomb deciders are not constrained by the evidence at their disposal; in the context of deliberation, free agents can believe what they want about their current acts because such beliefs provide their own justification. References Armendt, B. (1986). A foundation for causal decision theory. Topoi, 5, 3–19. Arntzenius, F. (1992). The common cause principle. Proceedings of the 1992 PSA Conference, Vol. 2, 227–237. Cartwright, N. (1979). Causal laws and effective strategies. Nous, 13, 419–437. Eells, E. (1982). Rational decision and causality. Cambridge, MA: Cambridge University Press. Eells, E. (2000). Review: The foundations of causal decision theory, by James M. Joyce. The British Journal for the Philosophy of Science, 51, 893–900. Gaifman, H. (1988). A Theory of higher order probabilities. In B. Skyrms, & W. Harper (Eds.), Causation, chance, and credence. Dordrecht: Kluwer. Gibbard, A. & Harper, W. (1978). Counterfactuals and two kinds of expected utility. In C. Hooker, J. Leach, & E. McClennen (Eds.), Foundations and applications of decision theory (pp. 125–162). Dordrecht: Reidel. Hitchcock, C. (1996). Causal decision theory and decision-theoretic causation. Nous, 30, 508–526. Jeffrey, R. (1983). The logic of decision, 2nd edn. Chicago: The University of Chicago Press. Jeffrey, R. (1993). Causality and the logic of decision. Philosophical Topics, 21, 139–151. Jeffrey, R. (2004). Subjective probability the real thing. Cambridge, UK: Cambridge University Press. Joyce, J. M. (1999). The foundations of causal decision theory. Cambridge, UK: Cambridge University Press. Lewis, D. (1981). Causal decision theory. Australasian Journal of Philosophy, 59, 5–30. Pearl, J. (2000). Causality: Models, reasoning, and inference. Cambridge, UK: Cambridge University Press. Price, H. (1993). The direction of causation: Ramsey’s ultimate contingency. In D. Hull, M. Forbes, & K. Okruhlik (Eds.), PSA 1992, Vol. 2, pp. 253–267. East Lansing: Philosophy of Science Association. Reichenbach, H. (1956). The direction of time. Berkeley: University of Los Angeles Press. Savage, L. J. (1972). The Foundations of Statistics. New York: Dover. Spirtes, P. Glymour, C. & Scheines, R. (2000). Causation, prediction, and search, 2nd edn. Cambridge, MA: MIT Press. Skyrms, B. (1980). Causal necessity. New Haven: Yale University Press. Skyrms, B. (1990). The Dynamics of Rational Deliberation. Cambridge: Harvard University Press. 562 Synthese (2007) 156:537–562 Skyrms, B. (1987). The value of knowledge. In C. Wade Savage (Ed.), Justification, discovery, and the evolution of scientific theories. Minneapolis: University of Minnesota Press. Sobel, J. H., (1985). Circumstances and dominance in a causal decision theory. Synthese, 62, 167–202. van Fraassen, B. (1984). Belief and the will. Journal of Philosophy, 81, 235–256. Velleman, J. D. (1989). Epistemic freedom. Pacific Philosophical Quarterly, 70, 73–97. Reproduced with permission of the copyright owner. Further reproduction prohibited without permission.