RISK AND TRADEOFFS by Lara Buchak, UC Berkeley1 1. Introduction Decision theories are theories of instrumental rationality: they formalize constraints of consistency between rational agents’ ends and the means they take to arrive at these ends. We can model the possible actions an agent might take each as a gamble whose outcomes depend on the state of the world: for example, the action of not bringing an umbrella is a gamble that results in getting wet if it rains and staying dry if it doesn’t. Decision theory places constraints on the structure of rational agents’ preferences among the actions available to them and, as a result, can represent the beliefs and desires of any agent who meets these constraints by precise numerical values. The prevailing view is that subjective expected utility theory, which dictates that agents prefer the gamble with the highest expected utility, is the correct theory of instrumental rationality. Subjective expected utility theory (hereafter, EU theory) is thought to characterize the preferences of all rational decision makers. And yet, there are some preferences that violate EU theory that seem both intuitively appealing and prima facie consistent. An important group of these preferences stem from how ordinary decision makers take risk into account: ordinary decision makers seem to care about “global” properties of gambles, but EU theory rules out their doing so. If one is sympathetic to the general aim of decision theory, there are three potential lines of response to the fact that EU theory does not capture the way that many people take risk into account in their preferences among gambles. The first is to claim that contrary to initial appearances, expected utility theory can represent agents who care about global properties, by re-describing the outcomes that these agents face. The second response is to claim that while many people care about global properties (and that these patterns of preferences cannot be represented by the theory), these people are simply not rational in doing so. I think that neither of these responses can succeed. I advocate a third response: modifying our normative theory to broaden the range of rationally permissible preferences. In particular, I advocate broadening the set of attitudes towards risk that count as rationally permissible. Although I won’t directly argue against the first two responses here, formulating an alternative will be important to evaluating them. We need to know what it is that agents are doing when they systematically violate EU theory in order to discover whether doing what they are doing constitutes taking the means to their ends in a rationally permissible way. In this paper, I will explain the alternative to EU theory that I favor, and I will in particular explain how it does a better job at explicating the components of instrumental rationality than does EU theory. 2 2. Risk, EU Theory, and Instrumental Rationality I begin by briefly explaining subjective expected utility theory; explaining how it must analyze the phenomenon of risk aversion; and showing that as a result, EU theory cannot capture certain preferences that many people have. I will then argue that this problem arises for EU theory because it neglects an important component of instrumental rationality. EU theory says that rational agents maximize expected utility: they prefer the act with the highest mathematical expectation of utility, relative to their utility and credence (subjective probability) functions. So, if we think of an act as a gamble that yields a particular outcome in each state of the world—for example, g = {E1, x1; E2, x2; …; En, xn} is the act that yields xi if Ei obtains, for each i —then the value of this act is: = According to EU theory, a rational agent strictly prefers f to g if and only if EU(f) > EU(g), and if she weakly prefers f to g if and only if EU(f) ≥ EU(g). So utility and credence are linked to rational preferences in the following way: if we know what an agent’s utility function and credence function are, we can say what her preferences ought to be. They are also linked in another way that will be of central interest in this paper: if we know an agent’s preferences, and if these preferences conform to the axioms of EU theory, then we can determine her credence function uniquely and her utility function uniquely up to positive affine transformation:2 we can represent her as an expected utility maximizer relative to a some particular p and u. It is crucial for the EU theorist that the preferences of all rational agents can be represented in this way. It is uncontroversial that many people’s preferences display risk aversion in the following sense: an individual would rather have $50 than a fair coin-flip between $0 and $100, and, in general, would prefer to receive $z rather than to receive a gamble that will yield $z on average.3 If such an agent is representable as an EU maximizer and the preference for $50 rather than the coin-flip is strict, then it must be that u($50) – u($0) > u($100) – u($50), i.e., that getting the first $50 makes more of a utility difference than getting the second $50 does. More generally, her utility function in money must diminish marginally: 3 Diminishing Marginal Utility 100 80 Utility (utils) 60 40 20 0 0 20 40 Diagram 1: Diminishing Marginal Utility 60 Money ($) 80 Diagram 1: Diminishing Marginal Utility 100 The converse holds as well: when a utility function is concave, a gamble will always have an expected utility that is no greater than the utility of its expected dollar value.4 On EU theory, then, aversion to risk is equivalent to diminishing marginal utility. Intuitively, though, there are two different psychological phenomena that could give rise to riskaverse behavior. On the one hand, how much one values additional amounts of money might diminish the more money one has. As an extreme case of this, consider Alice, who needs exactly $50 for a bus ticket, and doesn’t have much to buy beyond that. On the other hand, one might value small amounts of money linearly, but care about other properties of the gamble besides its average value: for example, the minimum value it might yield, the maximum, or the spread of possible outcomes. In other words, one might be sensitive to global properties. Consider Bob, who gets as much pleasure out of the first $50 as the second, but would rather guarantee himself $50 than risk having nothing for the possibility of $100. Both Alice and Bob prefer $50 to the coin-flip, and the EU theorist must interpret both agents as having a non-linear utility function, on the basis of this preference. What is the relationship between an agent’s psychology and the utility function that is derived from her preferences? There are two views about this: on the realistic picture, the utility function represents some pre-existing value that the agent has; furthermore, this value is independent from preferences, so could in principle be known apart from preferences. For the realist, EU theory will have misinterpreted Bob, since Bob’s strength of desire for money is linear. On the constructivist picture, 4 which seems to have more widespread endorsement among contemporary philosophers, utility cannot be determined independently of what the agent prefers.5 The utility function is a construction from preferences: it is the quantity whose mathematical expectation an agent maximizes. Utility may or may not correspond to some psychological quantity, but even if it does, it is not a quantity that we can give content to apart from preferences. So the constructivist won’t necessarily care about the agent’s own claims about how much she values various outcomes. There is nothing yet in Alice’s and Bob’s preferences to distinguish them from the point of view of constructivist EU theory. However, an agent’s reasons for the preferences she has will turn out to matter, because if the constructivist EU theorist accidentally commits himself to something false about an agent’s reasons, then although he will be able to explain an isolated preference (such as that for $50 over the $0/$100 coin-flip), his explanation will commit the agent to having preferences that she does not in fact have and will therefore fail to represent her. Matthew Rabin presents a “calibration theorem” to show that in order to describe the preferences of decision makers that display risk aversion in modest-stakes gambles, EU theory is committed to absurd conclusions about preferences between gambles when the stakes are higher (absurd in the sense that no one actually has these preferences). As mentioned, on EU theory, modest stakes risk-aversion entails a concave utility function. Rabin’s results assume nothing about the utility function except that it continues to be concave in higher stakes, and so doesn’t, for example, have an inflection point above which it increases marginally. Here are some examples of the sorts of conclusions EU theory is committed to.6 If an agent prefers not to take a fair coin-flip between losing $100 and gaining $110 (that is, if she prefers a sure-thing $0 to the -$100/$110 coin-flip), regardless of her initial wealth level, then she must also prefer not to take a coin-flip between losing $1,000 and gaining any amount of money. Similarly, if an agent prefers not to take a coin-flip between losing $1,000 and gaining $1050 for any initial wealth level, then she will also prefer not to take a coin-flip between losing $20,000 and gaining any amount of money. Furthermore, if an agent prefers not to take a coin-flip between losing $100 and gaining $105 as long as her lifetime wealth is less than $350,000, then from an initial wealth level of $340,000, she will turn down a coin-flip between losing $4,000 and gaining $635,670. In other words, she will prefer a surething $340,000 to the gamble {$339,600, 0.5; $975,670, 0.5}.7 Rabin’s results are problematic for both the realist and constructivist EU theorist: if most people have the modest-stakes preferences but lack the high-stakes preferences that ‘follow’ from them, then EU theory with a diminishing marginal utility function will fail to represent most people. In case the reader is worried that Rabin’s results rely on knowing a lot of the agent’s preferences, there are also examples of preferences that EU theory (under either interpretation) cannot account for that involve very few preferences. One example is Allais’s famous paradox. Consider Maurice, who is 5 presented with two hypothetical choices, each between two gambles.8 He is first asked whether he would rather have L1 or L2: L1: 10% chance of $5,000,000, 90% chance of $0. L2: 11% chance of $1,000,000, 89% chance of $0. He reasons that the minimum he stands to walk away with is the same either way, and there’s not much difference in his chances of winning some money. So, since L1 yields much higher winnings at only slightly lower odds, he decides he would rather have L1. He is then asked whether he would rather have L3 or L4: L3: 89% chance of $1,000,000, 10% chance of $5,000,000, 1% chance of $0. L4: 100% chance of $1,000,000. He reasons that the minimum amount that he stands to win in L4 is a great deal higher than the minimum amount he stands to win in L3, and that although L3 comes with the possibility of much higher winnings, this fact is not enough to offset the possibility of choosing L3 and ending up with nothing. So he decides he would rather have L4. Most people, like Maurice, prefer L1 to L2 and L4 to L3. However, there is no way to assign utility values to $0, $1m, and $5m such that L1 has a higher expected utility than L2 and L4 has a higher expected utility than L3; therefore these preferences cannot be represented as maximizing expected utility.9 Allais’s example does not require any assumptions about an agent’s psychology – it relies only on the agent having the two preferences mentioned – and so again presents a problem for both the realist and the constructivist EU theorist. Most people have preferences like those that Allais and Rabin show cannot be captured by EU theory; and there are many other examples of preferences that EU theory cannot capture. The reason EU theory fails to capture the preferences in the Rabin and Allais examples is that it fails to separate two different sorts of reasons for risk averse preferences: local considerations about outcomes, like those that Alice advanced in order to determine that she prefers $50 (“this particular amount of money is more valuable…”) and global considerations about gambles as a whole, like those that Bob advanced in order to determine that he prefers $50 (“I would rather be guaranteed $50 than risk getting less for the possibility of getting more”). Why would an agent find this second kind of consideration relevant to decision making? Let us examine the idea that decision theory formalizes and precisifies means-ends rationality. We are presented with an agent who wants some particular end and can achieve that end through a particular means. Or, more precisely, with an agent who is faced with a choice among means that lead to different ends, which he values to different degrees. To figure out what to do, the agent must make a judgment about which ends he cares about, and how much: this is what the utility function captures.10 In typical cases, none of the means available to the agent will lead with certainty some particular end, so he must also make a 6 judgment about the likely result of each of his possible actions. This judgment is captured by the subjective probability function. Expected utility theory makes precise these two components of meansends reasoning: how much one values various ends, and which courses of action are likely to realize these ends. But this can’t be the whole story: what we’ve said so far is not enough for an agent to reason to a unique decision, and so we can’t have captured all that’s relevant to decision making. An agent might be faced with a choice between one action that guarantees that he will get something he desires somewhat and another action that might lead to something he strongly desires, but which is by no means guaranteed to do so.11 Knowing how much he values the various ends involved is not enough to determine what the agent should do in these cases: the agent must make a judgment not only about how much he cares about particular ends, and how likely his actions are to realize each of these ends, but about what strategy to take towards realizing his ends as a whole. The agent must determine how to structure the potential realization of the various aims he has. This involves deciding whether to prioritize definitely ending up with something of some value or instead to prioritize potentially ending up with something of very high value, and by how much: specifically, he must decide the extent to which he is generally willing to accept a risk of something worse in exchange for a chance of something better. This judgment corresponds to considering global or structural properties of gambles. How should an agent trade off the fact that one act will bring about some outcome for sure against the fact that another act has some small probability of bringing about some different outcome that he cares about more? This question won’t be answered by consulting the probabilities of states or the utilities of outcomes. Two agents could attach the very same values to particular outcomes (various sums of money, say), and they could have the same beliefs about how likely various acts are to result in these outcomes. And yet, one agent might hold that his preferred strategy for achieving his general goal of getting as much money as he can involves taking a gamble that has a small chance of a very high payoff, whereas the other might hold that he can more effectively achieve this same general goal by taking a gamble with a high chance of a moderate payoff. Knowing they can only achieve some of their aims, these agents have two different ways to structure the potential realization of them. This dimension of instrumental reasoning is the dimension of evaluation that standard decision theory has ignored. To be precise, it hasn’t ignored it but rather supposed that there is a single correct answer for all rational agents: one ought to take actions that have higher utility on average, regardless of the spread of possibilities. There may or may not be good arguments for this, but we are not in a position to address them before we get clear on what exactly agents are doing when they answer the question differently, and how this relates to instrumental reasoning. The aim of this paper is to make this clear. 7 3. An Alternative Theory To explain the alternative theory of instrumental rationality I endorse, I will start with the case of gambles with only two outcomes: gambles of the form {E, x1; E, x2}. As mentioned, the EU of such a gamble is p(E)u(x1) + p(E)u(x2). We can state this equivalently as u(x1) + p(E)[u(x2) – u(x1)]. Taking x2 to be weakly preferred to x1, this is equivalent to taking the minimum utility value the gamble might yield, and adding to it the potential gain above the minimum – the difference between the high value and the low value – weighted by the probability of the event in which that gain is realized. For example, the value of the $0/$100 coin-flip will be u($0) + (0.5)[u($100) – u($0)]. The value of a gamble is its instrumental value, a measure of how the agent rates it in terms of satisfying her aims: we might say, a measure of its effectiveness as a means to her various ends. To review, on EU theory, while it is up to agents themselves how valuable each outcome is and how likely they believe each event is to obtain, these two evaluations are of set significance to the instrumental value of a gamble. If two decision makers agree about the values of various outcomes and on the probabilities involved, they must evaluate gambles in exactly the same way: they must have identical preference orderings.12 They must agree about how the gambles rank in terms of satisfying their aims. However, it is plausible to think that some people are more concerned with the worst-case scenario than others, again, for purely instrumental reasons: because they think that guaranteeing themselves something of moderate value is a better way to satisfy their general aim of getting some of the things that they value than is making something of very high value merely possible. More realistically, the minimum value won’t always trump the maximum in their considerations, but it will weigh more heavily. Alternatively, an agent might be more concerned with the best-case scenario: the maximum weighs more heavily in the estimation of a gamble’s value than the minimum does, even if these two outcomes are equally likely. So, it is plausible that two agents who attach the same values as each other to $100 and $0 will not both attach the same value to a coin-flip between $0 and $100. One agent will take the fact that he has a 50% chance of winning the better prize to be a weaker consideration than it is for the other. Thus, in addition to having different attitudes towards outcomes and different evaluations of likelihoods, two agents might have different attitudes towards some way of potentially attaining some of these outcomes. A natural way to interpret these different attitudes is to postulate that different decision makers take the fact that they might improve over the minimum to be a more or less important consideration in evaluating a gamble. Formally, they weight the potential gain above the minimum differently from each other. In EU theory, the instrumental value of a gamble is at least its minimum utility value, and the potential utility gain above the minimum is weighted by the probability of attaining the higher value. But this latter feature is too restrictive: a potential gain over the minimum might increase a gamble’s 8 instrumental value over the minimum value by the size of the gain multiplied by the probability of realizing that gain, but it might instead improve it by more or by less, depending on what the agent cares about. Of course, the probability and the size of the improvement will be relevant: the higher the probability of some particular gain or the larger the size of a gain with some particular probability, the better. Therefore, I propose that the possibility of a potential utility gain over the minimum improves the gamble above its minimum utility value by the size of the gain multiplied by a function of the probability of realizing that gain, instead of by the bare probability. This function represents the agent’s attitude towards risk in the “global properties” sense. Put formally, we might calculate the risk-weighted expected utility (REU) of a gamble {E, x1; E, x2}, where u(x1) ≤ u(x2), to be u(x1) + r(p(E))[u(x2) – u(x1)], where r is the agent’s “risk function,” adhering to the constraints 0 ≤ r(p) ≤ 1 for all p; r(0) = 0; r(1) = 1; and r is non-decreasing. The equation says that the instrumental value of a two-outcome gamble will be its low value plus the interval between the low value and the high value, weighted by the output of the risk function when the input is the probability of getting the high value. The equation is equivalent to r(p(E))u(x2) + (1 – r(p(E))u(x1). So we can think of r(p) either as the weight a particular improvement-possibility gets when this possibility has probability p, or as the weight that the better outcome gets when this outcome has probability p. If the risk function has a high value for some value p, then the value of the better outcome will count for a lot in the agent’s evaluation of the gamble, and if it has a low value for some value p, then the value of the worse outcome will count for a lot. This formulation also makes it clear how an agent’s evaluation of gambles rests on factors that are irreducibly global: the amount by which each outcome gets weighted will depend on which outcome is the minimum.13 For example, for an agent who values money linearly and has a risk function of r(p) = p2, the coin-flip will be worth $25: u({HEADS, $0; TAILS, $100}) = u($0) + (0.5)2[u($100) – u($0)] = u($25). 9 r(p) = p2 1.0 0.8 0.6 r(p) 0.4 0.2 0 0 0.2 0.4 Diagram 4: Sample Risk Function: r(p) = p2 0.6 p 0.8 Diagram 2: Sample Risk Function: r(p) = p2 1.0 And here we come to the crux of the difference between how EU theory represents risk aversion and what I think instead merits the term. On EU theory, to be risk averse is to have a concave utility function. On a theory like mine, to be risk averse is to have a convex risk function.14 The intuition behind the diminishing marginal utility analysis of risk aversion was that adding money to an outcome is of less value the more money the outcome already contains. The intuition behind the present analysis of risk aversion is that adding probability to an outcome is of more value the more likely that outcome already is to obtain. Risk averters prefer to “get to certainty,” so to speak. Of course, theories like mine allow that the utility function is concave (or, indeed, any shape). But I claim that this feature, which describes how an agent evaluates outcomes, pulls apart from his attitude towards risk properly called. So I claim that what we might appropriately describe as an agent’s attitude towards risk is captured by the shape of his risk function. There is a natural way to extend this theory to gambles with more than two possible outcomes. The way I’ve set up the risk-weighted expected utility equation emphasizes that an agent considers his possible gain above the minimum (the interval between the low outcome and the high outcome), and weights that gain by a factor which is a function of the probability of obtaining it, a function that depends on how he regards risk. Now consider a situation in which a gamble might result in one of more than two possible outcomes. It seems natural that the agent should consider the possible gain between each neighboring pair of outcomes and his chance of attaining the higher outcome or better, and, again, 10 subjectively determine how much that chance of attaining this adds to the instrumental value of the gamble. One way to state the value of a gamble with more than two outcomes for a standard EU maximizer is as follows. Start with the minimum value. Next, add the interval difference between this value and the next highest value, weighted by the probability of getting at least that higher value. Then add the interval difference between this value and the next highest value, weighted by the probability of getting at least that value. And so forth. Just as we replaced subjective probabilities by subjective weights of subjective probabilities in the two-outcome case, we can do so in this case. So the value of a gamble for the REU maximizer will be determined by following this same procedure but instead weighting by a function of the probability at each juncture. For example, consider the gamble that yields $1 with probability ½, $2 with probability ¼, and $4 with probability ¼. The agent will get at least $1 for certain, and he has a ½ probability of making at least $1 more. Furthermore, he has a ¼ probability of making at least $2 beyond that. So the EU of the gamble is u($1) + ½[u($2) – u($1)] + ¼[u($4) – u($2)], and the REU of the gamble is u($1) + r(½)[u($2) – u($1)] + r(¼)[u($4) – u($2)]. So the gamble g = {E1, x1; E2, x2; … ; En, xn}, where u(x1) ≤ … ≤ u(xn), is valued under expected utility theory as ∑ = n i i i x u E p 1 ) ( ) ( , which is equivalent to: EU(g) = ∑ ∑ = − = − + + − + − + n i n n n i n i i x u x u E p x u x u E p x u x u E p x u 3 1 2 3 2 1 2 1 )) ( ) ( )( ( … )) ( ) ( ))( ( ( )) ( ) ( ))( ( ( ) ( And that same gamble will be valued under risk-weighted expected utility theory as follows: REU(g) = ∑ ∑ = − = − + + − + − + n i n n n i n i i x u x u E p r x u x u E p r x u x u E p r x u 3 1 2 3 2 1 2 1 )) ( ) ( ))( ( ( … )) ( ) ( ))( ( ( )) ( ) ( ))( ( ( ) ( We can now see how the standard Allais preferences are captured by REU theory: they maximize riskweighted expected utility only if r is convex.15 This functional form is an instance of the “rank-dependent” approach in non-expected utility theory, discovered by economists around the 1980s, in which the agent maximizes a sum of utility values of outcomes, weighted by a factor that is related to the probability of that outcome but that depends on the outcome’s rank among possible outcomes. In particular, two of these theories are formally equivalent to REU theory when we abstract away from what their “weighting factor” is a function of. The first is Choquet expected utility (CEU), due to David Schmeidler and Itzhak Gilboa,16 and the second is anticipated utility (AU), due to John Quiggin.17 However, CEU employs a weighting function of states, 11 not of probabilities of states: it does not include an agent’s judgments about probabilities at all. Indeed, it is meant to apply to decision making under uncertainty, in which agents do not always have sharp probability judgments.18 AU does attach decision weights to probabilities, but it uses an “objective” probability function: it takes the probabilities as given. What is missing from each of these theories, for the purposes of philosophers interested in instrumental rationality, is a way to separate beliefs from decision weights: each theory contains only one subjective parameter that goes into fixing the weight an outcome gets in the “expectation” equation. Therefore, neither theory can formalize what it is to have a belief separately from what it is to have a decision weight. But as I argued in my discussion of instrumental rationality, the question of which strategy to employ is entirely separate from the question of what one believes. The REU formulation allows that an agent attaches subjective probabilities to states and then employs a weighting function of these probabilities. Furthermore, the theorem presented later in the paper will allow us to derive each of these uniquely from an agent’s preferences, so we can see exactly how each contributes to the evaluation of a gamble. Nonetheless, it might be helpful to the reader to think of REU theory as a generalization of AU theory to decision making with subjective probabilities: i.e., as “subjective” anticipated utility theory. Alternatively, one could think of REU theory as a restriction of CEU to the case in which agents are probabilistically sophisticated and whose decision weights are a function of their subjective probabilities: i.e., as “CEU with probabilistic sophistication.”19 This accurately represents the formalism, if not the philosophical commitments. If we set r(p) = p, we get the standard subjective expected utility equation. And again, for agents who care more about the worst-case scenario – agents with convex r-functions – the possibility of getting more than the minimum will contribute to the value of the gamble less than it will for the expected utility maximizer. The most extreme case of this is the maximinimizer, who simply takes the gamble with the highest minimum. He can be represented using r(p) = {0 if p ≠ 1, 1 if p = 1}. And for agents who care more about the best-case scenario – agents with concave r-functions – the possibility of getting higher than the minimum will contribute to the value of the gamble more. The maximaximizer, who takes the gamble with the highest maximum, can be represented using r(p) = {0 if p = 0, 1 if p ≠ 0}. The REU equation also ensures that the value of a gamble is always at least its minimum and at most its maximum, and, since r is non-decreasing, that increasing the probability of getting a good outcome will never make a gamble worse (preferences respect weak stochastic dominance). If we require in addition that r is increasing, then increasing the probability of getting a good outcome will always make a gamble better (preferences respect strict stochastic dominance). What ingredient of instrumental rationality does the risk function represent? The utility function is traditionally supposed to represent desire, and the probability function belief – both familiar 12 propositional attitudes. We try to make beliefs “fit the world,” and we try to make the world fit our desires. But the risk function is neither of these things: it does not quantify how we see the world – it does not, for example, measure the strength of an agent’s belief that things will go well or poorly for him – and it does not describe how we would like the world to be. It is not a belief about how much risk one should tolerate, nor is it a desire for more or less risk. The risk function corresponds to neither beliefs nor desires. Instead, it measures how an agent structures the realization of his aims. We will see in the remainder of this paper exactly how it does this. On REU theory, the agent subjectively determines three things: which outcomes he prefers, how likely various acts are to result in various outcomes, and the extent to which he is generally willing to accept the risk of something worse in exchange for the possibility of something better. First, like EU theory, REU theory allows agents to attach subjective values to outcomes. It is up to agents themselves to choose their ends, and hence, REU theory includes a subjective utility function, which is not necessarily linear in money (or any other good). Second, also like EU theory, REU theory allows agents to set their credences, and hence, REU theory includes a subjective probability function. Third, unlike EU theory, REU theory allows agents to decide how to structure the potential realization of the outcomes they care about. It allows them to judge which gamble is better from the point of view of their general aim of getting more money (which includes their particular aims of, say, getting $50 or of getting $100, which they may value twice as much). It is up to them whether they will better fulfill their general aim by prioritizing a high minimum or by prioritizing a high maximum. And it is up to them exactly how these two features of gambles trade off, e.g., how much possibly doing better than the minimum is worth, or how much weight to put on what happens in the top p portion of outcomes. Hence, REU theory includes a subjective risk function, which is not necessarily linear in probability. Every agent has beliefs and desires, and determines for himself a norm for translating these two things into preferences. EU theory claims that r(p) = p is the correct norm: that we ought to be globally neutral. But just as rationality does not dictate a single utility function or credence function for all agents, I claim that it also does not dictate a unique norm. 4. From Preferences to Beliefs and Desires I have so far been focusing on the question of how an agent might aggregate her beliefs (credence function) and desires (utility function) to arrive at a single value for an act. I’ve claimed that agents need not aggregate according to expected utility, but instead might weight utility intervals by a function of their probabilities. Thus we can model decision makers as using a more general decision rule, which includes a utility function, a credence function, and a risk function. However, the question that has received the most attention in philosophy is not how an agent might arrive at preferences given her beliefs and desires, 13 but rather how we might determine an agent’s beliefs and desires from her preferences. Specifically, decision theorists are interested in what restrictions on preferences will allow us to fix unique beliefs and desires. So the question that arises for REU theory is what restrictions are needed in order to extract beliefs, desires, and attitudes towards risk. As mentioned, a utility and probability function represent an agent under EU theory just in case for all acts f and g, the agent weakly prefers f to g iff EU(g) ≤ EU(f), where expected utility is calculated relative to the agent’s subjective probability function of states. A representation theorem for EU theory spells out a set of axioms such that if an agent’s preferences obey these axioms, then she will be representable under EU theory by a unique probability function and a utility function that is unique up to positive affine transformation: she will be an EU maximizer relative to these functions.20 Representation theorems are important in decision theory, but their upshot depends on the use to which decision theory is put. There are at least two very different ways in which decision theory has been used, which I refer to as the prescriptive use and the interpretive use.21 When the theory is taken prescriptively, an agent uses it to identify the choice he should make or the preferences he should have; or the decision theorist uses the theory to assess whether the agent’s choices and preferences are rational. The agent himself can use decision theory prescriptively in at least two ways. First, if he has already formed preferences over enough items, he can look to decision theory to tell him the preferences he should have over other items: to tell him how to attain his ends of getting, on balance, things that he (already) more strongly prefers. (Under realist decision theory, knowing his utility and probability values will also be enough for decision theory to tell him the preferences he should have.) Second, if an agent realizes that his preferences are not in accord with decision theory, then he can conclude that he has done something wrong and, insofar as he is rational, that he should alter his preferences so that they do so accord. In addition, the theorist using decision theory prescriptively ascertains whether the agent’s choices in fact accord with the theory, and it is by this criterion that she judges whether the agent’s preferences are rational. Representation theorems state the conditions under which an agent can count as an EU maximizer, and thus the conditions under which an agent’s preferences are rational (according to the standard theory). Therefore, they are useful for prescriptive decision theory because they provide a criterion for determining when an agent has irrational preferences that doesn’t require knowing his precise numerical values. Furthermore, this criterion can be employed if we think there are no precise numerical values to know aside from those that result from the theorem, so they are especially useful to the constructivist. For the constructivist, rationality just is conformity to the axioms of decision theory, and it is a convenience that this also guarantees representability as an expected utility maximizer. Thus, representation theorems are useful because they allow us to refocus the debate about rationality: instead 14 of arguing that a rational agent ought to maximize expected utility because, say, he ought to care only about average utility value, the EU theorist can argue that a rational agent ought to conform to the axioms. In contrast to prescriptive decision theory, a portion of the modern philosophical literature treats decision theory interpretively: not as a useful guide to an agent’s own decisions, but rather as a framework to interpret an agent’s desires, his beliefs, and perhaps even the options that he takes himself to be deciding among. The interpretive use of decision theory arises in response to a worry about how to discover what an agent believes and desires, given that we have no direct access to these mental states – and, if constructivism (or a version of realism on which one’s desires are opaque) is true, neither do the agents themselves, since these states cannot be discovered by introspection. However, it seems that it is relatively easy to discover agents’ preferences: preferences do manifest themselves directly, if not perfectly, in behavior, and are ordinarily open to introspection. It should be clear how representation theorems are useful to interpretive theorists. If an agent’s preferences obey the axioms of EU theory, then the interpretive theorist can start with the agent’s (observable) preferences and derive the probability function and utility function that represent her. It should also be clear why it is important that the theorems result in unique probability and utility functions. If there were multiple pairs that each could represent the agent as an expected utility maximizer, we wouldn’t know which way of representing the agent accurately captures “her” beliefs and desires.22 We can see that representation theorems are crucial to decision theory, so any alternative to EU theory needs a representation theorem if it can hope to serve the purposes EU theory is traditionally put to. Furthermore, comparing the axioms of an EU representation theorem to those of an alternative will allow us to see the difference between what each theory requires of rational agents. In the remainder of this paper, I explain the representation theorem for REU theory. This theorem presents a set of axioms such that if a decision maker’s preferences obey these axioms, we can determine a unique probability function, a unique risk function, and a utility function that is unique up to positive affine transformation such that the agent maximizes REU relative to these three functions. One thing this will illuminate is what attitude the risk function corresponds to at the level of preferences, and exactly how it differs from a credence function. My main aim here is to show how REU theory captures what it is to be risk-averse, by contrasting the axioms of the REU representation theorem and the stronger axioms of the analogous representation theorem for EU theory. This will provide a way to frame the debate about whether REU maximizers are rational around the question of whether agents ought to obey the axioms of EU theory or only the weaker axioms of REU theory. 15 5. The Axiomatic Dispute What is crucial to the present paper is presenting the axiomatic difference between REU theory and EU theory in a way that explains how the difference between the axioms of REU theory and the stronger axioms of EU theory amounts to the difference between allowing agents to care about global properties – or to determine for themselves the third component of instrumental rationality – and prohibiting them from doing so. Thus, I will leave the technical presentation of the axioms and theorem, as well as a more formal discussion of them, to the appendix. I will here concentrate on explaining the crucial differences between the theories.23 First, though, let me briefly say something about the relationship of the theorem here to other related theorems. The theorem here draws on two other results, one by Veronika Köbberling and Peter Wakker and the other by Mark Machina and David Schmeidler.24 Köbberling and Wakker prove a representation theorem for another rank-dependent theory, CEU theory. CEU, like REU theory, applies to preferences over acts, in which outcomes are tied to events whose probabilities are not given. However, as mentioned, CEU does not represent the agent as having a function that assigns additive probabilities to events, and thus the representation theorem for CEU does not provide us with a way of extracting a rational agent’s degrees of belief from his preferences. Machina and Schmeidler give conditions under which an agent can be represented as probabilistically sophisticated – as having a unique probability function relative to which his preferences respect stochastic dominance – and as maximizing some value function, but their result does not allow us to determine the values of outcomes aside from the gambles they are embedded in. Combining their results, I give conditions under which we can represent an agent as a probabilistically sophisticated decision maker maximizing the specific function that this paper is concerned with: conditions under which we can extract from an agent’s preferences a probability function, a utility function, and a function that represents how he structures the realization of his aims in the face of risk. Thus the set of axioms in the REU representation theorem are a combination of Köbberling and Wakker’s and Machina and Schmeidler’s axioms, strictly stronger than either set of axioms. The crux of the disagreement between REU theory and EU theory concerns the status of an axiom known as Unrestricted Tradeoff Consistency. According to the EU theorist, rational preferences must satisfy this condition. But according to the REU theorist, rational preferences must only satisfy two jointly weaker conditions, Comonotonic Tradeoff Consistency (from Köbberling’s and Wakker’s axiomatization) and Strong Comparative Probability (from Machina’s and Schmeider’s axiomatization).25 We can see the difference between these commitments by observing the pattern of preferences that Unrestricted Tradeoff Consistency (UTC) rules out but that REU theory allows. The basic idea behind UTC is that which tradeoffs you are willing to accept – tradeoffs about what happens in various 16 states – reveals your values. For example, if it is worth it to you to make one state worse by $10 in order to make another state better by $20, that shows that the utility difference $10 makes to the first state is equivalent to the utility difference $20 makes to the second state, as long as the two states are equally likely. But things are slightly more complicated in the formal theory: since we are trying to extract both utilities and probabilities from preferences, we cannot use the terms “equally likely” or “utility difference” in our axioms themselves. So EU theorists have to capture the spirit of this idea just using preferences. To capture this idea, the EU theorist defines a precise notion of equal tradeoffs. Let us say that x rather than y in E is an appropriate tradeoff for f rather than g in E if an agent is indifferent between the gamble that yields x if event E obtains and f if event E obtains and the gamble that yields y if event E obtains and g if event E obtains.26 For example, if the agent is indifferent between these two gambles:
Leave a Reply