Suppose there are only two possible worlds, the small world w that contains a small number of people, and the large world W that contains a much larger number of people. You current estimate there is a .5 chance that w is the real world, and a .5 chance that W is the real world. You don't have an intuitive way of "comparing utilities" between these worlds, so as a muddled compromise you currently spend half your resources on activities AW that will maximize utility function UW if W is real, and the other half on Aw to maximize utility Uw if w is real.
This is not a pure utilitarian strategy, so we know that our activities may be suboptimal; that we can probably do better in terms of maximizing our total expected utility. Here is one utilitarian strategy, among many, that usually dominates our current muddled compromise.
Calculate expected utility EU' of a local state s as:
EU'(s) = P(W) U'(s|W) * P(w) U'(s|w)
With P being the probablity, and U' being any new utility function that obeys these constraints:
1. U'(s|W), the utility of s if W is real, is an affine transformation of UW.
2. U'(s|w), the utility of s if w is real, is an affine transformation of Uw.
3. If the only activities you could do were AW and Aw, and P(W) and P(w) are both .5, then EU' is maximized when you spend half of your resources on AW and half on Aw.
What have we gained?
1. We now have a way to evaluate the utility of a single action that will help in both W and w.
2. We now have a way to evaluate the utility of a single action that helps in W but harms in w, or vice versa.
3. As future evidence comes in about whether W or w is real, you have a way to optimally shift resources between AW activities and Aw activities.
Subscribe to:
Post Comments (Atom)
2 comments:
This is generally a poor solution, (and one that very closely resembles the sorts of solutions Kahneman&Tversky argued against) as it gives radically different proposed behaviors depending on how you frame the bundling of actions.
It's also going to give radically different proposed behaviors based on what point you decide to adopt the heuristic, and so is still in some sense suboptimal. However, my argument is that the proposed solution still tends to dominate the original algorithm. Is there a better general solution that you would propose, in the case where the rationale for the original algorithm is inaccessible or nonexistent?
Post a Comment