The
Prisoner's Dilemma
There are thousands of Web-Sites about this,
and a huge amount of relevant published books and articles, as well as
a substantial and important literature on Game Theory in general.
I just want you now to think for yourselves whether we can learn
from the dilemma, and if so, what we can learn. You can start from
a short article of mine (F.C.T. Moore, `The Martyr's Dilemma', Analysis,
45:1, January 1985, pp. 29-33), reproduced below. [If you want
to look at a somewhat wider discussion, you can borrow a copy of my article
'Taking the sting out of the Prisoner's Dilemma' (Philosophical Quarterly,
44: 175, April 1994, pp. 223-234) from the General Office of the Department
of Philosophy.] The article
in the Stanford Encyclopedia of Philosophy is not too bad, but I find
it rather heavy (if you use this link to it, you'll have
to use the BACK button to get back here).
It is sometimes stated or implied
that the Prisoner's Dilemma arises only for rationally self-interested
behaviour. Hillel Steiner, for example, has said (`Prisoner's Dilemma
as an insoluble problem, Mind, XCI (1982), pp. 285-6) that the dilemma
depends upon the statement that `all persons pursue rationally self-interested
strategies'. Commentators on Steiner's article appear to have
accepted this widespread view (see Gordon, D., `Is the Prisoner's Dilemma
an insoluble problem ?', and Porter, J.P., `Relevant Interest and the Prisoner's
Dilemma, Mind, XCIII (1984), pp. 98-102). Thus the question
of the rationality of altruism may be held to be settled, or begged, by
the dilemma. Indeed, in the usual matrix (for two players A and B,
and two choices C and D), where T>R>P>S (the latter, of course, being a
preference ordering), C is commonly described as the `altruistic' (but
irrational) choice.
|
|
|
B
|
|
|
|
C
|
|
D
|
|
C
|
R,R
|
|
S,T
|
A
|
|
|
|
|
|
D
|
T,S
|
|
P,P
|
In fact, moralizing readings of
the Prisoner's Dilemma can be quite elaborate: thus C is read "co-operate",
and R is the "reward" obtained by A and B for their co-operation; D is
read as "deviation" or "defection", and P is the "punishment" both receive
for their wickedness; in the case of asymmetrical choices, the player gaining
T is described as a "traitor", while the loser (who takes S) is a "saint"
or even a "sucker". David Lewis has claimed it to be "rat"ional to
"rat" (in `Prisoner's Dilemma is a Newcomb problem', Philosophy and
Public Affairs, 83 (1979), pp. 235-40), while Anatol Rapaport in an
early article (`Formal games as probing tools for investigating behaviour
motivated by trust and suspicion', Journal of Conflict Resolution,
VII (1963), pp. 570-9) described a player's repeated co-operation in iterated
games (in spite of defection by the other player), as a "martyr run".
Various authors have questioned
such readings of the Prisoner's Dilemma, including recently M.C. Geach
(Analysis, 44:1, January 1984, pp. 46-8: "a similar sort of case
can arise even when people agree about the desire effect, and are not selfish.).
Similarly, J.W.N. Watkins has maintained, in `Self Interest and Morality'
(Körner, S. (ed.), Practical Reason, Oxford: Blackwell, 1974,
pp. 67-77), that the dilemma arises for moral agents who give equal weight
to their own and the other's preferences. And Derek Parfit
in `Is commonses morality self-defeating ?' (Journal of Philosophy
76:10, 1979, pp. 533-45), applies Prisoner's Dilemma reasoning to cases
where the parties are pursuing different goals which each flow from the
same moral theory, in order to construct an argument that any such theory
is self-defeating. My claim is more general, that Prisoner's
Dilemma arises for any actions whatsoever in whose case the outocme depends
upon the interplay of different actions: it is thus a problem for the theory
of social action, and does not in any way suppose or show the irrationality
of altruism.
Note also recent work
on applications of Game Theory to the evolution of behaviour in Maynard
Smith, J., Evolution and the Theory of Games (Cambridge: Cambridge
University Press, 1982), and `Game Theory and the Evolution of Behavior'
(in Behavioral and Brain Sciences 7 (1984), pp. 95-125, including
pp. 114-115 the commentary by Rapoport `Game Theory without Rationality').
These applications assume the possibility of different strategies, or "behavioural
phenotypes", being present in a population (whether because some individuals
are genetically determined to do x, and others to do y, or because an individual
is genetically determined to do x and y with certain probabilities, or
because an individual tends to do x and y with certain probabilities as
a result of learning). It is then possible to calculate whether there
is an "evolutionarily stable strategy" (pure or mixed) by interpreting
payoff in terms of Darwinian fitness, i.e. in terms of the numbers of offspring
produced by organisms with various strategies. Efforts have
been made using this approach to offer an empirical explanation of the
phenomenon of altruism. But it is more relevant to the present argument
that the approach itself theoretically allows any behavioural tendencies
to occur. We might, for example, have entire populations of doves,
or martyrs. The empirical questions are: which behavioural
phenotypes actually occur, and which have survival value.
It is not my aim here to say
anything about solutions to the Prisoner's Dilemma, but rather to establish
a point about its character. Indeed, one interesting line of solution
was developed formally by Howard, N., in `The Theory of Meta-Games' and
`The Mathematics of Meta-Games' (in General Systems 11, 1966, pp.
167-200), and explained and elaborated by Pörn, Ingmar, Action
Theory and Social Science: some formal models, Dordrecht: Reidel, 1977,
pp.98-102. This line of solution involves the introduction
of policies in the first-level meta-game (by which a makes his choice
of C or D conditional on a's choice of C or D), and then in the
second-level meta-game a can make his choice of C or D conditional
upon b's choice of policy. If this line of reasoning
is accepted, it turns out that a co-operative policy is rationally preferable
to a non-co-operative policy (see also Rapoport, A., `Escape from Paradox',
Scientific American 217 (1967), pp.50-6).
My aim here is rather to support
the view that the Prisoner's Dilemma is morally neutral by arguing that
it arises out of interdependent choices under uncertainty, whatever the
moral character of the objectives of the agents, and, in particular, independently
of whether those objectives are self-interested or altruistic in character.
The rational choice may commonly have been as a "selfish" or self-interested
one, but this is only because self-interested objectives happend to have
been selected in illustrating the dilemma.
It is sufficient proof of
this contention to construct another instantiation of the same problem
in which the objective is admittedly altruistic.
Three people, A, B and C,
are in a dug-out. A had-grenade lands close to and between A and
B, but some distance from C. This is how the situation appears to
A and B: if A and B both throw themselves away from the grenade, all three
will be injured. If A throws himself on the grenade, he will die,
and B will be injured, but C will be unscathed. Correspondingly,
if B throws himself on the grenade. If both A and B throw themselves
on the grenade, both will be seriously injured, but C will be unscathed.
If A is a martyr, all he will
be concerned with is the fate of the others: that is, in the case of the
four possible outcomes as they affect the three men in the dug-out,
we measure the pay-off from A's point of view by ignoring the outcome
for himself. Now let us suppose that in fact both A and
B are martyrs. Let them accordingly share the following preference
ordering over the possible outcomes (for both of them, as explained, the
outcome for only two people out of the three in the dug-out is relevant):
1 |
least bad |
1 injured |
2 |
bad |
2 injured |
3 |
very bad |
1 seriously injured |
4 |
worst |
1 dead |
What then should A do ?
Consider this diagram:
|
|
|
B
|
|
|
|
saves
himself
|
|
sacrifices
himself
|
|
saves
himself
|
2i/2i
|
|
1d/1i
|
A
|
|
|
|
|
|
sacrifices
himself
|
1i/1d
|
|
1si/1si
|
outcomes as judged by A and B, each
ignoring himself
(i=injured; d=dead; u=unscathed; si=seriously
injured)
Given A's preferences specified
above (i.e. 1 least bad, 2 bad, 3 very bad, 4 worst), he faces the following
situation. If he sacrifices himself, there are two possible
outcomes - 3 very bad (if B does the same), or 1 least bad (if B tries
to save his skin); but if he saves his skin, the two possible outcomes
are - 4 worst (in case B sacrifices himself) or 2 bad (in case B also saves
his skin). By minimax, as a martyr,
A should sacrifice himself, since the outcome cannot be worse than 3 very
bad, and there is chance of 1, the least bad outcome. He should not
save his skin, since though there is a chance of 2 the (merely) bad outcome
- which is collectively the best for two martyrs, there is also the chance
of 4, the worst outcome, if the other sacrifices himself. Let me
emphasize again that if A is a martyr (that is, if his preference ordering
of the outcomes is the one already described), then it would be rational
for A, using minimax, to sacrifice himself; but if both A and B are martyrs
the best result for them collecitvely, i.e. the outcome which would produce
the best net result taking both their preferences into account, would be
achieved by both saving their skin, for there will then be two injured,
which both avgree to be better than one seriously injured, or one dead.
Thus if both A and B are martyrs,
they will in 'rationality' both sacrifice themselves, producing an outcome
3, which is very bad (i.e. one seriously injured, from either point of
view); but irrational martyrs, both of whom saved their skin, would actually
produce the best net result. This can be checked on the following
diagram:
|
|
|
B
|
|
|
|
saves
himself
|
|
sacrifices
himself
|
|
saves
himself
|
A B C
i i
i
|
|
A B C
i d
u
|
A
|
|
|
|
|
|
sacrifices
himself
|
A B C
d i
u
|
|
A B C
si si u
|
I conclude that we can readily
construct a dilemma formally identical to the Prisoner's Dilemma, but in
which the parties have altruistic objectives, in the sense of ignoring
their own interests. My dilemma could be called the "Martyr's Dilemma"
- except that if my argument here is cogent, the two dilemmas are in fact
just the same dilemma in different instantiations. I am uncertain
whether the example tells us anything about martyrdom; but it does, I believe,
prove that moralizing interpretations of the Prisoner's Dilemma are wrong.
University of Hong Kong
© F. C. T. MOORE
1985
Pokfulam Road, Hong Kong
GO
BACK