Game Theory

Game tree from A. Rubinstein, 1991, <i>Econometrica</i>

Click here for an accompanying essay on the Taxonomy of Game Theory.

Game Theory has emerged recently as a powerful challenger to the conventional method of examining economics. Although many illustrious predecessors worked on problems in what can be called "game theory", the fundamental, formal conception of game theory as part and parcel of economic theory were first organized in John von Neumann and Oskar Morgenstern's 1944 classic, Theory of Games and Economic Behavior (1944) which used the "maximin" solution concept derived earlier by John von Neumann (1928) to solve simple strategic normal games.

Game Theory can be roughly divided into two broad areas: non- cooperative (or strategic) games and co-operative (or coalitional) games. The meaning of these terms are self evident, although John Nash claimed that one should be able to reduce all co-operative games into some non-cooperative form. This position is what is known as the "Nash Programme". Within the non-cooperative literature, a distinction can also be made between "normal" form games (static) and "extensive" form games (dynamic).

The main issue about game theory is that instead of agents making decisions as reactions to exogenous "prices" or some other "dead variable", their decisions are strategic reactions to other agents actions ("live variables"). He is faced with a set of moves he can play and will form a strategy, a best response to his environment, which he will play by. Strategies can be either "pure" (play a particular move) or "mixed" (random play). An " Nash Equilibrium" will be reached when each agent's actions begets a reaction by all the other agents which, in turn, begets the same initial action. In other words, all best responses of all players are in accordance with each other.

von Neumann and Morgenstern (1944) introduced the strategic normal game, strategic extensive game, the concept of pure/mixed strategies, coalitional games as well as the axiomatization of expected utility theory - so useful for later Neo-Walrasian theory.

In 1950, John Nash introduced the concept of a "Nash Equilibrium" (NE), which became the organizing concept under Game Theory - even though the concept actually stretched as far back as Cournot (1838). Nash followed this up in 1951 with the concept of a "Nash Bargaining Solution" (NBS) for coalitional games.

Then the floodgates opened for the refinement of Nash Equilibrium. In the field of non-cooperative games, R. Duncan Luce and Howard Raiffa (1957) provided the first popular textbook on Game Theory and, in it, they formalized the idea of the Iterated Elimination of Dominated Strategies (IEDS) method for Strategic Normal Games and introduced the concept of "Repeated Game" (static games which are played several times over). H.W. Kuhn (1953) introduced extensive games with "imperfect information" (i.e. where one does not know what moves have already been played by other players). William Vickrey (1961) provided the first formalization of "auctions". Reinhard Selten (1965) developed the concept of a "Subgame Perfect Equilibrium" (SPE) (i.e. elimination by backward induction) as a refined solution for extensive form games. John C. Harsanyi (1967-8) developed the concept of a "Bayesian Nash Equilibrium" (BNE) for Bayesian games (i.e. games with incomplete information - where there is some uncertainty surrounding moves, or "nature plays").

In coalitional (co-operative) games further refinements also occurred. Lloyd Shapley (1953) introduced the concept of the "Shapley Value" and the "Core" (which had been originally conceived by F.Y. Edgeworth (1881)) as solutions to Coalitional Games. Throughout the early 1960s, Robert J. Aumann and Martin Shubik began to apply Game Theory extensively throughout economics (e.g. industrial organization, general equilibrium, monetary theory, etc.) as well as invent several solution concepts for Coalitional Games (e.g. Bargaining Set, Strong Equilibrium), develop theorems for "large games" with infinite players, as well as provide some early statements of the "Folk Theorems" (solution concepts for Repeated Games). David Schmeidler (1969) developed the "Nucleolus" solution to Coalitional Games.

Further developments emerged in the 1970s. John C. Harsanyi (1973) provided a remarkably insightful new interpretation of the concept of a "mixed strategy". Robert J. Aumann (1974) defined "Correlated Equilibrium" for Bayesian Games while Reinhard Selten (1975) introduced "Trembling Hand Equilibrium" for Bayesian Games. Further definitions came: Robert J. Aumann (1976) formally defined the concept of "Common Knowledge", opening a floodgate of literature on the concept of "knowledge" whereas B.D. Bernheim and D.G. Pearce (1984) formalized the concept of "rationalizability".

Advancements continued apace: David M. Kreps and Robert Wilson (1982) introduced the concept of "Sequential Equilibrium" (SEQE) for extensive games with imperfect information while Ariel Rubinstein (1982), following the "Nash Programme" and an early insight by Frederik Zeuthen (1930), transformed the co-operative Nash Bargaining Solution (NBS) into a non-cooperative Strategic Extensive Game of sequential bargaining. Elon Kohlberg and Jean-François Mertens (1986) developed the concept of "Forward Induction" for extensive games. Drew Fudenberg and E.S. Maskin (1986) developed one of the more complicated "Perfect Folk Theorems" for infinitely repeated games. Finally, J.C. Harsanyi and R.Selten (1988) developed the idea of "equilibrium selection" for any type of game while D. Fudenberg and Jean Tirole (1991) developed the "Bayesian Perfect Equilibrium" (BPE) for Extensive Bayesian Games .

Evolutionary Game Theory started its development slightly later - although it can be thought of as part of the Nash Programme. Its objective is to apply the concepts of non-cooperative game theory to explain such phenomena which are often thought to be the result of cooperation or human design - i.e. "institutions" and "conventions" such as market formation, price mechanisms, social rules of conduct, money and credit, etc. One of the earliest exponents of the theory of evolutionary games was Thomas C. Schelling (1960, 1981) who argued that apparently "cooperative" social institutions (in this case, settlements to conflicts) are maintained by essentially by "threats" of punishment and retaliation.

Several Nobel Prizes have been awarded to some of major figures of Game Theory: the Nobel was shared by John Nash, J.C. Harsanyi and R. Selten in 1994 and William Vickrey in 1996. Herbert Simon won the Nobel in 1979 for concepts (e.g. bounded rationality) which have since been incorporated into the corpus of (Evolutionary) Game Theory.

Predecessors

Pioneers

The New Generation in Game Theory

Evolutionary Games Resources on Game Theory

HET Home Page
Alphabetical Index
Schools of Thought
GF Surveys and Essays
HET Web Links