Chapter 3

Zero-Sum Games

Zero-sum games abstract the model of combinatorial games. Fundamental examples arise naturally as LAGRANGE games from mathematical optimization problems and thus furnish an important link between game theory and mathematical optimization theory. In particular, strategic equilibria in such games correspond to optimal solutions of optimization problems. Conversely, mathematical optimization techniques are important tools for the analysis of game-theoretic situations.

As in the previous chapter, we consider games Γ with 2 agents (or players). However, we shift the viewpoint and no longer assume players taking turns in moving a system from one state into another. Instead, we assume that the players decide on strategies according to which they play the game.

More precisely, the players are assumed to have sets X and Y of possible strategies (or decisions, or actions, etc.) at their disposal. Hence a “state” of the game under this new perspective is a pair (x, y) of strategies xX and yY.1 In fact, we consider the set

of joint strategy choices as the system underlying the game Γ, where σ0 is just an initial state, which we may attach formally to the set X × Y of joint strategies. We therefore refer to the one player as the x-player and to the other as the y-player.

Γ is a zero-sum game if there is a function

that encodes the utility of the strategic choice (x, y) in the sense that the utility values of the individual players add up to zero:

(1) u1(x, y) = U(x, y) is the gain of the x-player;

(2) u2(x, y) = –U(x, y) is the gain of the y-player.

It follows that the two players have opposing goals:

(X) The x-player wants to choose xX as to maximize U(x, y).

(Y) The y-player wants to choose yY as to minimize U(x, y).

We denote the corresponding zero-sum game by Γ = Γ(X, Y, U).

EX. 3.1. A combinatorial game with respective strategy sets X and Y for the two players is, in principle, a zero-sum game with the utility function

EX. 3.2. The Prisoner’s dilemma is a 2-person game. However, the utilities of Ex. 1.10 do not lead to a zero-sum game.

1.Matrix games

In the case of finite strategy sets, say X = {1, . . . , m} and Y = {1, . . . , n}, a function U : X × Y → ℝ can be presented in matrix form:

The associated matrix game is the zero-sum game Γ = (X, Y, U) where the x-player chooses a row i and the y-player a column j. This joint selection (i, j) has the utility value uij for the row player and the value (–uij) for the column player. As an example, consider the game with the utility matrix

In the example (14), there is no obvious overall “optimal” choice of strategies. No matter what row i and column j are selected by the players, one of the players will find2 that the other choice would have been more profitable. In this sense, this matrix game has no “solution”.

Before pursuing this point further, we will introduce a general concept for the solution of a zero-sum game in terms of an equilibrium between both players.

2.Equilibria

Let us assume that both players in the zero-sum game Γ = (X, Y, U) are risk avoiding and want to ensure themselves optimally against the worst case. So they consider the worst case functions

The x-player thus faces the primal problem

while the y-player is to solve the dual problem

From the definition, one immediately deduces for any xX and yY the primal-dual inequality:

We say that (x*, y*) ∈ X × Y is an equilibrium of the game Γ if it yields the equality

i.e., if the primal-dual inequality, in fact, attains equality:

In the equilibrium (x*, y*), none of the risk avoiding players has an incentive3 to deviate from the chosen strategy. In this sense, equilibria represent optimal strategies for risk avoiding players.

EX. 3.3. Determine the best worst-case strategies for the two players in the matrix game with the matrix U of (14) and show that the game has no equilibrium.

Give furthermore an example of a matrix game that possesses at least one equilibrium.

Finding equilibria. If the strategy sets X and Y are finite and hence the zero-sum game Γ = (X, Y, U) is a matrix game, the question whether an equilibrium exists, can — in principle — be answered in finite time by a simple procedure:

Check each strategy pair (x*, y*) ∈ X × Y for the property (19).

If X and Y are infinite, usually the existence of equilibria can only be decided if the function U : X × Y → ℝ has special properties. From a theoretical point of view, the notion of convexity is very helpful and important.

3.Convex zero-sum games

Recall4 that a convex combination of points x1, . . . , xk ∈ ℝn is a linear combination

An important interpretation of is based on the observation that the coefficient vector λ = (λ1 , . . . , λk) of a convex combination is a probability distribution:

If a point xi is selected from the set {x1, . . . , xk} with probability λi, then the components of the convex combination are exactly the expected component values of the stochastically selected point.

Another way of looking at is:

If weights of size λi are placed on the points xi, then is their center of gravity.

A set X ⊆ ℝn is convex if X contains all convex combinations of all possible finite subsets {x1, . . . , xk} ⊆ X.

EX. 3.4. Let S = {s1, . . . , sm} be an arbitrary set with m ≥ 1 elements. Show that the set of all probability distributions λ on S forms a compact convex subset of ℝm.

A function f : X → ℝ is convex (or convex up) if X is a convex subset of some coordinate space ℝn and for every x1, . . . , xkX and probability distribution λ = (λ1 , . . . , λk), one has

f is concave (or convex down) if g = –f is convex (up).

With this terminology, we say that the zero-sum game Γ = (X, Y, U) is convex if

(1)  X and Y are non-empty convex strategy sets;
(2) the utility U : X × Y → ℝ is such that
(a)  for every yY, the map xU(x, y) is concave.
(b)  for every xX, the map yU(x, y) is convex.

The main theorem on general convex zero-sum-games guarantees the existence of at least one equilibrium in the case of compact strategy sets:

THEOREM 3.1. A convex zero-sum game Γ = (X, Y, U) with compact strategy sets X and Y and a continuous utility U admits a strategic equilibrium (x*, y*) ∈ X × Y.

Proof. Since X and Y are convex and compact sets, so is the set Z = X × Y and hence also the set Z × Z.

Consider the continuous function G : Z × Z → ℝ with the values

Since U is concave in the first variable x and (–U) concave in the second variable y, we find that G is concave in the second variable (x, y). Hence Corollary A.1 (Appendix) allows us to deduce the existence of an element (x*, y*) ∈ Z that satisfies

for all (x, y) ∈ Z and hence the inequality

This shows that x* is the best strategy for the x-player if the y-player chooses y*Y. Similarly, y* can be proved to be optimal against x*. In other words, (x*, y*) is an equilibrium of (X, Y, U).

Theorem 3.1 has important consequences not only in game theory but also in the theory of mathematical optimization in general, which we will sketch in more details in Section 4 below. To illustrate the situation, let us first look at the special case of randomizing the strategic decisions in zero-sum matrix games.

3.1.Randomized matrix games

Recall from Ex. 5.2 that a zero-sum game Γ = (X, Y, U) with finite strategy sets X and Y does not necessarily admit an equilibrium.

Suppose the players randomize the choice of their respective strategies. That is to say, the x-player decides on a probability distribution on X and chooses an iX with probability i. Similarly, the y-player chooses a probability distribution on Y and selects jY with probability j. Then the x-player’s expected gain is

where the uij are the coefficients of the utility matrix U.

So we arrive at a zero-sum game where is the set of probability distributions on X and the set of probability distributions on Y. and are compact convex sets (cf. Ex. 3.4).

Moreover, the function is linear, and thus continuous and both concave and convex in both components.

It follows that is a convex game that satisfies the hypothesis of Theorem 3.1. Therefore, admits an equilibrium. This proves the VON NEUMANN’s Theorem:

THEOREM 3.2 (VON NEUMANN [32]). Let U ∈ ℝm×n be an arbitrary matrix with coefficients uij. Then there exist x*X and y*Y such that

where X is the set of all probability distributions on {1, . . . , m} and Y the set of all probability distributions on {1, . . . , n}.

3.2.Computational aspects

While it is generally not easy to compute equilibria in zero-sum games, the task becomes tractable for randomized matrix games. Consider, for example, the two sets X = {1, . . . , m} and Y = { 1 , . . . , n} and the utility matrix

For the probability distributions x and y, the expected utility for the x-player is

The worst case for the row player occurs when the column player selects a probability distribution that puts the full weight 1 on kY such that

Hence

Similarly, the worst case for the column player it attained when the row player places the full probability weight 1 onto X such that

This yields

This analysis shows:

PROPOSITION 3.1. If (z*, x*) is an optimal solution of (20) and (w*, y*) and optimal solution of (21), then

REMARK 3.1. As further outlined in Section 4.5 below, the optimization problems (20) and (21) are linear programs that are dual to each other. They can be solved very efficiently in practice. For explicit solution algorithms, the interested reader may consult the standard literature on mathematical optimization.5

4.LAGRANGE games

The analysis of zero-sum games is closely connected with a fundamental technique in mathematical optimization. A very general formulation of an optimization problem is

where could be any set and f : → ℝ an arbitrary objective function. In our context, however, we will look at more concretely specified problems and understand by a mathematical optimization problem a problem of the form

where X is a nonempty subset of some coordinate space ℝn with an objective function f : X → ℝ.

The vector valued function g : X → ℝm is a restriction function and combines m real-valued restriction functions gi : X → ℝ as its components. The set of feasible solutions of (22) is

REMARK 3.2. The model (22) formulates an optimization problem as a maximization problem. Minimization problems can, of course, also be formulated within this model because of

The optimization problem (22) defines a zero-sum game Λ = with the so-called LAGRANGE function

as its utility. We refer to Λ as a LAGRANGE game.6 The worst-case utility functions of the two players in Λ are:

EX. 3.5 (Convex LAGRANGE games). If X is convex and the objective function f : X → ℝ as well as the restriction functions gi : X → ℝ are concave, then the LAGRANGE game Λ = is a convex zero-sum game. Indeed, L(x, y) is concave in x for every fixed y ≥ 0 and linear in y for every fixed xX. Since linear functions are in particular convex, the game Λ is convex.

4.1.Complementary slackness

The choice of an element xX with at least one restriction violation gi(x) < 0 would allow the y-player in the LAGRANGE game Λ = to increase its utility value infinitely with the choice yi ≈ ∞. So the risk avoiding x-player will always try to select a feasible x.

On the other hand, if all of the feasibility requirements gi(x) ≥ 0 are satisfied, the best the y-player can do is the selection of y such that the so-called complementary slackness condition

is met. Consequently, one finds:

The primal LAGRANGE problem is identical with the original problem:

LEMMA 3.1. If (x*, y*) is an equilibrium of the LAGRANGE game Λ, then x* is an optimal solution of problem (22).

Proof. For every feasible x, y* ≥ 0 yields (y*)Tg(x) ≥ 0 and, therefore,

So x* is optimal.

4.2.The KKT-conditions

Lemma 3.1 indicates the importance of being able to identify equilibria in LAGRANGE games. In order to establish necessary conditions, i.e., conditions which candidates for equilibria must satisfy, we impose further assumptions on problem (22):

(1)X ⊆ ℝn is a convex set, i.e., X contains with every x, x′ also the whole line segment

(2)The functions f and gi in (22) have continuous partial derivatives ∂f(x)/∂xj and ∂gi(x)/∂xj for all j = 1, . . . , n.

It follows that also the partial derivatives of the LAGRANGE function L exist. So the marginal change of L into the direction d of the x-variables is

REMARK 3.3 (JACOBI matrix). The (m × n) matrix Dg(x) having as coefficients the partial derivatives

of a function g : ℝn → ℝm is known as a functional or JACOBI7 matrix. It provides a compact matrix notation for the marginal change of the Lagrange function:

LEMMA 3.2 (KKT-conditions). The pair (x, y) ∈ X × cannot be an equilibrium of the LAGRANGE game Λ unless:

(K0) g(x) ≥ 0;

(K1) yTg(x) = 0;

(K3) ∇xL(x, y)d ≤ 0 holds for all d such that x + dX.

Proof. We already know that the feasibility condition (K0) and the complementary slackness condition (K1) are necessarily satisfied by an equilibrium. If (K3) were violated and ∇xL(x, y)d > 0 were true, the x-player could improve the L-value by moving a bit into direction d. This would contradict the definition of an “equilibrium”.

REMARK 3.4. The three conditions of Lemma 3.2 are the so-called KKT-conditions.8 Although they are always necessary, they are not always sufficient to conclude that a candidate (x, y) is indeed an equilibrium.

4.3.Shadow prices

Given m functions and m scalars b1, . . . , bm ∈ ℝ, the optimization problem

is of type (22) with the m restriction functions gi(x) = biai(x) and has the LAGRANGE function

For an intuitive interpretation of the problem (26), think of the data vector

as a plan for n products to be manufactured in quantities xj and of f(x) as the market value of x.

Assume that x requires the use of m materials in quantities a1(x) , . . . , am(x) and that the parameters b1, . . . , bm describe the quantities of the materials already in the possession of the manufacturer.

If the numbers y1, . . . , ym represent the market prices (per unit) of the m materials, we find that L(x, y) is the total value of the manufacturer’s assets:

The manufacturer would like to have that value as high as possible by deciding on an appropriate production plan x.

“The market” is an opponent of the manufacturer and looks at the value

which is the value of the materials the manufacturer must still buy on the market for the production of x minus the value of the production that the market would have to pay to the manufacturer for the production x.

The market would like to set the prices yi so that –L(x, y) is as large as possible. Hence:

The manufacturer and the market play a LAGRANGE game Λ.

An equilibrium (x*, y*) of Λ reflects an economic balance: Neither the manufacturer nor the market has a guaranteed way to improve their value by changing the production plan or by setting different prices.

In this sense, the production plan x* is optimal if (x*, y*) is an equilibrium. The optimal market prices are the so-called shadow prices of the m materials.

The complementary slackness condition (K1) says that a material which is in stock but not completely used by x* has zero market value:

The condition (K2) implies that x* is a production plan of optimal value f(x*) = L(x*, y*) under the given restrictions. Moreover, one has

which says that the price of the materials used for the production x* equals the value of the inventory under the shadow prices

Property (K3) says that the marginal change ∇xL(x*, y*)d of the manufacturer’s value L is negative in any feasible production modification from x* to x* + d and only profitable for the market because

We will return to production games in the context of cooperative game theory in Section 8.3.2.

4.4.Equilibria of convex LAGRANGE games

Remarkably, the KKT-conditions turn out to be not only necessary but also sufficient for the characterization of equilibria in convex LAGRANGE games with differentiable objective functions. This gives a way to compute such equilibria and hence to solve optimization problems of type (22) in practice9 :

Find a solution (x*, y*) ∈ X × for the KKT-inequalities. (x*, y*) will yield an equilibrium in Λ = and x* will be an optimal solution for (22).

Indeed, one finds:

THEOREM 3.3. A pair (x*, y*) ∈ X × is an equilibrium of the convex LAGRANGE game Λ = if and only if (x*, y*) satisfies the KKT-conditions.

Proof. From Lemma 3.2, we know that the KKT-conditions are necessary. To show sufficiency, assume that (x*, y*) ∈ X × satisfies the KKT-conditions. We must demonstrate that (x*, y*) is an equilibrium of the convex LAGRANGE game Λ = , i.e., satisfies

for the function L(x, y) = f(x) + yTg(x).

Since xL(x, y*) is a concave function, we have for every xX,

Because (K2) guarantees ∇xL(x*, y*)(xx*) ≤ 0, we conclude

which implies the first equality in (27). For the second equality, (K0) and (K1) yield g(x*) ≥ 0 and (y*)Tg(x*) = 0 and therefore:

4.5.Linear programs

A linear program (LP) in standard form is an optimization problem of the form

where c ∈ ℝn and b ∈ ℝm are parameter vectors and A ∈ ℝm×n a matrix, and thus is a mathematical optimization problem with a linear objective function f(x) = cTx and restriction function g(x) = bAx so that

The feasibility region of (28) is the set of all nonnegative solutions of the linear inequality system Axb:

The LAGRANGE function is

and yields for any x ≥ 0 and y ≥ 0:

The optimum value of L2 is found by solving the dual associated linear program

EX. 3.6. Since yTb = bTy and (cTyTA)x = xT(cATy) holds, the dual linear program (29) can be formulated equivalently in standard form:

The main theorem on linear programming is:

THOREM 3.4 (Main LP-Theorem). For the LP (28) the following holds:

 

(A)  An optimal solution x* exists if and only if both the LP (28) and the dual LP (29) have feasible solutions.
(B)  A feasible x* is an optimal solution if and only if there exists a dually feasible solution y* such that

Proof. Assume that (28) has an optimal solution x* with value z* = cTx*. Then cTxz* holds for all feasible solutions x. So the FARKAS Lemma10 guarantees the existence of some y* ≥ 0 such that

Noticing that y* is dually feasible and that L1(x*) ≤ L2(y) holds for all y ≥ 0, we conclude that y* is, in fact, an optimal dual solution:

This argument establishes property (B) and shows that the existence of an optimal solution necessitates the existence of a dually feasible solution. Assuming that (28) has at least one feasible solution x, it therefore remains to show that the existence a dual feasible solution y implies the existence of an optimal solution.

To see this, note first

So each dually feasible y satisfies –bTy ≤ −w*. Applying now the FARKAS Lemma to the dual linear program in the form (30), we find that a parameter vector x* ≥ 0 exists with the property

On the other hand, the primal-dual inequality yields L1(x*) ≤ w*. So x* must be an optimal feasible solution.

General linear programs. In general, a linear program refers to the problem of optimizing a linear objective function over a polyhedron, namely the set of solutions of a finite system of linear equalities and inequalities and, therefore, can be formulated as

with coefficient vectors c ∈ ℝn, b ∈ ℝm, d ∈ ℝk and matrices A ∈ ℝm×n and B ∈ ℝk×n.

If no equalities occur in the formulation (31), one has as a linear program in canonical form:

Because of the equivalence

the optimization problem (31) can be presented in canonical form:

Moreover, since any vector x ∈ ℝn can be expressed as the difference

of two (nonnegative) vectors one sees that each linear program in canonical form is equivalent to a linear program in standard form:

The LAGRANGE function of the canonical form is the same as for the standard form. Since the domain of L1 is now X = ℝn, the utility function L2(y) differs accordingly:

Relative to the canonical form, the optimum value of L2 is found by solving the linear program

Nevertheless, it is straightforward to check that Theorem 3.4 is literally valid also for a linear program in its canonical form.

Linear programming problems are particularly important in applications because they can be solved efficiently. In the theory of cooperative games with possibly more than two players (see Chapter 8), linear programming is a structurally analytical tool. We do not go into algorithmic details here but refer to the standard mathematical optimization literature.11

4.6.Linear programming games

A linear programming game is a LAGRANGE game that arises from a linear program. If c ∈ ℝn, b ∈ ℝm and A ∈ ℝm×n are the problem parameters, we may denote is by

where

is the underlying LAGRANGE function. Linear programming games are convex. In particular, Theorem 3.4 implies:

(1)A linear programming game admits an equilibrium if and only if the underlying linear program has an optimal solution.

(2)The equilibria of a linear programming game are precisely the pairs (x*, y*) of primally and dually optimal solutions.

EX. 3.7. Show that every randomized matrix game is a linear programming game. (See Section 3.3.2.)

 

 

 

_____________________________

1As in the Prisoner’s dilemma (Ex. 1.10).

2In hindsight!

3In hindsight.

4See also Appendix A.2 for more details.

5e.g., FAIGLE et al. [17].

6The idea goes back to J.-L. LAGRANGE (1736-1813).

7C.G. JACOBI (1804-1851).

8Named after the mathematicians KARUSH, KUHN and TUCKER.

9It is not the current purpose to investigate further computational aspects in details, which can be found in the established literature on mathematical programming.

10See Lemma A.6 in the Appendix.

11e.g., FAIGLE et al. [17].