Chapter 7

Potentials and Temperature

The temperature of a system depends on the measuring device in use, which is mathematically represented as a potential function. BOLTZMANN’s approach to the notion of temperature in statistical thermodynamics extends to general systems. Of particular interest are n-person matrix games where the temperature reflects the activity of the player set as a whole with respect to the total utility. The interpretation of the activity as a METROPOLIS process moreover indicates how the strategic decisions of individual players influence the expected value of the measuring device.

Consider a finite system that is in a state σ with probability πσ. Then has the entropy

The expected value of a potential υ ∈ ℝ will be

Let us think of υ as a numerical measuring device for a certain characteristic feature of . In a physical model, the number υσ could describe the level of inherent “energy” of in the state σ, for example. In economics, the function υ : → ℝ could be a representative statistic for the general state of the economy. In the context of an n-person game, υσ could possibly measure a degree of “activity” of the set N of players in the state σ, etc.

Of course, the activity level υσ depends on the particular characteristic feature that is investigated under υ. Different features of may display different activity levels in the same state σ.

1. Temperature

If the expected value μ = E(υ, π) is a measure for the degree of activity of relative to the probability distribution π and the potential υ, there may nevertheless be other probability distributions π′ with the same expected value

The idea is now to derive a canonical probability distribution β with a given expected value μ. To this end, we select β as the distribution with the largest entropy:

As it turns out (Lemma 7.1 below), every other probability distribution π yielding the same expected value μ of υ will have a strictly smaller entropy H(π) < H(β), which means that its specification would require more information. In this sense, β is the unique “freest” (i.e., least biased) distribution with expectation μ.

1.1.BOLTZMANN distributions

Given the potential υ : → ℝ with values υσ = υ(σ), any t ∈ ℝ defines a related BOLTZMANN1 (probability) distribution β(t) on with the components

REMARK 7.1. The distributions β(t) are also known as BOLTZMANN-GIBBS2 distributions. The function Z(t) is the associated so-called partition function.

Computing derivatives, one finds3 that the expected value of υ relative to b(t) can be expressed as the logarithmic derivative of the partition function:

NOTA BENE. Under a BOLTZMANN distribution β(t) with t ∈ ℝ, no state of will be impossible (i.e., occur with probability 0) and no state will occur with certainty (i.e., with probability 1) if has more than one state.

In the special case t = 0, one has Z(0) = ||. So β(0) is the uniform distribution on with the average potential value as its expectation:

Moreover, one observes the limiting behavior

In fact, one finds that any value μ between the minimum and the maximum of υ can be achieved as the expected value relative to a unique BOLTZMANN distribution. Moreover, the BOLTZMANN distribution is the one with the largest entropy relative to a prescribed expected value:

LEMMA 7.1. Let υ = minσ and υ* = maxσ υσ. Then:

 

(0)  υ* = υ*μ(t) = υ*(= υ*) holds for all t ∈ ℝ.
(1)  For every υ* < μ < υ*, there exists a unique t ∈ ℝ such that μ = μ(t).
(2)  H(π) < H(β(t)) holds for every probability distribution πβ(t) on mit expectation E(υ, π) = μ(t).

1.2.BOLTZMANN temperature

From Lemma 7.1, it is clear that one could characterize the expected value μ of a non-constant potential υ equally well by specifying the parameter t ∈ ℝ ∪ {−∞, +∞} of the BOLTZMANN distribution β(t) with expectation

In analogy with the BOLTZMANN model in statistical thermodynamics for the temperature, we call the related parameter

the temperature of the system relative to a potential with the expected value μ(1/T). Adjusting the notation accordingly to

the BOLTZMANN distribution β(T) has the coefficients

As the system “freezes” to the temperature T = 0, one obtains the extreme values of the potential υ as the expectations in the limit, depending on whether the limit 0 is approached from the positive or the negative side:

In contrast, all states of are equally likely at when the temperature T is infinite.

2.The METROPOLIS process

METROPOLIS et al. [29] have pointed out that BOLTZMANN distributions may result dynamically as limiting distributions of quite general stochastic processes.

To formulate the process, we associate with each σ a neighborhood (σ) ⊆ that renders connected in the following sense:

For any σ, τ, there are σ1, . . . , σk such that

where σ0 = σ and σk+1 = τ.

The METROPOLIS process specifies a parameter t ∈ ℝ and proceeds from an initial state σ0 in a (possibly infinite) sequence of state transitions relative to the potential υ in the following way.

When the process has reached the state σ, a neighbor τ(σ) is selected at random as a candidate for a transition στ and then proceeds as follows:

(1)If υτυσ, then the transition στ is made with certainty, i.e., with probability

(2)If υτ < υσ, then the transition στ is made with probability

and repeated otherwise. Hence the transition selection procedure is repeated with probability 1 − αστ.

It follows that the METROPOLIS state transitions define a MARKOV chain4 on with transition probabilities

The METROPOLIS process converges as a MARKOV chain to a limiting distribution on under quite general conditions. A sufficient condition is, for example:

PROPOSITION 7.1. Assume that all neighborhoods (σ) have the same size f = |F(σ)|. Then the METROPOLIS process converges to the BOLTZMANN probability distribution β(t) in the limit.

In other words: The process will arrive at the state σ after n iterations with probability

Proof. By Ex. A.12 of Section 7 of the Appendix, it suffices to check that the condition

is satisfied by any σ, τ. So assume υτ < υσ, for example. Then

Simulated annealing. The METROPOLIS process suggests a simple intuitive method for maximizing a function υ : X → ℝ over a finite set X:

(S0 Associate with each xX a neighborhood set (x) and designate an initial element x0X.
(S1 Choose a sequence t0, t1, . . . , tn, . . . of natural numbers tn with tn → ∞.
(Sn) Iterate the following procedure for n = 0, 1, 2, . . .:
(1)   If xnX has been constructed, choose a neighbor y(xn) randomly as a candidate for the next element xn+1.
(2)  If υyυxn, set xn+1 = y.
(3)  If υy < υxn, set

There is no guarantee that the simulating annealing heuristic will find a maximizer of υ. However, the analogy with the METROPOLIS procedure makes it intuitively plausible that the method might generate elements xn with the property

The schematic description of the simulated annealing heuristic leaves many implementation details vague. For example:

What is the best neighborhood structure in (S0)?

How should the numbers t0, t1, . . . in (S1)be chosen best?

Nevertheless, the method often produces quite good results in practice5 and has become a standard in the tool box for discrete optimization problems.

REMARK 7.2. The terminology “annealing” (= cooling) derives from the interpretation of the related parameters Tn = 1/tn as temperatures which simulate a cooling schedule Tn → 0.

REMARK 7.3. The import of the simulated annealing method is not so much the generation of BOLTZMANN distributions at fixed temperature levels but the convergence of the expected potential values μ(t) to the extremum:

The latter property can be guaranteed under much more general conditions than are formulated in Proposition 7.1.6

3.Temperature of matrix games

Let Γ = Γ(ui | iN) be an n-person game with player set N = {1, . . . , n} where each player iN has a finite set Xi of strategic resources and a utility function

In the model of randomized matrix games, it is assumed that the players i choose probability distributions π(i) on their strategy sets Xi independently from each other and then select elements xiXi according to those distributions.

Let us drop the stochastic independence assumption and consider the more general model where the joint strategy

would be chosen by the player set N with a certain probability πx. The aggregated total utility value is then expected to be

The players’ total utility

is a potential on . So one may consider the (BOLTZMANN) temperature relative to u. In the case

we say that Γ is is played at temperature T. If |T| ≈ ∞ (i.e., |T| is very large), we expect about the average value of the total utility:

If T > 0 is very small (i.e., T ≈ 0), then we may expect about the maximal total utility:

Similarly, if T ≈ 0 and T < 0 holds, about the minimal total utility value is to be expected:

A METROPOLIS model. In order to model the dynamic behavior of the player set N as a METROPOLIS process, define the set

A selection (i, y) ∈ , when the game is in the state x, corresponds to a possible transition

with the interpretation that player i considers the alternative strategy yXi as a candidate for an exchange against the current xiXi. Hence, if (i, y) is chosen randomly, the probability for the corresponding transition is

So each x has the same number f = || of neighbors. Clearly, this neighborhood structure is connected. Hence Proposition 7.1 shows that the process converges to a BOLTZMANN probability distribution:

Recall that every state x occurs with strictly positive probability in a BOLTZMANN distribution at temperature T ≠ 0. So we should not expect convergence towards an equilibrium with respect to u.

This observation does not contradict the guaranteed existence of an equilibrium in the randomized version of a matrix game (Theorem 6.1). The reason can be found in the different roles of the players in the models:

(1)In a randomized matrix game, each player iN chooses, individually, a probability distribution π(i) for the selection a xiXi with the goal of optimizing the individual utility ui.

(2)In the METROPOLIS process, the total utility u of the group N of players is the key objective.

REMARK 7.4 (Social justice). It appears to be in the joint interest of the community N of players to play Γ at a temperature T that is close to 0 but positive if a large total utility value is desired and negative if a minimal value is sought.

The potential function u is equivalent (up to the scaling factor n) to the average utility function of the members of N:

A high group average does not necessarily imply a guaranteed high utility value for each individual member in N, however.

To formulate it bluntly:

Even when a high average utility value is used as a criterion for “social justice” in N, there may still be members of N that are not treated “fairly”.

The interplay of different interests (individual utility of the players vs. combined utility of the set of all players) is studied in more details within the framework of cooperative games in Chapter 8.

 

 

 

_____________________________

1 L. BOLTZMANN (1844–1906).

2 J.W. Gibbs (1839–1903).

3 See also Section 6 of the Appendix.

4 See Section 7 in the Appendix.

5 See, e.g., S. KIRKPATRICK et al. [26].

6 See, e.g., U. FAIGLE and W. KERN [14].