Differential Games
()
About this ebook
Geared toward graduate students, Differential Games will be of particular interest to professionals in the fields of electrical engineering, industrial engineering, economics, and mathematics. Although intended primarily for self-study, it can be used as a core or ancillary text in courses in differential games, game theory, and control theory.
Related to Differential Games
Titles in the series (100)
History of the Theory of Numbers, Volume II: Diophantine Analysis Rating: 0 out of 5 stars0 ratingsAn Adventurer's Guide to Number Theory Rating: 4 out of 5 stars4/5Methods of Applied Mathematics Rating: 3 out of 5 stars3/5Dynamic Probabilistic Systems, Volume II: Semi-Markov and Decision Processes Rating: 0 out of 5 stars0 ratingsLaplace Transforms and Their Applications to Differential Equations Rating: 5 out of 5 stars5/5A Catalog of Special Plane Curves Rating: 2 out of 5 stars2/5Calculus: An Intuitive and Physical Approach (Second Edition) Rating: 4 out of 5 stars4/5Gauge Theory and Variational Principles Rating: 2 out of 5 stars2/5The Calculus Primer Rating: 0 out of 5 stars0 ratingsFirst-Order Partial Differential Equations, Vol. 2 Rating: 0 out of 5 stars0 ratingsFirst-Order Partial Differential Equations, Vol. 1 Rating: 5 out of 5 stars5/5Calculus Refresher Rating: 3 out of 5 stars3/5Applied Multivariate Analysis: Using Bayesian and Frequentist Methods of Inference, Second Edition Rating: 0 out of 5 stars0 ratingsInfinite Series Rating: 4 out of 5 stars4/5Mathematics for the Nonmathematician Rating: 4 out of 5 stars4/5Theory of Approximation Rating: 0 out of 5 stars0 ratingsLinear Algebra Rating: 3 out of 5 stars3/5Analytic Inequalities Rating: 5 out of 5 stars5/5Advanced Calculus: Second Edition Rating: 5 out of 5 stars5/5A History of Mathematical Notations Rating: 4 out of 5 stars4/5Optimization Theory for Large Systems Rating: 5 out of 5 stars5/5Numerical Methods Rating: 5 out of 5 stars5/5Chebyshev and Fourier Spectral Methods: Second Revised Edition Rating: 4 out of 5 stars4/5Journey into Mathematics: An Introduction to Proofs Rating: 4 out of 5 stars4/5Theory of Games and Statistical Decisions Rating: 4 out of 5 stars4/5Topology for Analysis Rating: 4 out of 5 stars4/5An Introduction to Lebesgue Integration and Fourier Series Rating: 0 out of 5 stars0 ratingsFifty Challenging Problems in Probability with Solutions Rating: 4 out of 5 stars4/5The History of the Calculus and Its Conceptual Development Rating: 4 out of 5 stars4/5Fourier Series and Orthogonal Polynomials Rating: 0 out of 5 stars0 ratings
Related ebooks
Elementary Theory of Numbers Rating: 4 out of 5 stars4/5Game Theory: Fundamentals and Applications Rating: 0 out of 5 stars0 ratingsTwo-Person Game Theory Rating: 0 out of 5 stars0 ratingsThe Master Book of Mathematical Recreations Rating: 5 out of 5 stars5/5The Mathematics of Games of Strategy Rating: 4 out of 5 stars4/5A Course of Mathematics for Engineers and Scientists Rating: 0 out of 5 stars0 ratingsMystifying Mathematical Puzzles: Golden Spheres, Squared Eggs and Other Brainteasers Rating: 0 out of 5 stars0 ratingsThe Mathematics of Games Rating: 0 out of 5 stars0 ratingsCalculus and Its Applications Rating: 4 out of 5 stars4/5Modern Algebra Essentials Rating: 0 out of 5 stars0 ratingsIntroduction to Real Analysis: An Educational Approach Rating: 0 out of 5 stars0 ratingsStaying On The Leading Edge Rating: 0 out of 5 stars0 ratingsA Geometric Algebra Invitation to Space-Time Physics, Robotics and Molecular Geometry Rating: 0 out of 5 stars0 ratingsKähler Metric and Moduli Spaces: Advanced Studies in Pure Mathematics, Vol. 18.2 Rating: 0 out of 5 stars0 ratingsMeasure and Integration: A Concise Introduction to Real Analysis Rating: 0 out of 5 stars0 ratingsThe Power of Strategy Innovation: A New Way of Linking Creativity and Strategic Planning to Discover Great Business Opportunities Rating: 0 out of 5 stars0 ratingsReal Analysis: A Historical Approach Rating: 0 out of 5 stars0 ratingsPascal Triangle Analogues Introduction Rating: 0 out of 5 stars0 ratingsA Course of Higher Mathematics: Adiwes International Series in Mathematics, Volume 3, Part 1 Rating: 4 out of 5 stars4/5An Introduction to Proof through Real Analysis Rating: 0 out of 5 stars0 ratingsWhy Cats Land on Their Feet: And 76 Other Physical Paradoxes and Puzzles Rating: 4 out of 5 stars4/5Automorphic Forms and Geometry of Arithmetic Varieties Rating: 0 out of 5 stars0 ratingsMathematical Methods and Theory in Games, Programming, and Economics: Volume 2: The Theory of Infinite Games Rating: 0 out of 5 stars0 ratingsDiscrete Optimization: The State of the Art Rating: 0 out of 5 stars0 ratingsThe Gentle Art of Mathematics Rating: 0 out of 5 stars0 ratingsThe Umbral Calculus Rating: 3 out of 5 stars3/5What Is Random?: Chance and Order in Mathematics and Life Rating: 4 out of 5 stars4/5Invariant Variational Principles Rating: 5 out of 5 stars5/5A Course in Complex Analysis Rating: 0 out of 5 stars0 ratingsPath Integrals and Quantum Processes Rating: 0 out of 5 stars0 ratings
Mathematics For You
Standard Deviations: Flawed Assumptions, Tortured Data, and Other Ways to Lie with Statistics Rating: 4 out of 5 stars4/5The Little Book of Mathematical Principles, Theories & Things Rating: 3 out of 5 stars3/5Algebra I Workbook For Dummies Rating: 3 out of 5 stars3/5Calculus Made Easy Rating: 4 out of 5 stars4/5Flatland Rating: 4 out of 5 stars4/5Quantum Physics for Beginners Rating: 4 out of 5 stars4/5My Best Mathematical and Logic Puzzles Rating: 5 out of 5 stars5/5Basic Math & Pre-Algebra For Dummies Rating: 4 out of 5 stars4/5Algebra - The Very Basics Rating: 5 out of 5 stars5/5The Thirteen Books of the Elements, Vol. 1 Rating: 0 out of 5 stars0 ratingsReal Estate by the Numbers: A Complete Reference Guide to Deal Analysis Rating: 0 out of 5 stars0 ratingsAlan Turing: The Enigma: The Book That Inspired the Film The Imitation Game - Updated Edition Rating: 4 out of 5 stars4/5A Mind for Numbers | Summary Rating: 4 out of 5 stars4/5The Everything Everyday Math Book: From Tipping to Taxes, All the Real-World, Everyday Math Skills You Need Rating: 5 out of 5 stars5/5Limitless Mind: Learn, Lead, and Live Without Barriers Rating: 4 out of 5 stars4/5Sneaky Math: A Graphic Primer with Projects Rating: 0 out of 5 stars0 ratingsGame Theory: A Simple Introduction Rating: 4 out of 5 stars4/5The Golden Ratio: The Divine Beauty of Mathematics Rating: 5 out of 5 stars5/5Mental Math Secrets - How To Be a Human Calculator Rating: 5 out of 5 stars5/5The Everything Guide to Algebra: A Step-by-Step Guide to the Basics of Algebra - in Plain English! Rating: 4 out of 5 stars4/5Relativity: The special and the general theory Rating: 5 out of 5 stars5/5Algebra I For Dummies Rating: 4 out of 5 stars4/5Introducing Game Theory: A Graphic Guide Rating: 4 out of 5 stars4/5The Math Book: From Pythagoras to the 57th Dimension, 250 Milestones in the History of Mathematics Rating: 3 out of 5 stars3/5Geometry For Dummies Rating: 5 out of 5 stars5/5Algebra II For Dummies Rating: 3 out of 5 stars3/5
Reviews for Differential Games
0 ratings0 reviews
Book preview
Differential Games - Avner Friedman
MATHEMATICS
Introduction
In a differential game between two players, y and z, we are given a dynamical system
(1)
an initial condition
(2)
and a payoff P(y, z), usually in the form of
(3)
where is the first time that (t, x(t)) intersects a given set; y wishes to maximize the payoff, and z wishes to minimize it. The players y, z are entirely free to choose any control functions y(t), z(t), that is, any measurable functions y(t), z(t) with values in given sets Y, Z, respectively.
What makes the situation complicated is the information pattern: each player at time t has perfect information of the controls chosen prior to t by himself and by his opponent, but he has none regarding the future choice of control by his opponent.
Suppose there exists controls ӯ(t(t) such that
(4)
for any controls y, z and set V = P(ӯ, z). We can then state the following: when the player z , no matter what y does, z will always render the payoff ≤ V. Similarly, when the player y chooses ӯ, no matter what z does, y will always render the payoff ≥ V. Thus, if y and z do not act on the assumption that their opponent may play in a stupid
way, their best choice is y = ӯ, z . The number V is the payoff when both players are doing their best; it is called the value of the game. The pair (ӯ(t(t)) is called a saddle point.
Unfortunately, such a saddle point (ӯ(t(t)) does not exist in general.
In the heuristic work on the subject (see Isaacs [2]), we introduce the concept of pure strategies y(t, x), z(t, x) for the players y, z. These are measurable functions with values in Y, Z, respectively, piecewise Lipschitz continuous in x. To such a pair we may match a payoff P(y, z), provided a solution of (1) and (2) exists. We then define a saddle point in pure strategies as a pair (ӯ(t, x(t, x)) of pure strategies for which (4) holds for any pure strategies y = y(t, x), z = z(t, x). Again, a saddle point in pure strategy does not exist in general.
In this book we introduce more general concepts of value and saddle point. We form two sequences of δ-approximating models {Gδ} and {Gδ} (δ varies over a sequence that converges to 0). In the model Gδ, y has a δ-advantage of information and in the model Gδ, z has a δ-advantage of information. In the models Gδ and Gδ there is a natural way of defining values Vδ and Vδ, respectively, since both players do not act simultaneously, but rather step-by-step (the size of each step being δ). It is clear that Vδ ≥ Vδ. If lim Vδ = lim Vδ = V, we say that the game has value equal to V. A fundamental theorem asserts that V exists (under suitable conditions).
Next, a strategy Γ for the player y is defined as a sequence of δ-strategies Γδ; each Γδ is a vector of maps telling y what to do in a scheme G(δ) (similar to Gδ) which δ-approximates the actual game. This notion of strategy includes the notion of pure strategy. In fact, to a pure strategy ӯ(t, x. In particular, if ӯ = ӯ(tis called a constant strategy. If Δ is a strategy for z, the payoff, set P) is a pair of strategies for which
(5)
for any strategies Γ, Δ for y, z, respectively. A fundamental theorem asserts that a saddle point exists (under suitable conditions).
) coincides with the saddle point in pure strategies when the latter is known to exist (by solving the Hamilton-Jacobi equation). Thus the abstract existence theory ties together various heuristic computational results in a rigorous manner.
The fact that differential games are played in a continuous (and not in a discrete) manner causes fundamental difficulties in the formulation of the theory. On the other hand, in a differential game we are given a dynamical system of fixed structure. This enables us to use various tools of analysis that are not applicable in discrete games. As a result of these two differences, the two theories of games, differential and discrete, have (at present) nothing substantial in common.
CHAPTER 1
Definition of a Differential Game
1.1 Results on Ordinary Differential Equations
We shall denote the m-dimensional euclidean space by Rm. A point (x1,..., xm) in Rm will be denoted, briefly, by x. The norm |x)¹/²- An m-vector function (f1, ... , fm) will be denoted, briefly, by f.
Consider a system of m ordinary differential equations
(1.1.1)
together with an initial condition
(1.1.2)
where f(t, x) = (f1(t, x), ... , f,m(t, x)) is defined in some bounded domain D of the (m + 1)-dimensional (t, x)-space and (t0, xo) is a point of D. If we integrate both sides of (1.1.1), making use of (1.1.2), we get
(1.1.3)
The domain D is contained in a strip {α ≤ t ≤ β} x Rm.
THEOREM 1.1.1 Assume that (i) f(t, x) is measurable in t for each fixed x and continuous in x for each fixed t; (ii) |f(t, x)| ≤ k0(t), where . Then there exists a solution x(t) of (1.1.3) in an interval a < t < b containing to; furthermore, x(a + 0) and x(b − 0) exist and the points (a, x(a + 0)), (b, x(b − 0)) lie on the boundary of D.
THEOREM 1.1.2 Assume, in addition to (i), (ii) of Theorem 1.1.1, that there exists a non-negative function k(t) (α ≤ t ≤ β) such that and
(1.1.4)
for all (t, x) and ) in D. Then, in any interval (a, b) containing to there is at most one solution of (1.1.3).
Thus, under the assumptions of Theorem 1.1.2, the solution of (1.1.3) asserted in Theorem 1.1.1 is unique.
The proofs of Theorems 1.1.1 and 1.1.2 can be found, for instance, in Coddington and Levinson [1].
THEOREM 1.1.3 Let D = {α < t < β} x Rm and assume that f(t, x) satisfies the condition (i) of Theorem 1.1.1, and the inequality
(1.1.5)
where k0(t) dt < ∞. Assume also that for every R > 0 there is a nonnegative function kR(t) such that kR(t) dt < ∞ and
if .
Then there exists a unique solution of (1.1.3) for α ≤ t ≤ β.
Proof. We shall prove that there is a constant K such that, for any subinterval (a, b) of (α, β), the solution x(t) of (1.1.3) in (a, b), if existing, must satisfy
(1.1.6)
If we then apply Theorem 1.1.1 with D = {x; lxl < K + 1}, we find that the numbers a, b in the assertion of that theorem must be equal to α and β, respectively; but this, together with Theorem 1.1.2, gives the assertion in Theorem 1.1.3.
To prove (1.1.6) we use (1.1.5) in (1.1.3). We get
(1.1.7)
This inequality easily yields (1.1.6).
Note that the solution x(t) of (1.1.3) is an indefinite integral (which is the same as an absolutely continuous function) and its derivative ẋ(t) exists and equals f(t, x(t)) almost everywhere. Conversely, if an absolutely continuous function x(t) satisfies (1.1.1) almost everywhere and also satisfies (1.1.2), it is a solution of (1.1.3).
From now on when we speak of a solution x(t) of (1.1.1) and (1.1.2) we shall understand that x(t) is absolutely continuous, it satisfies (1.1.1) almost everywhere, and it satisfies (1.1.2). Thus x(t) is a solution of (1.1.1) and (1.1.2) if and only if it is a solution of (1.1.3).
Problems
1. Let ϕ, ψ be continuous functions in a closed interval [a, b], and let χ be non-negative and an integrable function in (a, b). Suppose that
Prove that
Use this result to deduce (1.1.6) from (1.1.7).
2. Assume that f(t, x) satisfies all the conditions of Theorem 1.1.3, except for (1.1.5), and
x∙f(t, x) ≤ γ lxl² + γ0 (γ , γ0 constants),
for all (t, x) in D. Prove that the assertion in Theorem 1.1.3 is still valid.
1.2 Differential Equations with Control Functions
Consider a differential system of m equations
(1.2.1)
with an initial condition
(1.2.2)
If we take y = y(t) and z = z(t), we obtain a differential system of the form considered in Section 1.1.
We fix two sets, Y and Z; Y is a compact (i.e., closed and bounded) subset of some euclidean space Rp and Z is a compact subset of some euclidean space Rq. We shall call Y the control set for y and Z the control set for z.
A measurable function y(t) whose values belong to Y for almost all t is called a control function for y. Similarly, a measurable function z(t) whose values belong to Z for almost all t is called a control function for z.
We shall assume the following:
(A1) f(t, x, y, z) is continuous in (t, x, y, z) ∈ [to, To] × Rm × Y × Z.
(A2) There is a non-negative function k(t) (t0 ≤ t ≤ T0, such that
. ∣f(t, x, y, z)∣ ≤ k(t)(1 + |xl)
for all (t, x, y, z) in [t0, T0] × Rm × Y × Z.
(A3) For each <R > 0 there is a non-negative function kR(t) (t0 ≤ t ≤ T, such that
for all t ∈ [t0, T0], y ∈ Y, z ∈ Z and lxl ≤ R, |≤ R.
Now take any control functions y = y(t), z = z(t) in (1.2.1). Then (1.2.1) and (1.2.2) become
(1.2.3)
Using the assumptions (A1)−A3), we easily see that the system (1.2.3) satisfies all the conditions imposed in Theorem 1.1.3; in verifying the condition (i) of Theorem 1.1.1, we use the fact that a continuous function of a measurable function is measurable.
THEOREM 1.2.1 Let the conditions (A1)–(A3) hold. Then for any control functions y = y(t), z = z(t) in t0 ≤ t ≤ T0 there exists a unique solution x(t) of (1.2.3) in t0 ≤ t ≤ T0. Furthermore, there is a constant H such that for any control functions y(t), z(t) the corresponding solution x(t) of (1.2.3) satisfies |x(t)| ≤ H for t0 ≤ t ≤ T0.
Proof. Extend f (t, x, y, z) to t0−1 ≤ t < t0 by f (t, x, y, z) = f (t0, x, y, z). Extend y(t) and z(t) to t0 − 1 ≤ t < t0 as constants. Now apply Theorem 1.1.3 with α = t0 − 1,β= T0 to deduce the existence and uniqueness of a solution of (1.2.3). The last assertion of Theorem 1.2.1 follows from the proof of Theorem 1.1.3 [cf. the derivation in (1.1.6)].
Remark. The condition (A2) is fairly restrictive. It may be replaced by the slightly weaker condition
x∙f(t, x, y, z) ≤ γ|x|² + γ0 (γ, γ0 constants)
(see Problem 2, Section 1.1). However, without such a condition we cannot assert, in general, the existence of a solution x(t) of (1.2.1) (in t0 ≤ t ≤ T and (1.2.2) for any control functions y(t), z(t). Note that if (A1) and (A3) hold, and we can verify (for a particular f) that the solution of (1.2.1) (in t0 ≤ t ≤ T) and (1.2.2) exists for any y(t), z(t) and that |x(t)| ≤ H, H a constant independent of y(t), z(t), the condition (A2) is superfluous. Indeed, we can then modify the definition of f(t, x, y, z) for |x| > H in such a way that the modified f satisfies (A1)−(A3). Since this will not affect the solutions x(t) with the original f, we may replace the original f with the modified one in all our considerations.
The solution x(t) of, (1.2.3) is called the trajectory corresponding to the controls y(t), z(t).
In the future we shall call the variables y and z players. When we substitute y = y(t), z = z(t) in (1.2.1) we shall say that the player y chooses a control function y(t) and the player z chooses a control function z(t).
Let F be a closed set in the (t, x) space. We assume that
(1.2.4)
We shall call F the terminal set.
Given control functions y(t), z(t) for t0 ≤ t ≤ T0, denote the corresponding trajectory by x(t, x) belongs to F. Since F exists. We shall denote it by t(x) or t(y, z) and call it the capture time of the trajectory x(t), or the capture time corresponding to the controls y = y(t), z = z(t). We shall say that capture occurs at time t(y, z).
Let g(t, x) be a given function defined for all (t, x) in [t0, T0] x Rm and bounded on bounded sets. Let h(t, x, y, z) be a given function that satisfies the following condition:
(B1) h(t, x, y, z) is continuous in (t, x, y, z) ∈ [t0, T0] x Rm x Y x Z.
We introduce a functional
(1.2.5)
and call it the payoff functional. The functional g(t(y, z), x(t(y, z))) is called the terminal part of the payoff, and the functional
is called the integral part of the payoff. When y and z make a particular choice of y = y(t), z = z(t), the corresponding trajectory x(t) and the capture time t(y, z) are well defined. Thus the payoff P(y, z) is well defined for each choice of controls made by the players y and z.
The aim of y will be to maximize the payoff and the aim of z will be to minimize it. The precise circumstances under which these aims are to be carried out will be explained in the following sections.
In conclusion of this section we want to generalize slightly the concept of a payoff in the case in which the terminal set is given by t = T0; that is,
F = {(T0, x); x ∈ Rm}.
In this case t(y, z) = T0 for all y, z. the set of all trajectories [when y = y(t), z = z(t) vary over the sets of all the control functions defined on t0 < t ≤ Tbe a given functional defined on the space Xt0, T0. We assume that the functional is bounded; that is, there exists a constant K such that
We now introduce a payoff, of the form
(1.2.6)
Problems
1. Let (A1)−(A3) hold. Prove that there is a constant C such that for any trajectory x(t) of (1.2.1), (1.2.2) [x(t) − x(s)| ≤ C|t − s|.
2. Consider the system
Let F = {(t, x1, x2); 0 ≤ t < 3, x1, = x2} ∪ {(t, x1, x2); 3 ≤ t < ∞, (x1,x2) ∈ R²}. Compute the payoff
for y = 4t³, z = 2t.
1.3 Upper and Lower δ-Strategies and δ-Games
As mentioned in Section 1.2, the player y will want to maximize the payoffs (1.2.5), or (1.2.6) and the player z will want to minimize it. We assume that the circumstances are such that each player can observe his own past behavior and that of his opponent; that is, each player has perfect information of the past. None, however, has information regarding the future behavior of his opponent. The question is how he should act in the future.
Before introducing a rigorous scheme for this situation, we consider an approximate situation in which y has better up-to-date information than z. More precisely, y knows all the past choices made by y and z; that is, y(s) and z(s) for all s ≤ t, whereas z knows the choices made only for s ≤ t − δ. Here δ is some small positive number. We shall introduce a scheme for this δ-approximate situation.
Let n be any positive integer. Set
We divide the interval [t0, T0] into n intervals Ij of equal length δ:
Ij = {t; tj−1 < t ≤ tj}
where tj = t0 + jδ, 1 ≤ j ≤ n.
Denote by Yj and Zj the sets of all measurable functions on Ij whose values are, almost everywhere, contained in Y and Z, respectively. (Two measurable functions in Yj, or in Zj, are considered equal if they agree almost everywhere on Ij.)
Given sets A1,..., Ak, we denote by A1 × ⋯ × Ak their (cartesian) product. Thus, a point in the product has the form (a1,..., ak), where ai ∈ Ai for each i, and two points (a1,..., ak) and (b1,..., bk) are equal if and only if ai = bi for all i.
Let Γδ,j be any map from Z1 × Y1 × Z2 × Y2 × ⋯ × Zj-1 × × Yj-1 × × Zj into Yj. The vector
Γδ = (Γδ,1,..., Γδ,n)
is then called an upper δ-strategy for y. Similarly, we define an upper δ-strategy for z as a vector
Δδ = (Δδ,1,...,Δδ,n,
where Δδ,j is any map from Y1 × Z1 × Y2 × Z2 × ··· × Yj × 1 × Zj−1 × Yj into Zj.
Next let Γδ,j (2 ≤ j ≤ n) be any map from Y1 × Z1 × Y2 × Z2 × ··· × Yj−1 × Zj−1 into Yj and denote by Γð,1 any function in Y1. The vector
Γð = (Γδ,1, ..., Γδ,n)
is called a lower δ-strategy for y. Similarly, we define a lower δ-strategy for z as
Δδ = (Δδ,1,...., Δδ,n),
where Δδ;1 is any function in Z1 and Δδ,j (2 ≤ j ≤ n) is any map from Z1 × Y1 × Z2 × Y2 × ·· × Zj-1 × Yj-1 into Zj.
Given any pair (Δδ, Γδ), we can uniquely construct control functions yδ(t), zδ(t) with components Yj, Zj on Ij by z1= Δδ,1,y1 = Γδ,1(z1), and, for 2 ≤ j ≤ n,
(1.3.1)
(1.3.2)
We call yδ, zδ) the outcome of (Δδ, Γδ). We write the corresponding payoff, given by either (1.2.5) or (1.2.6), in the form
(1.3.3)
Note that (1.3.1) means that the player z chooses a control on Ij based only on the information he has [on y(t) and z(t)] in the intervals I1,..., Ij-1. On the other hand, (1.3.2) means that y chooses a control on Ij based on the information he has [on y(t) and z(t)] in the intervals I1,...,Ij-1 and on the additional information he has on z(t) in Ij.
We shall call the above scheme an upper δ-game and denote it by Gδ. Thus, in Gδ, y chooses an upper δ-strategy Γδ and z chooses a lower δ-strategy Δδ and then they play
according to (1.3.1), (1.3.2). The payoff of this play
is given by (1.3.3). Let us call a choice made by a player in an interval Ij a move; y is the one that makes the last move (in In). He should therefore try to choose Γδ,n = such that
Since a maximum may not exist, we should modify the aim of y: his aim in the last move is to make the payoff approximately
equal to
(1.3.4)
With this anticipated behavior on the part of y in In, z should make his move in In in such a way that the functional in (1.3.4) is approximately
equal to
These considerations extend step-by-step to the intervals In-1 , In-2,..., I1 We conclude that if each player tries to do his best (in the sense explained above) the resulting payoff should be approximately
equal to the number
(1.3.5)
The number Vδ is called the upper δ-value of the game Gδ.
Similarly, we can define the concept of a lower δ-game Gδ and lower δ-value Vδ. Here y chooses a lower δ-strategy Γδ and z chooses an upper δ-strategy Δδ. We then construct control functions yδ(t) and zδ(t) with components yj, zj on Ij, given by y1 = Γδ,1, z1 = Δδ, 1(y1), and, for 2 ≤ j ≤ n,
yj = Γδ,j (y1,z1, y2,z2,..... yj-1, zj-1),
zj = Δδ,j (y1, z1, y2, z2, ... , yj−1, zj−1, yj).
We call (yδ, zδ) the outcome of (Γδ, Δδ) and write
P(yδ, zδ) = P[Γδ, Δδ] = P[Γδ,1, Δδ,¹,..., Γδ,n, Δδ,n].
The lower δ-value Vδ is defined by
Remark. In the definition of an upper δ-strategy Γδ = (Γδ,1,...., Γδ,n) we have required that each Γδ,j be a map from
Aj ≡ Z1 × Y1 × Z2 × Y2 × ··· × Zj−1 × Yj−1 x Zj iinto Yj.
Actually, in determining the outcome of (Δδ, Γδ) (for any Δδ) we need only to know the values of Γδ,j on the subspace
Âj ≡ Zi × Ŷ1 × Z2 × Ŷ2 x ··· × Zj-1 × Ŷj−1 Zj,
where Ŷi is the range of Γδ,i. Therefore, when in the future we explicitly construct components Γδ,j of an upper δ-strategy Γδ = (Γδ,1,..., Γδ,n), we shall often define the Γδ,j only on the subspace Âj and not bother to extend them into Aj; the results will not depend on whatever extension is chosen. The same remark applies to upper δ-strategies for z and to lower δ-strategies for y and z.
Problems
1. Consider the differential system
with control sets 0 ≤ y ≤ 1, 0 ≤ z ≤ 2, and payoff
Show that Vδ = Vδ = 3 for δ = 3/n, n = 1, 2,....
2. If in the preceding example
show that Vδ = Vδ = 6(2 − α).
1.4 Properties of Upper and Lower δ-Values
THEOREM 1.4.1 Let the conditions (A1)−(A3) and (B1) hold. Then
(1.4.1)
where
Proof. Set
(1.4.2)
We can easily verify, step-by-step with respect to k, for k = n, n–1,..., 1, that
For k = 1 this gives
Hence B ≤ Vδ. We shall now prove that
(1.4.3)
Let e be any positive number. Denote by P(z1, y1,... Zn, yn) the payoff corresponding to the controls y and z with components yj, zj on Ij. Then for any z1,y1,....., zn-1 yn-1, zn we can find
such that
Denote the map ỹn Then for any Δδ,1, Γδ,1,..., Δδ,n−1, Γδ,n−1, Δδ,n we have
It follows that
(1.4.4)
Set
(1.4.5)
Proceeding as before, with P replaced by Pnsuch that
Combining this with (1.4.4), (1.4.5), we get
(1.4.6)
Setting
we can proceed as before with P replaced by Pnsuch that
Combining this with (1.4.6), we get
such that
that is,
Hence
Taking ε → 0, (1.4.3) follows.
It remains to prove the first equality in (1.4.1). Set
we have
we conclude that A ≥ Vδ. It thus remains to show that
(1.4.7)
Define
(1.4.8)
such that
(1.4.9)
Next, define
(1.4.10)
such that
Combining this with (1.4.8)–(1.4.10), we conclude that
Proceeding similarly step-by-step, we arrive at the inequality
Hence
Taking ε → 0, we obtain the inequality (1.4.7). This completes the proof of Theorem 1.4.1.
THEOREM 1.4.2 Let the conditions (A1)−(A3) and (B1) hold. Then
(1.4.11)
The proof is similar to the proof of Theorem 1.4.1.
Let yj and zj be the components in Ij of y and z, respectively, where (y, z) is the outcome of (Δδ, Γδ). Write, for any k = 1, 2,..., n,
(1.4.12)
Then we can easily prove step-by-step with respect to k, for k = n, n–1, ..., 1, that
(1.4.13)
Here, by infzk we mean infzk ε Zk. Similarly, by supyk we mean SUPYk ε Yk·
The relation (1.4.13) with k = 1 reduces to
(1.4.14)
Similarly we show that
(1.4.15)
Next we can easily show, step-by-step with respect to k, for k = n, n–1, ..., 1, that
Combining this inequality for k = 1 with (1.4.14), (1.4.15), we obtain the following result:
THEOREM 1.4.3 Let the conditions (A1)−(A3) and (B1) hold. Then
(1.4.16)
We shall finally prove the following:
THEOREM 1.4.4 Let the conditions (A1)−(A3) and (B1) hold and let n and k be any positive integers n′ = kn, δ = (To − t0)/n, δ′ = (T0 − t0)/n′. Then
(1.4.17)
Proof. For simplicity we shall give the proof only in case k = 2. Divide each interval Ij into two intervals I′j and I″j of equal length:
where Ij = {t; tj−1 < t ≤ tj}, tj = t0 + jδ the spaces of all control functions of y similarly with respect to zcorrespond in a one-to-one way to the points yj of Yj by the mapping
Any functional W(yj) defined on Yj can therefore be written as a functional Wj(i.e., W) = W(yj)), and, as easily seen,
(1.4.18)
where
Similarly, if we define
then, for any functional U(zj) = U) we have
(1.4.19)
Denote the component of a control y on Ij by yj and the components of y by y’j , respectively. Define similarly the components zj of a control z. Finally, write the payoff P(y, z) in either one of the forms
By (1.4.18), (1.4.19) we deduce that
(1.4.20)
where the last inequality follows from the fact that for any functional Q(u, v),
Proceeding to the interval In-1, we obtain from (1.4.20) and (1.4.18), (1.4.19) for j = n = 1,
Proceeding in this, way step-by-step, we arrive at the inequality
(1.4.21)
By (1.4.14) the left-hand side of (1.4.21) is equal to Vδ. Similarly, the right-hand side of (1.4.21) is equal to Vδ’. Consequently (1.4.21) reduces to Vδ ≥ Vδ’. Similarly we prove that Vδ ≤ Vδ’
Problems
1. Prove (1.4.13).
2. Prove (1.4.18).
3. Let Q be a bounded real-valued function defined on a set V x W and denote by S the set of all maps from V into W. Show that
4. Give an example of a continuous function M(y, z) of two real variables y, z,