Preprint
Article

This version is not peer-reviewed.

Linear Quadratic Pursuit–Evasion Games on Time Scales

A peer-reviewed article of this preprint also exists.

Submitted:

22 September 2025

Posted:

23 September 2025

You are already at the latest version

Abstract
In this paper, we unify and extend the linear quadratic pursuit-evasion games to dynamic equations on time scales. Here we seek to a mixed strategy for a pair of linear players. We show that when the final state is fixed, these (open–loop) strategies can be written in terms of a zero-input state difference. On the other hand, when the final states are free, we find closed-loop strategies in terms of an extended state.
Keywords: 
;  ;  ;  

1. Introduction

The theory of deterministic pursuit-evasion games can single-handedly be attributed to Isaacs in the 1950s [1,2]. Here, Isaacs first considered differential games as two-player zero-sum games. One early application was formulation of missile guidance systems during his time with the RAND Corporation. Shortly thereafter, Kalman among others initiated the linear quadratic regulator and tracking (LQR) and (LQT) in the continuous and discrete cases (see [3,4,5,6]). Since then, the concept of pursuit-evasion games and optimal control have been closely related, each playing a fundamental role in control engineering and economics. One breakout paper to combine these concepts was written by Ho, Bryson, and Baron. Together, they studied linear quadratic pursuit-evasion games (LQPEG) as regulator problems [7,8]. In particular, this work included a three-dimensional target interception problem. Since then, there have been a number of papers that have extended these results in the continuous and discrete cases. One of the issues that researchers have faced in the past is the discrete nature of these mixed strategies.
In 1988, Stefan Hilger initiated the theory of dynamic equations of time scales, which seeks to unify and extend discrete and continuous analysis [9]. As a result, we can generalize a process to account for both cases, or any combination of the two provided we restrict ourselves to closed, nonempty subsets of the reals (a time scale). From a numerical viewpoint, this theory can be thought of a generalized sampling technique that allows a researcher to evaluate processes with continuous, discrete, or uneven measurements. Since its inception, this area of mathematics has gained a great deal of international attention. Researchers have since found applications of time scales to include heat transfer, population dynamics, as well as economics. For a more in depth study of time scales, it is suggested that one see Bohner and Peterson’s books [10,11].
There have been a number of researchers who have sought to combine this field with the theory of control. A number of authors have contributed to generalizing the basic notions of controllability and observability (see [12,13,14,15,16]). Bohner first provided the conditions for optimality for dynamic control processes in [17]. DaCunha unified the theory of Lyapunov and Floquet theory in his dissertation [18]. Hilscher along with Zeidan have studied optimal control for sympletic systems [19]. Additional contributions can be found in [20,21,22,23,24,25], among several others.
In this paper, we study a natural extension of the LQR and LQT previously generalized to dynamics equations on time scales (see [26,27]). Here, we consider the following separable dynamic systems
x P Δ ( t ) = A P x P ( t ) + B P u ( t ) , x P ( t 0 ) = x 0 P x E Δ ( t ) = A E x E ( t ) + B E v ( t ) , x E ( t 0 ) = x 0 E ,
where x i R n represent our states and u , v R m represent our controls. Note that the subscripts P and E to stand for the pursuer and the evader respectively. The pursuing state seeks to intercept the evading state at time t f while the latter state seeks to do the opposite. For simplicity, we make the following assumptions. First, we assume the given systems are linear-time invariant (although the strategies for the time-varying case can be determined in a similar fashion). Second, we assume that both states are controllable and are being evaluated on the same time scale. Finally, we assume our state equations are associated with the cost functional
J ( u , v ) = 1 2 | | x P x E | | M ( t f ) + 1 2 t 0 t f ( | | x P x E | | Q + | | u | | R P | | v | | R E ) ( τ ) Δ τ = 1 2 ( x P x E ) T ( t f ) M ( x P x E ) ( t f ) + 1 2 t 0 t f ( x P x E ) T Q ( x P x E ) + u T R P u v T R E v ( τ ) Δ τ ,
where M 0 and diagonal, Q 0 and R P , R E > 0 . Note that the goal of the pursuing state is to minimize (1.2) while the evading state seeks to maximize it. Since these states represent opposing players, evaluating this cost can be thought of as a minimax problem.
The pursuit-evasion framework remains an active area across multiple disciplines, as found in [28,29,30,31,32,33,34]. It should be noted that there have been other excursions in combining dynamic games with time scales calculus. Libich and Stehlík introduced macroeconomic policy games on times scales with inefficient equilibria in [35]. Martins and Torres considered n player games where each player sought to minimize a shared cost functional. Mozhegova and Petrov introduced a simple pursuit problem in [36] and a dynamic analogue to the “Cossacks-robbers” in [37]. Minh and Phuong have previously studied linear pursuit-evasion games on time scales in [38]. However, these results do include a regulator/saddle point framework, nor are they complete when compared to this manuscript.
The organization of this paper is as follows. Section 2 presents core definitions and concepts of the time scales calculus. We offer the variational properties needed such that an optimal strategy exists in Section 3. In Section 4, we seek a mixed strategy when the final states are both fixed. In this setting, we can rewrite our cost functional (1.2) in terms of the difference in Gramians of each system. For Section 5, we find a pair of a controls in terms of an extended state. In Section 6, we offer some examples including a numerical result. Finally, we provide some concluding remarks and future plans in Section 7.

2. Time Scales Preliminaries

Here we offer a brief introduction to the theory of dynamic equations on time scales. For a more in-depth study of time scales, see Bohner and Peterson’s books [10,11].
Definition 1.
A time scale  T is an arbitrary nonempty closed subset of the real numbers. We let T κ = T { max T } if max T exists; otherwise T κ = T .
Example 2.
The most common examples of time scales are T = R , T = Z , T = h Z for h > 0 , and T = q N 0 for q > 1 .
Definition 3.
We define the forward jump operator  σ : T T and the graininess function  μ : T [ 0 , ) by
σ ( t ) : = inf s T : s > t a n d μ ( t ) = σ ( t ) t .
Definition 4.
For any function f : T R , we define the function f σ : T R by f σ = f σ .
Next, we define the delta (or Hilger) derivative as follows.
Definition 5.
Assume f : T R and let t T κ . The delta derivative  f Δ ( t ) is the number (when it exists) such that given any ε > 0 , there is a neighborhood U of t such that
[ f ( σ ( t ) ) f ( s ) ] f Δ ( t ) [ σ ( t ) s ] ε | σ ( t ) s | for all s U .
In the next two theorems, we consider some properties of the delta derivative.
Theorem 6
(See Theorem 1.16 [10]). Suppose f : T R is a function and let t T κ . Then we have the following:
a. 
If f is differentiable at t, then f is continuous at t.
b. 
If f is continuous at t, where t is right-scattered, then f is differentiable at t and
f Δ ( t ) = f ( σ ( t ) ) f ( t ) μ ( t ) .
c. 
If f is differentiable at t, where t is right-dense, then
f Δ ( t ) = lim s t f ( t ) f ( s ) t s .
d. 
If f is differentiable at t, then
f ( σ ( t ) ) = f ( t ) + μ ( t ) f Δ ( t ) .
Note that (2.1) is sometimes called the “simple useful formula."
Example 7.
Note the following examples.
a.
When T = R , then (if the limit exists)
f Δ ( t ) = lim s t f ( t ) f ( s ) t s = f ( t ) .
b.
When T = Z , then
f Δ ( t ) = f ( t + 1 ) f ( t ) = : Δ f ( t ) .
c.
When T = h Z for h > 0 , then
f Δ ( t ) = f ( t + h ) f ( t ) h = : Δ h f ( t ) .
d.
When T = q Z for q > 1 , then
f Δ ( t ) = f ( q t ) f ( t ) ( q 1 ) t = : D q f ( t ) .
Next we consider the linearity property as well as the product rules.
Theorem 8
(See Theorem 1.20 [10]). Let f , g : T R be differentiable at t T κ . Then we have the following:
a. 
For any constants α and β, the sum ( α f + β g ) : T R is differentiable at t with
( α f + β g ) Δ ( t ) = α f Δ ( t ) + β g Δ ( t ) .
b. 
The product f g : T R is differentiable at t with
( f g ) Δ ( t ) = f Δ ( t ) g ( t ) + f σ ( t ) g Δ ( t ) = f ( t ) g Δ ( t ) + f Δ ( t ) g σ ( t ) .
Definition 9.
A function f : T R is said to be rd-continuous on T when f is continuous in points t T with σ ( t ) = t and it has finite left-sided limits in points t T with sup s T : s < t = t . The class of rd-continuous functions f : T R is denoted by C rd = C rd ( T ) = C rd ( T , R ) . The set of functions f : T R that are differentiable and whose derivative is rd-continuous is denoted by C rd 1 .
Theorem 10
(See Theorem 1.74[10]). Any rd-continuous function f : T R has an antiderivative F, i.e., F Δ = f on T κ .
Definition 11.
Let f C rd and let F be any function such that F Δ ( t ) = f ( t ) for all t T κ . Then the Cauchy integral of f is defined by
a b f ( t ) Δ t = F ( b ) F ( a ) for   all a , b T .
Example 12.
Let a , b T with a < b and assume that f C rd .
a.
When T = R , then
a b f ( t ) Δ t = a b f ( t ) d t .
b.
When T = Z , then
a b f ( t ) Δ t = t = a b 1 f ( t ) .
c.
When T = h Z for h > 0 , then
a b f ( t ) Δ t = h t = a / h b / h 1 f ( t h ) .
d.
When T = q N 0 for q > 1 , then
a b f ( t ) Δ t = a b f ( t ) d q ( t ) : = ( q 1 ) t [ a , b ) T t f ( t ) .
Next, we present the matrix exponential and some of its properties.
Definition 13.
An m × n matrix-valued function A on T is rd-continuous if each of its entries are rd-continuous. Furthermore, if m = n , A is said to be regressive (we write A R ) if
I + μ ( t ) A ( t ) is   invertible   for   all t T κ .
Theorem 14
(See Theorem 5.8 [10]). Suppose that A is regressive and rd-continuous. Then the initial value problem
X Δ ( t ) = A ( t ) X ( t ) , X ( t 0 ) = I ,
where I is the identity matrix, has a unique n × n matrix-valued solution X.
Definition 15.
The solution X from Theorem 14 is called the matrix exponential function on T and is denoted by e A ( · , t 0 ) .
Theorem 16
(See Theorem 5.21 [10]). Let A be regressive and rd-continuous. Then for r , s , t T ,
a. 
e A ( t , s ) e A ( s , r ) = e A ( t , r ) , hence e A ( t , t ) = I ,
b. 
e A ( σ ( t ) , s ) = ( I + μ ( t ) A ( t ) ) e A ( t , s ) ,
c. 
e A ( t , σ ( s ) ) = e A ( t , s ) ( I + μ ( s ) A ( s ) ) 1 ,
d. 
( e A ( · , s ) ) Δ = A e A ( · , s ) ,
e. 
( e A ( t , · ) ) Δ = e A σ ( t , · ) A ( s ) = e A ( t , · ) ( I + μ ( s ) A ( s ) ) 1 A ( s ) .
Next we give the solution (state response) to our linear system using variation of parameters.
Theorem 17
(See Theorem 5.24 [10]). Let A R be an n × n matrix-valued function on T and suppose that f : T R n is rd-continuous. Let t 0 T and x 0 R n . Then the solution of the initial value problem
x Δ ( t ) = A ( t ) x ( t ) + f ( t ) , x ( t 0 ) = x 0
is given by
x ( t ) = e A ( t , t 0 ) x 0 + t 0 t e A ( t , σ ( τ ) ) f ( τ ) Δ τ .

3. Optimization of Linear Systems on Time Scales

In this section, we make use of variational methods on time scales as introduced by Bohner in [17]. First, note that the state equations in (1.1) are uncoupled. For convenience, we rewrite (1.1) as
z Δ ( t ) = A ^ z ( t ) + B ^ u ( t ) + C ^ v ( t ) , z ( t 0 ) = z 0 ,
where z represents an extended state given by z = x P x E T , A ^ = A P 0 0 A E , B ^ = B P 0 T , and C ^ = 0 B E T . Associated with (3.1) is the quadratic cost functional
J ( u , v ) = 1 2 z T ( t f ) M ^ z ( t f ) + 1 2 t 0 t f ( z T Q ^ z + u T R P u v T R E v ) ( τ ) Δ τ ,
where M ^ , Q ^ 0 and R P , R E > 0 . To minimize (3.2), we introduce the augmented cost functional
J + ( u , v ) = 1 2 z T ( t f ) M ^ z ( t f ) + t 0 t f [ H ( x , u , v , λ σ ) ( λ σ ) T z Δ ] ( τ ) Δ τ ,
where the so-called Hamiltonian H is given by
H ( x , u , v , λ ) = 1 2 ( z T Q ^ z + u T R P u v T R E v ) + λ T ( A ^ z + B ^ u + C ^ v )
and λ = λ P λ E T represents a multiplier to be determined later.
Remark 18.
Our treatment of (1.1) differs from the argument used by Ho, Bryson, and Baron in [7]. In their paper, they appealed to state estimates of the pursuer and evader to evaluate the cost functional. Their motivation for their argument is due to notion that when they studied pursing and evading missiles, they considered difference in altitude as negligible. As a result of our rewriting of (1.1), we are not required to make such a restriction.
Next, we provide necessary conditions for an optimal control. We assume that
d d ε t 0 t f f ( τ , ε ) Δ τ = t 0 t f ε f ( τ , ε ) Δ τ
for all f : T × R R such that f ( · , ε ) , f ( · , ε ) / ε C rd ( T ) .
Lemma 19.
Let (3.2) be the cost functional associated with (3.1). Assume (3.4) holds. Then the first variation, Φ ˙ ( 0 ) , is zero provided that z, λ, u, and v satisfy
z Δ = A ^ z + B ^ u + C ^ v ,
λ Δ = Q ^ z + A ^ T λ σ ,
0 = R P u + B ^ T λ σ ,
0 = R E v + C ^ T λ σ .
Proof. 
First note that
Φ ( ε ) = J ( ( z , u , v , λ ) + ε ( η 1 , η 2 , η 3 , η 4 ) ) = 1 2 ( z + ε η 1 ) T ( t f ) M ^ ( z + ε η 1 ) ( t f ) + 1 2 t 0 t f [ ( z + ε η 1 ) T Q ^ ( z + ε η 1 ) ] ( τ ) Δ τ + 1 2 t 0 t f [ ( u + ε η 2 ) T R P ( u + ε η 2 ) ] ( τ ) Δ τ 1 2 t 0 t f [ ( v + ε η 3 ) T R E ( v + ε η 3 ) ] ( τ ) Δ τ + t 0 t f ( λ σ + ε η 4 σ ) T [ A ^ ( z + ε η 1 ) + B ^ ( u + ε η 2 ) ( τ ) Δ τ + t 0 t f ( λ σ + ε η 4 σ ) T [ C ^ ( v + ε η 3 ) ( z + ε η 1 ) Δ ] ( τ ) Δ τ .
Then
Φ ˙ ( ε ) = η 1 T ( t f ) M ^ ( z + ε η 1 ) ( t f ) + t 0 t f [ η 1 T Q ^ ( z + ε η 1 ) ] ( τ ) Δ τ + t 0 t f [ η 2 T R P ( u + ε η 2 ) η 3 T R E ( v + ε η 3 ) ] ( τ ) Δ τ + t 0 t f ( η 4 σ ) T [ A ^ ( z + ε η 1 ) + B ^ ( u + ε η 2 ) ] ( τ ) Δ τ + t 0 t f ( η 4 σ ) T [ C ^ ( v + ε η 3 ) ( z + ε η 1 ) Δ ] ( τ ) Δ τ + t 0 t f ( λ σ + ε η 4 σ ) T [ A ^ η 1 + B ^ η 2 + C ^ η 3 η 1 Δ ] ( τ ) Δ τ .
Then after rearranging terms, the first variation can be written as
Φ ˙ ( 0 ) = [ M ^ z ( t f ) λ ( t f ) ] T η 1 ( t f ) + λ T ( t 0 ) η 1 ( t 0 ) + t 0 t f [ ( A ^ T λ σ + Q ^ z + λ Δ ) T η 1 + ( R P u + B ^ T λ σ ) T η 2 ] ( τ ) Δ τ + t 0 t f [ ( R E v + C ^ T λ σ ) T η 3 + ( A ^ z + B ^ u + C ^ v z Δ ) T η 4 σ ] ( τ ) Δ τ .
Now in order for Φ ˙ ( 0 ) = 0 , we set each coefficient of independent increments η 1 , η 2 , η 3 , η 4 σ equal to zero. This yields the necessary conditions for a minimum of (3.2). Using the Hamiltonian (3.3), we have state and costate equations
z Δ = H λ ( z , u , v , λ σ ) = A ^ z + B ^ u + C ^ v
and
λ Δ = H z ( z , u , v , λ σ ) = Q ^ z + A ^ T λ σ .
Similarly, we have the stationary conditions
0 = H u ( z , u , v , λ σ ) = R P u + B ^ T λ σ
and
0 = H v ( z , u , v , λ σ ) = R E v + C ^ T λ σ
This concludes the proof. □
Remark 20.
We note that z, λ , u, and v solve (3.5) if and only if they solve
z Δ = A ^ z D ^ λ σ ,
λ Δ = Q ^ z + A ^ T λ σ ,
u = R P 1 B ^ T λ σ ,
v = R E 1 C ^ T λ σ ,
where D ^ is a “mixing term” given by
D ^ : = B ^ R P 1 B ^ T C ^ R E 1 C ^ T .
Throughout this paper, we assume that D ^ is regressive. As a result, we can determine an optimal strategy if we know the value of the costate.
Finally, we give the sufficient conditions for a local optimal control.
Lemma 21.
Let (3.2) be the cost functional associated with (3.1). Assume (3.4) holds. Then the second variation, Φ ¨ ( 0 ) , is positive provided that η 1 , η 2 , and η 3 satisfy the constraints η 1 Δ = A ^ η 1 + B ^ η 2 + C ^ η 3 where η 2 0 and η 3 is fixed.
Proof. 
Taking the second derivative of Φ ( ε ) , we have
Φ ¨ ( ε ) = η 1 T ( t f ) M ^ η 1 ( t f ) + t 0 t f [ η 1 T Q ^ η 1 + η 2 T R P η 2 η 3 T R E η 3 ] ( τ ) Δ τ + 2 t 0 t f [ ( A ^ η 1 + B ^ η 2 + C ^ η 3 η 1 Δ ) T η 4 σ ] ( τ ) Δ τ .
If we assume that η 1 , η 2 , and η 3 satisfy the constraint
η 1 Δ = A ^ η 1 + B ^ η 2 + C ^ η 3 ,
then the second variation is given by
Φ ¨ ( 0 ) = η 1 T ( t f ) M ^ η 1 ( t f ) + t 0 t f [ η 1 T Q ^ η 1 + η 2 T R P η 2 η 3 T R E η 3 ] ( τ ) Δ τ .
Note that M ^ and Q ^ 0 while R P and R E > 0 . Thus if η 2 0 and η 3 is fixed, then (3.7) is guaranteed to be positive. □
Definition 22.
The pair ( u * , v * ) is a saddle point to the system (3.1) associated with the cost (3.2) provided
J ( u , v * ) J ( u * , v * ) J ( u * , v ) .
Here, the stationary conditions needed to ensure a saddle are H u u = R P > 0 and H v v = R E < 0 (see [39]). For our purposes, this pair corresponds to when neither player wishes to deviate from this compromise without being penalized by the other player. It be understood that this compromise occurs when we have the natural caveat that the pursuer and evader belong to the same time scale. In this paper, we do not claim that this saddle point must be unique.

4. Fixed Final States Case

In this section, we seek an optimal strategy when the final states are fixed. In this setting we write the equations for the pursuer and evader separately. Here we consider the state and costate equations for the pursuer
x P Δ ( t ) = A P x P ( t ) B P R P 1 B P T λ P σ ( t ) , x P ( t 0 ) = x 0 P λ P Δ ( t ) = A P T λ P σ ( t ) , λ P ( t f ) = M ( x P x E ) ( t f )
as well as those for the evader
x E Δ ( t ) = A E x E ( t ) B E R E 1 B E T λ E σ ( t ) , x E ( t 0 ) = x 0 E λ E Δ ( t ) = A E T λ E σ ( t ) , λ E ( t f ) = M ( x E x P ) ( t f )
associated with the cost functional
J ( u , v ) = 1 2 ( x P x E ) T ( t f ) M ( x P x E ) ( t f ) + 1 2 t 0 t f ( x P x E ) T Q ( x P x E ) + u T R P u v T R E v ( τ ) Δ τ .
Definition 23.
The initial state difference, d 0 ( · ) , is the difference between the zero-input pursuing and evading states, i.e.,
d 0 ( t ) : = e A P ( t , t 0 ) x P ( t 0 ) e A E ( t , t 0 ) x E ( t 0 ) .
Next, we determine an open–loop strategy for both players. Note that the following theorem mirrors Kalman’s generalized controllability criterion as found in Theorem 3.2 [16].
Theorem 24.
Suppose that x P and λ P solve (4.1) while x E and λ E satisfy (4.2). Let the Gramians for the pursuer and evader
G P ( t o , t f ) : = t 0 t f e A P ( t f , σ ( τ ) ) B P R P 1 B P T e A P T ( t f , σ ( τ ) ) Δ τ
and
G E ( t o , t f ) : = t 0 t f e A E ( t f , σ ( τ ) ) B E R E 1 B E T e A E T ( t f , σ ( τ ) ) Δ τ ,
respectively, be such that I + ( G P G E ) ( t 0 , t f ) M is invertible for all t [ t 0 , t f ] T . Then u and v can be rewritten as
u ( t ) = R P 1 B P T e A P T ( t f , σ ( t ) ) M [ I + ( G P G E ) ( t o , t f ) M ] 1 d 0 ( t f )
and
v ( t ) = R E 1 B E T e A E T ( t f , σ ( t ) ) M [ I + ( G P G E ) ( t o , t f ) M ] 1 d 0 ( t f ) .
Proof. 
Solving (4.1) for λ P , we have
λ P ( t ) = e A P T ( t f , t ) λ P ( t f ) = e A P T ( t f , t ) M ( x P x E ) ( t f ) .
Using (2.1) and (3.5a), the state equation becomes
x P Δ ( t ) = A P x P ( t ) B P R P 1 B P T e A P T ( t f , σ ( t ) ) λ P ( t f ) .
Now solving (4.8) with Theorem 17 at time t = t f , we have
x P ( t f ) = e A P ( t f , t 0 ) x P ( t 0 ) t 0 t f e A P ( t f , σ ( τ ) ) B P R P 1 B P T e A P T ( t f , σ ( τ ) ) λ P ( t f ) Δ τ = e A P ( t f , t 0 ) x P ( t 0 ) G P ( t 0 , t f ) M ( x P x E ) ( t f ) .
Similarly, the final state for the pursuer can be written as
x E ( t f ) = e A E ( t f , t 0 ) x E ( t 0 ) G E ( t o , t f ) M ( x P x E ) ( t f ) .
Taking the difference in the final states and rearranging, we have
( x P x E ) ( t f ) = d 0 ( t f ) ( G P G E ) ( t 0 , t f ) M ( x P x E ) ( t f ) = [ I + ( G P G E ) ( t 0 , t f ) M ] 1 d 0 ( t f ) .
Finally, plugging λ into (3.6c) and using (4.9) yields
u ( t ) = R P 1 B P T e A P T ( t f , σ ( t ) ) λ P ( t f ) = R P 1 B P T e A P T ( t f , σ ( t ) ) M ( x P x E ) ( t f ) = R P 1 B P T e A P T ( t f , σ ( t ) ) M [ I + ( G P G E ) ( t o , t f ) M ] 1 d 0 ( t f ) .
The equation for v can be shown similarly. This concludes the proof. □
Next, we determine the optimal cost.
Theorem 25.
If u and v are given by (4.6) and (4.7), respectively, then the cost functional (4.3) can be rewritten as
J ( u , v ) = 1 2 d 0 T ( t f ) H ( t 0 , t f ) M T [ I + ( G P G E ) ( t 0 , t f ) ] M H ( t 0 , t f ) d 0 ( t f ) ,
where H ( t 0 , t f ) : = [ I + ( G P G E ) ( t 0 , t f ) M ] 1 .
Proof. 
First, plugging (4.6), (4.7), and (4.9) into (4.3), we have
J ( u , v ) = 1 2 d 0 T ( t f ) H ( t 0 , t f ) M H ( t 0 , t f ) d 0 ( t f ) + 1 2 d 0 T ( t f ) H ( t 0 , t f ) M T t 0 t f e A P ( t f , σ ( τ ) ) B P R P 1 B P T e A P T ( t f , σ ( τ ) ) Δ τ M H ( t 0 , t f ) d 0 ( t f ) 1 2 d 0 T ( t f ) H ( t 0 , t f ) M T t 0 t f e A E ( t f , σ ( τ ) ) B E R E 1 B E T e A E T ( t f , σ ( τ ) ) Δ τ M H ( t 0 , t f ) d 0 ( t f ) = 1 2 d 0 T ( t f ) H ( t 0 , t f ) M H ( t 0 , t f ) d 0 ( t f ) + 1 2 d 0 T ( t f ) H ( t 0 , t f ) M T ( G P G E ) ( t 0 , t f ) ] M H ( t 0 , t f ) d 0 ( t f ) ,
using the gramians (4.4) and (4.5). Since M 0 is symmetric, we can pull out common factors on the left and right to obtain our result. □
Remark 26.
Suppose that the pursuer wants to use a strategy u that intercepts the evader (using strategy v) with minimal energy. Note that det [ I + ( G P G E ) ( t 0 , t f ) ] 0 if and only if det [ ( G P G E ) ( t 0 , t f ) ] 0 . From the classical definition of controllability, this implies that the pursuer captures the evader when the pursuer is “more controllable" than the evader. A sufficient condition for the pursuing state to intercept the evader is given by ( G P G E ) ( t 0 , t f ) > 0 . As a result, this relationship is preserved in the unification of pursuit-evasion to dynamic equations on time scales.

5. Free Final States Case

In this section, we develop an optimal control law in the form of state feedback. In considering the boundary conditions, note that z ( t 0 ) is known (meaning η 1 ( t 0 ) = 0 ) while z ( t f ) is free (meaning η 1 ( t f ) 0 ). Thus the coefficient on η 1 ( t f ) must be zero. This gives the terminal condition on the costate to be
λ ( t f ) = M ^ z ( t f ) .
Remark 27.
Now in order to solve this two-point boundary value problem, we make the assumption that z and λ satisfy
λ ( t ) = S ( t ) z ( t ) .
for all t [ t 0 , t f ] . This condition (5.2) is called a “sweep condition," a term used by Bryson and Ho in [8]. Since the terminal condition M ^ 0 , it is natural to assume that S 0 as well.
Theorem 28.
Assume that S solves
S Δ = Q ^ + A ^ T S σ + ( I + μ A ^ T ) S σ ( I + μ D ^ S σ ) 1 ( A ^ D ^ S σ ) .
If x satisfies
z Δ = ( I + μ D ^ S σ ) 1 ( A D ^ S σ ) z
and λ is given by (5.2), then
λ Δ = Q ^ z + A ^ T λ σ .
Proof. 
Since λ is as given in (5.2), we may use the product rule, (5.3), (5.4), and (2.1) to arrive at
λ Δ = S Δ z S σ z Δ = Q ^ z + A ^ T S σ z + ( I + μ A ^ T ) S σ z S Δ z = Q ^ z + A ^ T S σ z + μ A ^ T S σ z Δ = Q ^ z + A ^ T S σ z σ = Q ^ z + A ^ T λ σ ,
which gives (5.5) as desired. □
Next we offer an alternative form of our Riccati equation.
Lemma 29.
If D ^ S σ is regressive, then S solves (5.3) if and only if it solves
S Δ = Q ^ + A ^ T S σ + ( I + μ A ^ T ) S σ A ^ ( I + μ A ^ T ) S σ D ^ S σ ( I + μ D ^ S σ ) 1 ( I + μ A ^ ) .
Proof. 
Note that
A D ^ S σ = A D ^ S σ ( I + μ A μ A ) = ( I + μ D ^ S σ ) A D ^ S σ ( I + μ A ) .
Plugging the above identity into (5.3) yields (5.6). □
Next we define our Kalman gains as follows.
Definition 30.
Let D ^ S σ be regressive. Then the matrix-valued functions
K P ( t ) = R P 1 B ^ T S σ ( t ) ( I + μ ( t ) D ^ S σ ( t ) ) 1 ( I + μ ( t ) A ^ )
and
K E ( t ) = R E 1 C ^ T S σ ( t ) ( I + μ ( t ) D ^ S σ ( t ) ) 1 ( I + μ ( t ) A ^ )
are called the pursuer feedback gain and evader feedback gain, respectively.
Theorem 31.
Let D ^ S σ be regressive and suppose that z and λ solve (4.8) such that (5.2) holds. Then
B ^ u + C ^ v = K P z + K E z .
Proof. 
Using (3.6), (5.2), and (2.1), we have
B ^ u + C ^ v = B ^ R P 1 B ^ T λ σ + C ^ R E 1 C ^ T λ σ = B ^ R P 1 B ^ T S σ z + μ z Δ + C ^ R E 1 C ^ T S σ z + μ z Δ = D ^ S σ [ ( I + μ A ^ ) z + μ ( B ^ u + C ^ v ) ] .
Now combining like terms yields
( I + D ^ S σ ) ( B ^ u + C ^ v ) = D ^ S σ ( I + μ A ^ ) z
Multiplying both side by the inverse of I + D ^ S σ and rearranging terms, we have
B ^ u + C ^ v = ( I + D ^ S σ ) 1 D ^ S σ ( I + μ A ^ ) z = D ^ S σ ( I + D ^ S σ ) 1 ( I + μ A ^ ) z = B ^ R P 1 B ^ T S σ ( I + D ^ S σ ) 1 ( I + μ A ^ ) z + C ^ R E 1 C ^ T S σ ( I + D ^ S σ ) 1 ( I + μ A ^ ) z .
Finally, (5.9) follows using (5.7) and (5.8). □
Next we rewrite our extended state equation under the influence of the pursuit-evasion control laws. This yields the closed-loop plant given by
z Δ ( t ) = ( A ^ B ^ K P ( t ) + C ^ K E ( t ) ) z ( t ) ,
which can be used to find an optimal trajectory for any given z ( t 0 ) .
Lemma 32.
If D ^ S σ is regressive and S is symmetric, then
( I + μ A ^ T ) S σ A ^ ( I + μ A ^ T ) S σ ( I + μ D ^ S σ ) 1 D ^ S σ ( I + μ A ^ ) = ( I + μ ( A ^ B ^ K P + C ^ K E ) T ) S σ ( A ^ B ^ K P + C ^ K E ) K P T B ^ T S σ + K E T C ^ T S σ + K P T R P K P K E T R E K E .
Moreover, both sides of (5.11) are equal to ( I + μ A ^ T ) S σ ( A ^ B ^ K P + C ^ K E ) .
Proof. 
We can use (5.7) and (5.8) to rewrite the left-hand side of (5.11) as
( I + μ A ^ T ) S σ A ^ ( I + μ A ^ T ) S σ D ^ S σ ( I + μ D ^ S σ ) 1 ( I + μ A ^ ) = ( I + μ A ^ T ) S σ A ^ ( I + μ A ^ T ) S σ ( B ^ K P C ^ K E ) = ( I + μ A ^ T ) S σ ( A ^ B ^ K P + C ^ K E ) .
Using (5.7) and (5.8), the right-hand side of (5.11) can be written as
( I + μ A ^ T ) S σ ( A ^ B ^ K P + C ^ K E ) K P T B ^ T S σ ( I + μ A ^ ) + K E T C ^ T S σ ( I + μ A ^ ) μ K P T B ^ T S σ ( B ^ K P + C ^ K E ) + μ K E T C ^ T S σ ( B ^ K P + C ^ K E ) + K P T R P K P K E T R E K E = ( I + μ A ^ T ) S σ ( A ^ B ^ K P + C ^ K E ) K P T B ^ T S σ ( I + μ A ^ ) + K E T C ^ T S σ ( I + μ A ^ ) μ K P T B ^ T S σ D ^ S σ ( I + μ D ^ S σ ) 1 ( I + μ A ^ ) + μ K E T C ^ T S σ D ^ S σ ( I + μ D ^ S σ ) 1 ( I + μ A ^ ) + K P T R P K P K E T R E K E = ( I + μ A ^ T ) S σ ( A ^ B ^ K P + C ^ K E ) K P T B ^ T S σ ( I + μ D ^ S σ ) 1 ( I + μ A ^ ) + K E T C ^ T S σ ( I + μ D ^ S σ ) 1 ( I + μ A ^ ) + K P T R P K P K E T R E K E = ( I + μ A ^ T ) S σ ( A ^ B ^ K P + C ^ K E ) .
Thus, (5.11) holds. □
Now we rewrite the Riccati equation (5.6) in so-called (generalized) Joseph stabilized form (see [39]).
Theorem 33.
If D ^ S σ is regressive and S is symmetric, then S solves the Riccati equation (5.6) if and only if it solves
S Δ = Q ^ + ( A ^ B ^ K P + C ^ K E ) T S σ + ( I + μ ( A ^ B ^ K P + C ^ K E ) T ) S σ ( A ^ B ^ K P + C ^ K E ) + K P T R P K P K E T R E K E .
Proof. 
The statement follows directly from Lemma 32. □
Finally, we rewrite the cost.
Theorem 34.
Suppose that S solves (5.12). If z, u, and v satisfy (5.10), and (5.9) respectively, then the cost functional (3.2) can be rewritten as
J ( u , v ) = 1 2 z T ( t 0 ) S ( t 0 ) z ( t 0 ) .
Proof. 
First note that we may use the product rule, (2.1), and (5.10) to find
( z T S z ) Δ = ( z T S ) Δ z + ( z T S ) σ z Δ = ( z Δ ) T S σ z + z T S Δ z + ( z + μ z Δ ) T S σ z Δ = z T [ ( A ^ B ^ K P + C ^ K E ) T S σ + S Δ ] z + z T [ I + μ ( A ^ B ^ K P + C ^ K E ) ] T S σ ( A ^ B ^ K P + C ^ K E ) z .
Using this and (5.9) in (3.2), we have
J ( u , v ) = 1 2 z T ( t 0 ) S ( t 0 ) z ( t 0 ) + 1 2 t 0 t f ( z T S z ) Δ ( τ ) Δ τ + 1 2 t 0 t f [ z T Q ^ z + u T R P u v T R E v ] ( τ ) Δ τ = 1 2 z T ( t 0 ) S ( t 0 ) z ( t 0 ) + 1 2 t 0 t f ( z T S z ) Δ ( τ ) Δ τ + 1 2 t 0 t f z T [ Q ^ + K P T R P K P K E T R E K E ] z ( τ ) Δ τ .
Using (5.14) and (5.12), the cost functional can be rewritten as
J ( u , v ) = 1 2 z T ( t 0 ) S ( t 0 ) z ( t 0 ) .
This concludes the proof. □
From Theorem 34, if the current state and S are known, we can determine the optimal cost before we apply the optimal control or even calculate it. The table below summarizes our results.
Table 5.1. The LQPEG on T
Table 5.1. The LQPEG on T
System:
z Δ = A ^ z + B ^ u + C ^ v
Cost:
J ( u , v ) = 1 2 z T ( t f ) M ^ z ( t f ) + 1 2 t 0 t f ( z T Q ^ z + u T R P u v T R E v ) ( τ ) Δ τ
Mixing Term:
D ^ = B ^ R P 1 B ^ T C ^ R E 1 C ^ T
Pursuer Feedback:
K P = R P 1 B ^ T S σ ( I + μ D ^ S σ ) 1 ( I + μ A ^ )
Evader Feedback:
K E = R E 1 C ^ T S σ ( I + μ D ^ S σ ) 1 ( I + μ A ^ )
Riccati Equation:
S Δ = Q + A ^ T S σ + ( I + μ A ^ T ) S σ A ^ ( I + μ A ^ T ) S σ D ^ S σ ( I + μ D ^ S σ ) 1 ( I + μ A ^ )

6. Examples

Example 35. (The Continuous LQPEG) Let T = R and consider
z ( t ) = A ^ z ( t ) + B ^ u ( t ) + C ^ v ( t ) ,
associated with the cost functional
J ( u , v ) = 1 2 z T ( t f ) M ^ z ( t f ) + 1 2 t 0 t f [ z T Q ^ z + u T R P u v T R E v ] ( τ ) d τ
(observe part (a) of Examples 7 and 12). Then the state, costate, and stationary equations (3.6) are given by
z = A ^ z D ^ λ , λ = A ^ T λ + Q ^ z , u = R P 1 B ^ T λ , v = R E 1 C ^ T λ .
In this case, our pursuer-evader feedback gains (5.7) and (5.8) are given as
K P ( t ) = R P 1 B ^ T S ( t ) a n d K E ( t ) = R E 1 C ^ T S ( t ) .
Now the pursuer-evader law (5.9) and the closed-loop plant (5.10) can be written as
B ^ u ( t ) + C ^ v ( t ) = B ^ K P ( t ) z ( t ) + C ^ K E ( t ) z ( t )
and
z = ( A ^ B ^ K P + C ^ K E ) z .
Similarly, the closed-loop Riccati equation (5.12) can be written as
S = Q ^ + S ( A ^ B ^ K P + C ^ K E ) + ( A ^ B ^ K P + C ^ K E ) T S + K P T R P K P K E T R E K E
while the optimal cost is given by (5.13).
Example 36. (The h-difference LQPEG) Let T = h Z and consider
Δ h z ( t ) = A ^ z ( t ) + B ^ u ( t ) + C ^ v ( t ) ,
By observing Example 7 (c) and introducing
A ˜ = I + h A ^ , B ˜ = h B ^ , C ˜ = h C ^ , Q ˜ = h Q ^ , R ˜ i = h R i , D ˜ = h D ^ ,
we can rewrite the system as
z ( t + h ) = A ˜ z ( t ) + B ˜ u ( t ) + C ˜ v ( t ) ,
and the associated cost functional takes the form (observe Example 12 (c))
J ( u , v ) = 1 2 z T ( t f ) M ^ z ( t f ) + 1 2 τ = t 0 / h t f / h 1 [ z T Q ˜ z + u T R ˜ P u v T R ˜ E v ] ( τ h ) .
Then the state, costate, and stationary equations (3.6) are given by
z ( t + h ) = A ˜ z ( t ) D ˜ λ ( t + h ) , λ ( t ) = A ˜ T λ ( t + h ) + Q ˜ z ( t ) , u ( t ) = R ˜ P 1 B ˜ T λ ( t + h ) , v ( t ) = R ˜ E 1 C ˜ T λ ( t + h ) .
Now our pursuer and evader feedback gains (5.7) and (5.8) are
K P ( t ) = R ˜ P 1 B ˜ T S ( t + h ) ( I + D ˜ S ( t + h ) ) 1 A ˜
and
K E ( t ) = R ˜ E 1 C ˜ T S ( t + h ) ( I + D ˜ S ( t + h ) ) 1 A ˜ .
Next, the control-tracker law (5.9) and the closed-loop plant (5.10) can be written as
B ˜ u ( t ) + C ˜ v ( t ) = B ˜ K P ( t ) z ( t ) + C ˜ K E ( t ) z ( t )
and
z ( t + h ) = ( A ˜ B ˜ K P ( t ) + C ˜ K E ( t ) ) z ( t ) ,
respectively. Similarly, the closed-loop Riccati equation (5.12) can be written as
S ( t ) = Q ˜ + ( A ˜ B ˜ K P ( t ) + C ˜ K E ( t ) ) T S ( t + h ) ( A ˜ B ˜ K P ( t ) + C ˜ K E ( t ) ) + K P T ( t ) R ˜ P K P ( t ) K E T ( t ) R ˜ E K E ( t )
while the optimal cost is given by (5.13).
Example 37. (The q-difference LQPEG) Let T = q N 0 with q > 1 and consider
D q z ( t ) = A ^ z ( t ) + B ^ u ( t ) + C ^ v ( t ) .
By observing Example 7 (d) and introducing
A ˜ ( t ) = I + ( q 1 ) t A ^ , B ˜ ( t ) = ( q 1 ) t B ^ , C ˜ ( t ) = ( q 1 ) t C ^ Q ˜ ( t ) = ( q 1 ) t Q ^ , R ˜ i ( t ) = ( q 1 ) t R i , D ˜ ( t ) = ( q 1 ) t D ^ ,
we can rewrite the system as
z ( q t ) = A ˜ ( t ) z ( t ) + B ˜ ( t ) u ( t ) + C ˜ v ( t ) ,
while the associated cost functional becomes (observe Example 12 (d))
J ( u , v ) = 1 2 z T ( t f ) M ^ z ( t f ) + 1 2 τ [ t 0 , t f ) T [ z T Q ˜ z + u T R ˜ P u v T R ˜ E v ] ( τ ) .
Then the state, costate, and stationary equations (3.6) are given by
z ( q t ) = A ˜ ( t ) z ( t ) D ˜ ( t ) λ ( q t ) , λ ( t ) = A ˜ T ( t ) λ ( q t ) + Q ˜ ( t ) z ( t ) , u ( t ) = R ˜ P 1 ( t ) B ˜ T ( t ) λ ( q t ) , v ( t ) = R ˜ E 1 ( t ) C ˜ T ( t ) λ ( q t ) .
In this case, our pursuer and evader feedback gains (5.7) and (5.8) are
K P ( t ) = R ˜ P 1 ( t ) B ˜ T ( t ) S ( q t ) ( I + D ˜ S ( q t ) ) 1 A ˜ ( t )
and
K E ( t ) = R ˜ E 1 ( t ) C ˜ T ( t ) S ( q t ) ( I + D ˜ S ( q t ) ) 1 A ˜ ( t ) .
Now the control-tracker law (5.9) and the closed-loop plant (5.10) can be written as
B ˜ ( t ) u ( t ) + C ˜ ( t ) v ( t ) = B ˜ ( t ) K P ( t ) z ( t ) + C ˜ ( t ) K E ( t ) z ( t )
and
z ( q t ) = ( A ˜ ( t ) B ˜ ( t ) K P ( t ) + C ˜ ( t ) K E ( t ) ) z ( t ) ,
respectively. Finally, the closed-loop Riccati equation (5.12) can be written as
S ( t ) = Q ˜ ( t ) + K P T ( t ) R ˜ P ( t ) K P ( t ) K E T ( t ) R ˜ E ( t ) K E ( t ) + ( A ˜ ( t ) B ˜ ( t ) K P ( t ) + C ˜ ( t ) K E ( t ) ) T S ( q t ) ( A ˜ ( t ) B ˜ ( t ) K P ( t ) + C ˜ ( t ) K E ( t ) )
while the optimal cost is given by (5.13).
Example 38.
In this last example, we provide a numerical of the LQPEG. In this setting, we sample a two-dimensional pursuer and evader on the same discrete, but uneven time scale
T = { 0 , 0.03 , 0.29 , 1.23 , 1.49 , 1.94 , 2.11 , 2.51 , 2.77 , 3.78 , 3.87 , 4.15 , 4.78 , 4.81 , 4.89 , 4.91 , 5.49 , 5.62 , 5.71 , 6.15 , 6.72 , 7.2 , 7.4 , 7.48 , 7.59 , 7.66 , 7.68 , 8.37 , 8.55 , 8.87 , 8.96 , 9.4 , 9.44 , 9.73 , 10 } .
Next, we consider the theoretical linear dynamic system
x P Δ ( t ) = 2 0 0 1 x P ( t ) + 1 3 u ( t ) , x P ( 0 ) = 2 1 x E Δ ( t ) = 3 1 1 1 x E ( t ) + 2 2 v ( t ) , x E ( 0 ) = 1 2 .
Note that the first component of each player represents its position while the second corresponds to its velocity. For simplicity, only the position is observed. Here, we set the weights in (3.2) to be R P = 1 , R E = 1.3 and Q = S ( t f ) = I 4 . The plots for the pursuer and evader’s positions are given in Figure 6.1 below.

7. Concluding Remarks and Future Work

In this project, we have established the LQPEG where the pursuer and evader belong to the same arbitrary time scale T . One potential application of this work is when the evader represented by a drone and the evader represents a missile guidance where their corresponding signals are unevenly sampled. Here. the cost in part represents the wear and tear on the drone. A saddle point in this setting would represent a “live and let live” arrangement, where the drone is allowed to spy briefly on the missile-guidance system and return home, but is not given opportunity to preserve enough of its battery to outstay its welcome. Similarly, in finance the pursuer and evader can represents competing companies where a saddle point would correspond to an effort to coexist, where a hostile takeover or unnecessarily expended resources can be avoided. We have sidestepped the setting where the pursuer and evader each belong to their own time scale T P and T E , respectively. However, these time scales can be merged using a sample-and-hold method as found in [40,41].
One potential extension of this work is the introduction of additional pursuers. In this setting, the cost must be adjusted to account for the closest pursuer, which can vary over the time scale. A second potential extension is to consider the setting when one player is subject to a delay. Here, both players can still belong to the same time scale. However, this allows one player to act after the other, perhaps with some knowledge of the opposing player’s strategy. Finally, a third possible approach is to such games in a stochastic setting. Here, we can discretize each player’s stochastic linear time-invariant system to a dynamic system on an isolated time scale, as found in [40,42]. However, the usual separability property is not preserved in this setting.

Author Contributions

D. Funk and R Williams contributed to the analysis and writing/editing of the manuscript as well as the numerical example. N. Wintz contributed to the project conceptualization/analysis, writing/editing, and the funding of the project. All authors have read and agreed to the published version of the manuscript.

Funding

This project was partially supported by the National Science Foundation grant DMS-2150226, the NASA West Virginia Space Grant Consortium, training grant #80NSSC20M0055, and the NASA Missouri Space Grant Consortium grant #80NSSC20M0100.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank Matthias Baur and Tom Cuchta for the use of their time scales Python package in producing the last example.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LQR linear quadratic regulator
LQT linear quadratic tracker
LQPEG linear quadratic pursuit-evasion games

References

  1. Isaacs, R. Differential Games I, II, III, IV; The RAND Corp. Memo. RM-1391, RM-1399, RM-1411, RM-1486, 1954-1956.
  2. Isaacs, R. Differential games. A mathematical theory with applications to warfare and pursuit, control and optimization; John Wiley & Sons Inc.: New York, 1965; pp. xvii+384. [CrossRef]
  3. Kalman, R.E. Contributions to the theory of optimal control. Bol. Soc. Mat. Mexicana (2) 1960, 5, 102–119.
  4. Kalman, R.E. When is a linear control system optimal? Trans. ASME Ser. D. J. Basic Engineering 1964, 86, 81–90.
  5. Kalman, R.E.; Koepcke, R.W. Optimal synthesis of linear sampling control systems using generalized performance indexes. Trans. ASME Ser. D. J. Basic Engineering 1958, 80, 1820–1826. [CrossRef]
  6. Kalman, R.E. The theory of optimal control and the calculus of variations. In Mathematical optimization techniques; Univ. California Press: Berkeley, Calif., 1963; pp. 309–331.
  7. Ho, Y.C.; Bryson, Jr., A.E.; Baron, S. Differential games and optimal pursuit-evasion strategies. IEEE Trans. Automatic Control 1965, AC-10, 385–389. [CrossRef]
  8. Bryson, Jr., A.E.; Ho, Y.C. Applied optimal control; Hemisphere Publishing Corp. Washington, D. C., 1975; pp. xiv+481. Optimization, estimation, and control, Revised printing.
  9. Hilger, S. Ein Maßkettenkalkül mit Anwendung auf Zentrumsmannigfaltigkeiten; PhD Thesis, Universität Würzburg, 1988.
  10. Bohner, M.; Peterson, A. Dynamic equations on time scales; Birkhäuser Boston Inc.: Boston, MA, 2001; pp. x+358.
  11. Bohner, M.; Peterson, A., Eds. Advances in dynamic equations on time scales; Birkhäuser Boston Inc.: Boston, MA, 2003; pp. xii+348.
  12. Bartosiewicz, Z.; Pawłuszewicz, E. Linear control systems on time scales: unification of continuous and discrete. Proceedings of the 10th IEEE International Conference on Methods and Models in Automation and Robotics MMAR’04 2004, pp. 263–266.
  13. Bartosiewicz, Z.; Pawłuszewicz, E. Realizations of linear control systems on time scales. Control Cybernet. 2006, 35, 769–786. [CrossRef]
  14. Davis, J.; Gravagne, I.; Jackson, B.; Marks II, R. Controllability, observability, realizability, and stability of dynamic linear systems. Electron. J. Diff. Eqns 2009, 2009, 1–32.
  15. Fausett, L.V.; Murty, K.N. Controllability, observability and realizability criteria on time scale dynamical systems. Nonlinear Stud. 2004, 11, 627–638.
  16. Bohner, M.; Wintz, N. Controllability and observability of time-invariant linear dynamic systems. Mathematica Bohemica 2012, 137, 149–163. [CrossRef]
  17. Bohner, M. Calculus of variations on time scales. Dynam. Systems Appl. 2004, 13, 339–349. [CrossRef]
  18. DaCunha, J.J. Lyapunov Stability and Floquet Theory for Nonautonomous Linear Dynamic Systems on Time Scales; PhD Thesis, Baylor University, 2004.
  19. Hilscher, R.; Zeidan, V. Weak maximum principle and accessory problem for control problems on time scales. Nonlinear Anal. 2009, 70, 3209–3226. [CrossRef]
  20. Bettiol, P.; Bourdin, L. Pontryagin maximum principle for state constrained optimal sampled-data control problems on time scales. ESAIM: Control, Optimisation and Calculus of Variations 2021, 27. [CrossRef]
  21. Zhu, Y.; Jia, G. Linear Feedback of Mean-Field Stochastic Linear Quadratic Optimal Control Problems on Time Scales. Mathematical Problems in Engineering 2020, 2020, 8051918. [CrossRef]
  22. Ren, Q.Y.; Sun, J.P.; et al. Optimality necessary conditions for an optimal control problem on time scales. AIMS Math. 2021, 6, 5639–5646. [CrossRef]
  23. Zhu, Y.; Jia, G. Stochastic Linear Quadratic Control Problem on Time Scales. Discrete Dynamics in Nature and Society 2021, 2021, 5743014. [CrossRef]
  24. Poulsen, D.; Defoort, M.; Djemai, M. Mean Square Consensus of Double-Integrator Multi-Agent Systems under Intermittent Control: A Stochastic Time Scale Approach. Journal of the Franklin Institute 2019, 356. [CrossRef]
  25. Duque, C.; Leiva, H. CONTROLLABILITY OF A SEMILINEAR NEUTRAL DYNAMIC EQUATION ON TIME SCALES WITH IMPULSES AND NONLOCAL CONDITIONS. TWMS Journal of Applied and Engineering Mathematics 2023, 13, 975–989.
  26. Bohner, M.; Wintz, N. The linear quadratic regulator on time scales. Int. J. Difference Equ 2010, 5, 149–174.
  27. Bohner, M.; Wintz, N. The linear quadratic tracker on time scales. International Journal of Dynamical Systems and Differential Equations 2011, 3, 423–447. [CrossRef]
  28. Mu, Z.; Jie, T.; Zhou, Z.; Yu, J.; Cao, L. A survey of the pursuit–evasion problem in swarm intelligence, volume = 24, journal = Frontiers of Information Technology & Electronic Engineering, doi = 10.1631/FITEE.2200590, 2023. pp. 1093–1116. [CrossRef]
  29. Sun, Z.; Sun, H.; Li, P.; Zou, J. Cooperative strategy for pursuit-evasion problem in the presence of static and dynamic obstacles. Ocean Engineering 2023, 279, 114476. [CrossRef]
  30. Chen, N.; Li, L.; Mao, W. Equilibrium Strategy of the Pursuit-Evasion Game in Three-Dimensional Space. IEEE/CAA Journal of Automatica Sinica 2024, 11, 446–458. [CrossRef]
  31. Venigalla, C.; Scheeres, D.J. Delta-V-Based Analysis of Spacecraft Pursuit–Evasion Games. Journal of Guidance, Control, and Dynamics 2021, 44, 1961–1971. [CrossRef]
  32. Feng, Y.; Dai, L.; Gao, J.; Cheng, G. Uncertain pursuit-evasion game. Soft Comput. 2020, 24, 2425–2429. [CrossRef]
  33. Bertram, J.; Wei, P., An Efficient Algorithm for Multiple-Pursuer-Multiple-Evader Pursuit/Evasion Game; AIAA Scitech 2021 Forum, 2021; [https://arc.aiaa.org/doi/pdf/10.2514/6.2021-1862]. [CrossRef]
  34. Ye, D.; Shi, M.; Sun, Z. Satellite proximate pursuit-evasion game with different thrust configurations. Aerospace Science and Technology 2020, 99, 105715. [CrossRef]
  35. Libich, J.; Stehlík, P. Macroeconomic games on time scales. Dynamic Systems and Applications 2008, 5, 274–278.
  36. Petrov, N.; Mozhegova, E. On a simple pursuit problem on time scales of two coordinated evaders. Chelyabinsk Journal of Physics and Mathematics 2022, 7, 277–286. [CrossRef]
  37. Mozhegova, E.; Petrov, N. The differential game “Cossacks–robbers” on time scales. Izvestiya Instituta Matematiki i Informatiki Udmurtskogo Gosudarstvennogo Universiteta 2023, 62, 56–70. [CrossRef]
  38. Minh, V.D.; Phuong, B.L. LINEAR PURSUIT GAMES ON TIME SCALES WITH DELAY IN INFORMATION AND GEOMETRICAL CONSTRAINTS. TNU Journal of Science and Technology 2019, 200, 11–17.
  39. Lewis, F.L.; Vrabie, D.L.; Syrmos, V.L. Optimal Control; John Wiley & Sons, Inc., 2012. [CrossRef]
  40. Poulsen, D.; Davis, J.; Gravagne, I. Optimal Control on Stochastic Time Scales. IFAC-PapersOnLine 2017, 50, 14861–14866. [CrossRef]
  41. Siegmund, S.; Stehlik, P. Time scale-induced asynchronous discrete dynamical systems. Discrete & Continuous Dynamical Systems - B 2017, 22. [CrossRef]
  42. Poulsen, D.; Wintz, N. The Kalman filter on stochastic time scales. Nonlinear Analysis: Hybrid Systems 2019, 33, 151–161. [CrossRef]
Figure 6.1. Two-dimensional LQPEG on an isolated, uneven T
Figure 6.1. Two-dimensional LQPEG on an isolated, uneven T
Preprints 177763 g001
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated