Performance Analysis of Pursuit-Evasion Game-Based Guidance Laws
- Author: Kim Young-Sam, Kim Tae-Hun, Tahk Min-Jea
- Organization: Kim Young-Sam; Kim Tae-Hun; Tahk Min-Jea
- Publish: International Journal Aeronautical and Space Sciences Volume 11, Issue2, p110~117, 15 June 2010
-
ABSTRACT
We propose guidance laws based on a pursuit-evasion game. The game solutions are obtained from a pursuit-evasion game solver developed by the authors. We introduce a direct method to solve planar pursuit-evasion games with control variable constraints in which the game solution is sought by iteration of the update and correction steps. The initial value of the game solution is used for guidance of the evader and the pursuer, and then the pursuit-evasion game is solved again at the next time step. In this respect, the proposed guidance laws are similar to the approach of model predictive control. The proposed guidance method is compared to proportional navigation guidance for a pursuit-evasion scenario in which the evader always tries to maximize the capture time. The capture sets of the two guidance methods are demonstrated.
-
KEYWORD
Pursuit-evasion game , Differential game , Direct method
-
The pursuit-evasion game is a kind of differential game demonstrated by Isaacs (1967). It is an important class of two-player zero-sum differential games with perfect information and is useful in military applications, especially in missile guidance applications. In the game, the pursuer and the evader try to minimize and maximize the intercept time or the miss-distance as a payoff function. Among previous studies, Guelman et al. (1988) studied a simple case of pursuit-evasion game that can be solved without addressing a two-point boundary value problem. Breitner et al. (1993) proved that a multiple shooting method is able to precisely solve practical pursuit-evasion games subject to state constraints. In addition to these studies, a variety of parameter optimization methods such as those described in (Hargraves and Paris, 1987) can be extended to solve realistic pursuit-evasion games. In this study, we propose a direct method for solving pursuit-evasion games (Tahk et al., 1998a, b).
We first summarize the algorithm proposed in (Tahk et al., 1998a, b) for the reader's convenience. The algorithm is a direct method based on discretization of control inputs for solving pursuit-evasion games for which the intercept time is the payoff function of the game. Every iteration of the algorithm has two features: update and correction. The update step improves the evader's control to maximize the intercept time, and modifies the pursuer's control to satisfy the terminal condition. After applying the update step several times, the correction step is used to minimize the pursuer's flight time.
Note that the solution of the pursuit-evasion game can be used for guidance of the pursuer and the evader if the solution is obtained in real time. For this purpose, the iteration number of the solver is limited to reduce the computation time, at the cost of solution accuracy (Kim et al., 2006). In this study, the work of Kim et al. (2006) was refined for better numerical efficiency. Also, the capture set of the proposed guidance method is compared with that of proportional navigation to confirm that the proposed differential game guidance laws provide a smaller no-escape envelope.
The problem is a two-dimensional pursuit-evasion game with the final time as a payoff function. The dynamics of the pursuer and the evader are expressed as follows:
The admissible control inputs
u p (t ) andu e (t ) are assumed to be piecewise-continuous functions subject to the following constraints:Our pursuit-evasion problem is written as
subject to the constraints in Eq. (2) and the final constraint,
To develop a numerical algorithm, we used a direct approach based on parameterization of the control inputs. The control inputs of both players are separated into the following form:
where
u p,k andu e,k , which are assumed to be constant in the k-th interval, should satisfy the constraints in Eq. (2) as follows:The solution technique proposed in (Tahk et al., 1998b) is composed of two procedures: update and correction. The update procedure determines δ
u e , a small perturbation ofu e , in the direction of maximizingt f * . It also includes an algorithm for the computation of a newu p * . Instead of finding the minimum time solution, this algorithm computes δu p, the smallest variation ofu p *, to satisfy the final constraint (Eq. 4). Ifu p * is far from the minimum time solution, the correction procedure executes and provides an adequate tuning ofu p * to satisfy the optimality within the error bounds defined by the user. The aforementioned procedures, illustrated in Fig. 1, are implemented repeatedly until there is no improvement int f * . In (Kim et al., 2006), the update and correction procedures are called step 1 and step 2 to avoid confusion of terminologies.We assume that a capture within a finite time is guaranteed. Any perturbations of both players' control input, δ
u e and δu p , are said to be admissible if they satisfy following conditions:The capture condition is expressed as
where
P andE are defined asand
v p andv e are the sensitivities ofr p (t f ) andr e (t f ), respectively, to the perturbation oft f. We definev r as follows:Then, Eq. (8) can be written as
Let
n 1 andn 2 denote two orthonormal unit vectors in the two-dimensional space. We definen 1 asNow,
P , E , and (특수문자)r f are expressed aswhere
p 1 ,p 2 ,e 1 , ande 2 areN '1 column vectors andd 1 andd 2 are scalars. For a given δu e , let δu p o , which is orthogonal to the null space ofp T 2 , be the minimum norm solution of the combination of Eqs. (11) and (13), which is expressed as1If δ
u p o is admissible, thenδu p *=δu p o. Foru p=u p* andt f=t f*, δu p satisfying the capture condition can be written aswhere δ
u p n is the vector in the null space ofp T 2 , which can be chosen freely without affecting the capture condition. After some manipulations of Eqs. (14) and (15), we obtainwhere
The evader's control input is then updated as follows:
where j denotes the iteration number, and subscript k is kth element of the separated evader's control variables.
When the deviation of
u p andt f from the minimum norm solution grows too much, an optimal δu p n should be found to correctu p andt f . δu p n , an undefined term of Eq. (16), is calculated as follows. We write δu p n aswhere δ
u p v is the amount of constraint violation ofu p+δu p o+δu p' n, and δu p r is used to make δu p n which belongs to the null space ofp T 2 . Given δu p v , δu p r can be expressed aswhere δ
u pro belongs to the null space of , and δu p rn is to be determined in an optimal way.For minimizing (수식삽입), we find the δ
u pro that minimizes the norm of -δu p v +δu p ro. The norm of δu p n=-δu p v +δu p rois minimized byAfter calculating δ
u p n , each component of is examined to check whether or not any violations of control input constraints occur. As soon as δ δu p n is determined, δu p is updated using the following relationship:Once
u e andu p areu pdated,t f isu pdated usingIn Eqs. 22) and (23),
i denotes the iteration number.2.2 Step 2 (correction algorithm)
In step 1, the pursuer's control input is determined to satisfy the capture condition, and
u p +δu p is assumed to close to the optimal control variable for the case of small perturbations of both players' control variables. If the pursuer's control inputu p is far from the optimum to minimize a payoff function, we optimize the pursuer's control variable during the correction procedure.Assume that the pursuer's control input is subject to variation, but the evader's control input is fixed; that is, δ
u e = 0 . The first variation necessary condition for optimality is given by for all admissible δu p . To inspect the optimality condition at a later time, we defineI S as the set of saturated parameters. It is represented aswhere
K S ands(i) denote the number of elements inI S, and itsi th element. We construct aN ×K s matrixS asThe purs
u er's control variable is restricted to constraints, and a saturated control variable can be perturbed only in one direction, as shown byThen, let
q 1 denotes the component ofp 1 normal top 2 , which is given byHence, the Kuhn-Tucker condition is written as
where mk > 0 for all . We also derive the following necessary conditions for optimality:
For the correction procedure, we set dδ
u e = 0 as previously mentioned. Then, we substitute all elements ofq 1 that do not satisfy the necessary conditions for optimality by zero to obtain (특수문자), and we compute (특수문자), the component of (특수문자) normal to (특수문자):We choose d
u prn asto reduce
t f because is the null space ofp 2T . We compute δu prn and the total correction of du p given by the following relations, starting from δu po=δu pv=δu pro, as follows:If there is any violation of the control input constraint, we use the
u pdate procedure ofu p as explained in step 1.3. Constructing the Proposed Guidance Law
In this section, we explain the method we used to construct the guidance law proposed in (Kim et al., 2006). The algorithm introduced in the preceding section consists of two steps. This algorithm initializes both players' control inputs, and then checks the optimality conditions. If the pursuer's control input and final time do not satisfy the optimality conditions, step 2 is performed; otherwise, step 1 is executed. The evader tries to maximize the final time or to avoid capture by the pursuer, and the pursuer tries to intercept the evader. After performing step 1, we examine whether a difference in the positions of the players at the final time is within the error bound. If it is within the error bound, a variation of the evader's position at the final time is examined. If the variation is within the error bound, the algorithm is terminated and gives optimal trajectories. Otherwise, step 1 is repeated. If the difference in the positions of the players at the final time is not within the error bound, the optimality conditions of the pursuer's control input and the final time are checked again.
The initializations of both players' control inputs are the same as the algorithm proposed in (Tahk et al., 1998a, b). In this case, the algorithm is started with step 2. At the proposed algorithm, step 2 is executed when a violation of the optimality condition has occurred and step 1 is repeated when a variation of the evader's position at the final time is not within the error bound. To construct a guidance law, this procedure is changed to execute step 2 once and step 1 ten times. The dynamics of both players are integrated with specific integration time interval δ
t using the initial value of the game solution after performing step 2. Then, state variables changed by integration are considered as initial conditions. The termination condition is the same as the proposed algorithm. The algorithm as a guidance law is described in Fig. 2.With integration, this modified algorithm changes control inputs. First, control input elements of both players are got rid of the control input sets after an integration and the others are replaced previous elements of each control inputs. The reason is that the first control input elements of the pursuer and the evader are used during integration, and both players' positions are updated by integration. By reason of integration with δ
t , both players have new initial conditions. After executing integration and replacement (movement) of control inputs, the control input elements of both players are improved during step 1 as previously mentioned. In this respect, the proposed guidance law is similar to the approach of model predictive control (MPC). A schematic diagram of the movement of control input elements is shown in Fig. 3. In addition, we reduced the integration time interval δt whent f /N was smaller than the specific integration time interval δt . In this case, time interval δt was reduced tot f /N . If δt is used consistently, pursuit-evasion game solutions are calculated with the small time interval compared to δt . In this case, integration with δt can result in a large miss distance because the dynamics of both players integrate with a much longer time interval than with a discrete time interval of their control inputs near the final time. Thus, we should reduce the integration time interval, which guarantees that integration time interval δt is close tot f /N . If the integration time interval continu es to decrease, δt becomes infinitesimal and the iteration executes infinitely. To prevent this, the integration time interval is fixed the specific value near the final time and the iteration number of the solver is limited. Applying this technique, we can improve the interception precision at the final stage. This guidance law is called the differential game guidance law.The differential game guidance law proposed in (Kim et al., 2006) wasapplied to the planar pursuit-evasion game problem illustrated in Fig. 4. The equations of motion of the evader are described as
In these equations,
x e andy e denote the position in Cartesian coordinates, γe is the flight path angle,v e is the speed,u e is the non-dimensional control input variable,R e is the minimum turn radius, andc is the drag coefficient. If the drag coefficient is zero, the evader does not produce a drag force and its speed does not decrease.The equations of motion of the pursuer, which are analogous to those of the evader, are as follows:
where
R p is the minimum turn radius, andxx pr,y pr,γ pr,v pr, andu p are analogous to the evader model. The c-3ptnstanta , which is composed of the zero-lift-drag coefficientC D0 and maximum lift coefficientC Lmax, is defined asThe constant
b consists of the maximum lift coefficient and the induced drag factorK :The engagement scenario for the numerical example is shown in Fig. 5. The evader is initially 6,724.97 m apart from the pursuer along the x-axis, and moves with
v e(0) = 300m /s and (0) = 131.86°. The pursuer initially moves withv p(0) = 920.83m /s and (0) = 85.36°. The parameters used in this engagement model are taken from (Guelman et al., 1988) asR e = 600m ,R p = 1515.15m ,a = 0.0875,b = 0.4. For a realistic evader, the drag coefficient was chosen as 0.4.The differential game guidance law parameters were chosen as
N = 50 and ε1 = 0.002m . The algorithm proposed in (Tahk et al., 1998a, b) includes the parameter ε1, but the differential game guidance law does not define it. The reason is that the proposed guidance law; i.e., the differential game guidance law, does not require it because of the difference in the execution control between step 1 and step 2, as mentioned in the previous section.The trajectories of the pursuer and the evader are shown in Fig. 6. The pursuer intercepts the evader at 18.565769 seconds. The time histories of the control input of both players are illustrated in Fig. 7. The figure shows that an initial control input of the evader is a zero. The figure also shows that the control inputs of both players are constrained by Eq. (2).
To compare the performance of the differential game guidance law with other guidance laws, proportional navigation guidance (PNG) was adopted as a guidance law of the pursuer. In spite of using PNG, we assumed that the evader knows the maneuvers of the pursuer which are optimal. In this case, the pursuer's control inputs are calculated from
where
N ' denotes the effective navigation ratio, andV c and are the closing velocity and line-of-sight rate, respectively. Froma cmd, the control input of the pursu er is determined as follows:It is also constrained by the first expression in Eq. (2). Using PNG as a guidance law, the termination condition is replaced by the miss distance. If the miss distance is smaller than 1 m, we believe that the purs
u er intercepts the evader.Trajectories of the pursuer using PNG as a guidance law, and of the evader, are given in Fig. 8. The pursuer captured the evader at 23.339497 seconds. Time histories of both control inputs are shown in Fig. 9, and are similar to Fig. 7. The pursuer's initial control input is a few different between the differential game guidance and PNG. This is because a control input of the pursuer using the proposed guidance law is initialized by PNG, and then it is improved by step 2.
Table 1 is a summary of the simulation results. The intercept time of the differential game guidance law is shorter than the time using PNG. In the case of using PNG, a computation time is approximately three times longer than that using pursuit-evasion game solution. The number of iterations needed to capture the evader is nearly two times greater than that of the proposed guidance law.
To compare the interception performance of both methods, which are both players using the differential game guidance law and the pursuer using PNG, the capture set was calculated. For simplicity, we considered only the initial condition for the previous example except that the initial flight path angles of the pursuer and the evader varied from 0° to 20° and from 0° to 350°, respectively. The reason why we selected the initial flight path angle ranges of the pursuer is that seeker's gimbals used typical guided missiles adopt lock-on type. The conventional field of view of a seeker is limited from 2° to 5°, but the field of view of seeker-adopted lock-on type gimbals is larger than 15°. The field of view of gimbaled seekers, however, is not much larger than 15°. Thus, the initial flight path angle ranges of the pursuer were chosen as 0° to 20°.
Figures 10 and 11 show the capture sets of the proposed guidance law, and the case using PNG as a guidance law of the pursuer. In these figures, the initial conditions marked with (특수문자) are those for which the pursuer can capture the evader within a finite time. The other symbols indicate the termination of the program without a capture. In Fig. 10, the symbol ★ indicates a circumstance for which the pursuer cannot capture the evader under the termination condition of the differential game guidance law, but is able to intercept the evader under the termination condition when using PNG as a guidance law of the pursuer. The symbol × implies that the pursuer is not able to intercept the evader under both termination conditions. In Fig. 11, the symbol × indicates that the miss distance is larger than 1 meter, so interception by the pursuer does not occur. From Figs. 10 and 11, we know that the capture set of the proposed guidance law is larger than one using PNG as a guidance law of the pursuer.
We carried out a performance analysis of the proposed guidance law based on pursuit-evasion game solutions that were sought by the pursuit-evasion game solver. The derivation of the proposed pursuit-evasion game solver was described, and construction of the proposed guidance law was explained in detail. The guidance law proposed in (Kim et al., 2006) consists of two procedures, update and correction, called step 1 and step 2. Control inputs of the pursuer and the evader were
initialized in the differential game guidance law, and then improved by step 1 and 2. Step 2 executes first that this procedure provides a way to optimize the pursuer trajectory when it deviates excessively from the optimal trajectory. Then, the dynamics of both players were integrated with a specific time interval. Step 1 performs that control inputs of the evader were updated to maximize the increment in the capture time, while the pursuer tried to minimize it. This procedure iterated until satisfying the termination condition. The differential game guidance law was applied to solve a numerical example. For a comparison of the interception performance, PNG was adopted to a guidance law of the pursuer. Simulation results of the engagement scenario were tabulated for each case--the differential game guidance law, and using PNG as the guidance law of the pursuer. Capture sets of both cases were calculated for performance analysis. Based on our numerical example simulation results, we know that the differential game guidance law provided a smaller no-escape envelop than PNG.
-
[Fig. 1.] Flowchart of the algorithm in (Kim et al., 2006).
-
[Fig. 2.] Flow chart of the algorithm after modification (Kim et al., 2006).
-
[Fig. 3.] Movement of the control input elements (Kim et al., 2006).
-
[Fig. 4.] Pursuer and evader engagement geometry (Kim et al., 2006).
-
[Fig. 5.] Intercept scenario (Kim et al., 2006).
-
[Fig. 6.] Trajectories of the pursuer and evader.
-
[Fig. 7.] Time histories of the control input of both players.
-
[Fig. 8.] Trajectories of the pursuer using PNG and the evader.
-
[Fig. 9.] Time histories of the control input of the pursuer using PNG and the evader.
-
[Table 1.] Summary of simulation results
-
[Fig. 10.] The capture set of the proposed guidance law.
-
[Fig. 11.] The capture set of the case using proportional navigation guidance.