Journal of Applied & Computational Mathematics

Page 1

Computat

i

Olotu et al., J Appl Computat Math 2019, 8:3

Journal of

Applied & Computational Mathematics

athem at i

of Appl ie nal ur

&

al M on

d

ISSN: 2168-9679

cs

Jo

Research Article

Open Access

Algorithm for Solving Multi-Delay Optimal Control Problems Using Modified Alternating Direction Method of Multipliers Olotu O, Dawodu KA* and Yusuf TT Department of Mathematical Sciences, Federal University of Technology, Akure, Nigeria

Abstract Purpose: This study presents an algorithm for solving Optimal Control Problems (OCPs) with objective function of the Lagrange-type and multiple delays on both the state and control variables of the constraints. Design/Methodology/Approach: The full discretization of the objective functional and the multiple delay constraints was carried out using the Simpson numerical scheme. The discrete recurrence relations of both the objective function and constraints were generated and used to develop the matrix operators of the optimal control system. The spectral and convergence analyses of the developed matrix operators and formulated unconstrained optimal control problem were carried out respectively to ascertain the well-posedness of the Modified Alternating Direction Method of Multipliers (M-ADMM). The primal-dual feasibility and their residuals were derived in other to obtain the point of convergence of the algorithm. Findings: The direct numerical approach for handling the multi-delay control problem was observed to obtain a very accurate result at a super-linear rate of convergence. This makes the algorithm faster when compared to the classical Pontryagin maximum principle. Research limitations/Implications: The research is limited to linear constraints and objective functional of the Lagrange-type. Practical implications: This research can address real-life models with multiple delays in epidemiology. Originality: The novelty of this research paper lies in the method of discretization of the multiple delay constraints and the adaptation of the modified ADMM algorithm in handling linearly constraint multiple delay optimal control problems for better performance in terms of speed and accuracy.

Keywords: Multiple delays; Recurrence relation; Alternating direction method of multipliers; Primal residual; Dual residual; Superlinear convergence; Composite Simpson’s rule Introduction Optimal Control Systems (OCS) with linear and nonlinear inequality constraints with delays in state and/or control variables play important roles in the modeling of real-life phenomena [1]. Many papers over the years have been devoted to developing the maximum principles for optimal control problems with constant and variable delays because the time delay effects in control systems cannot be neglected especially when it involves the transmission of information between different parts of a system [2]. This hereditary effect which frequently occur in economical, physical, chemical and biological processes have been deliberated upon by Bashier and Patidar [3]. Bertsekas [4] generalized the classical Pontryagin’s maximum principle for optimal control problems with delay in the state only. Contemporarily, Boley [5] obtained the maximum principle for the non-autonomous linear-quadratic optimal control problem with multiple delays in the state only [6]. Maximum principle for control systems with a timedependent delay in the state variable was also derived in Boyd et al. [7]. The maximum principle for linear-quadratic optimal control problems with multiple constant delays in both state and control variables was obtained by Chyung and Lee [8]. Colonius and Hinrichsen [9] provided a unified approach to control problems with delays in the state variable only while Ghadimi et al. [10] established maximum principle for the delay optimal control problem with multiple variable delays by the theory of optimal fields. The direct numerical approach was developed by Gilbert [11] using the Quasi-Newton embedded augmented Lagrangian functional for the proportional control class of Optimal J Appl Computat Math, an open access journal ISSN: 2168-9679

Control Problems. Goldfarb and Schneiberg [12] did an extensive work on mixed control-state inequality constraints with single delay on both state and delay variables with the initial and terminal boundary conditions in a general mixed form [13]. Later, Gollmann and Maurer [14] extended their work to multiple time delays in the control and state variables with mixed control-state constraints. The Pontryagintype minimum principle for the class of delayed control problems was derived to ascertain their solutions [15]. It is therefore the aim of this paper to introduce the Douglas-Rachford (D-R) splitting method called the Alternating Direction Method of Multipliers (ADMM) as a direct numerical approach in solving Multiple Delay Optimal Control Problems [16]. ADMM therefore enjoys the strong convergence properties of the method of multipliers and the decomposability property of dual ascent [17]. It has the potential to solve large-scaled structured convex optimization problems by solving the primal and dual feasibility updates in parallel [18]. The method gained wide acceptability due to its simplicity, versatility, scalability and ability to solve constrained optimization problems. A number of authors such

*Corresponding author: Dawodu KA, Department of Mathematical Sciences, Federal University of Technology, Akure, Nigeria, Tel: 08032541380; E-mail: dawodukazeem@yahoo.com Received April 24, 2019; Accepted May 10, 2019; Published May 17, 2019 Citation: Olotu O, Dawodu KA, Yusuf TT (2019) Algorithm for Solving MultiDelay Optimal Control Problems Using Modified Alternating Direction Method of Multipliers. J Appl Computat Math 8: 445. Copyright: © 2019 Olotu O, et al. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

Volume 8 • Issue 3 • 1000445


Citation: Olotu O, Dawodu KA, Yusuf TT (2019) Algorithm for Solving Multi-Delay Optimal Control Problems Using ModiďŹ ed Alternating Direction Method of Multipliers. J Appl Computat Math 8: 445.

Page 2 of 9 as Laarabi et al. [19], Nesterov [20], O’Donoghu et al. [21], Olotu et al. [22], Rihan and Anwar [23] and Soliman [24] further worked on the linear convergence of the ADMM with strictly convex inequality [25]. Others include: Yang and Zhang [26], Zhang et al. [27], who worked on the acceleration of the ADMM, called the Accelerated Alternating Method of Multipliers (A-ADMM), to improve on its efficiency and rate of convergence. Nesterov [20] and Zhang et al. [27] improved on the ADMM by deriving the optimal parameter using the GaussSeidel relaxation factor. However, the A-ADMM will be deployed and modiďŹ ed as an optimization tool for solving the discretized continuous multiple control problem.

the continuous-time multiple delayed constraint, thereby generating a recurrence relation that was used to obtain the large sparse matrix operators. The Augmented Lagrangian functional was used to convert the discretized constrained problem to an unconstrained form. An accelerator variant (relaxation factor) was introduced in the formulation of the modiďŹ ed ADMM algorithm to accelerate its rate of convergence. The spectral analyses of the matrix operators and the convergence analysis of the modiďŹ ed ADMM were carried out to ensure it is wellposed and implementable on the MATLAB subroutine.

Statement of Problem

The state variable x(t)=(x1,x2,...xn)∈Rn and the control variable u(t)=(u1,u2,....,um) ∈ Rm deďŹ ned over the intervals [t0-r,t0] and [t0-q,t0] respectively are sets of n and m dimensional Banach spaces of the continuous functions of the state and control vectors respectively. Suppose the time interval is discretized by letting δ = (T − t0 ) ∈ ď š + N such that NâˆŠď š+ and tk=t0+kδ, (for k=0, 1, ‌N), then the discretization operator fx maps each discrete point in the time interval [t0,T]⊆â„? into each discrete point of the concatenated state vector xi(k)(.)∊RnN for all i=1, 2, ‌, n; while, the operator fu maps the points into each discrete point of the concatenated control vector uj(k)(.)∊Rm(N+1) for all j=1, 2, ‌, m as expressed below:

The general form of the multiple delay optimal control problem is: Minimize J ( x, u ) =

1 T F ( t , x ( t ) , u ( t ) ) dt 2 âˆŤt0

f (t, x (t ) , u (t ) , xh (t ) , u h (t ))

Subject to x ( t )

(1) t ∈ [ t0 , T ]

(2)

= x (t ) Ď• (t ) ,

t ∈ [t0 − r , t0 ] ,

(3)

= u (t ) Ďˆ (t ) ,

t ∈ ( t0 − q, t0 ) ,

(4)

x ( t0 ) = x0 ,

(5)

where, x ∈ â„? , u ∈ â„? , r=max{rj} j=1, q=max{q} l=1 F:[t0,T] Ă— â„? Ă— â„? → â„? n

m

d

e

n

m

f :[t0, T] Ă— â„?(d+1)n Ă— â„?(e+1)m → â„?n, đ??‹(t) : [t0-r; t0] → â„?n, đ?&#x; (t): [t0-q; t0] → â„?m, rj < rj+1 and ql <ql+1.

However, with respect to all arguments, the functions F and f are assumed to be at least twice continuously differentiable while the functions đ??‹ and đ?&#x; are only continuous. The pair of functions (x,u)âˆˆâ„Ś:=([t0,T],Rn) Ă— ([t0,T],Rm) is an admissible pair of the problem above such that the conditions of eqn. (1) to eqn. (5) are satisďŹ ed. However, it is imperative to develop the numerical approach to solving the case where the continuous nonlinear functional F and constraint function f are quadratic and linear in nature respectively; hence the optimal control model below: Min = J ( x, u )

1 T t0 ( xT Px + u T Qu ) dt 2âˆŤ d

(6)

e

s.t x ( t ) ≤ Ax + Bu + ∑ Îą j x ( t − rj ) + ∑ β l u ( t − ql ) t0 ≤ t ≤ T (7) =j 1 =l 1

= x ( t ) Ď• ( t ) , t0 − r ≤ t ≤ t0

(8)

= u ( t ) Ďˆ ( t ) , t0 − q ≤ t ≤ t0

(9)

x ( t0 ) = x0

(10)

where,

= r max rj ,= q max ql ,= x 1≤ j ≤ d

1≤ j ≤ e

( x1 , x2 ,..., xn )

T

∈ � n ,= u

( u1 , u2 ,..., u m )

T

∈

â„? m , P ∈ â„? nĂ— n , Q ∈ â„? mĂ— m , A ∈ â„? pĂ— n , B ∈ â„? pĂ— m ,Îą j ∈ â„? pĂ— n ∀ j , β l ∈ â„? pĂ— m∀l

while, all assumptions are as in the general form of the multiple delay control model above.

Materials and Methods The concept of "ďŹ rst discretize fully and then optimize" approach was adopted in obtaining the optimal solution of the multiple delay OCP. This involves applying an existing numerical scheme to discretize J Appl Computat Math, an open access journal ISSN: 2168-9679

Background and preliminaries

f x : [t0 , T ] ⊆ � → xi(

k)

(.) ∈ x

(11)

f u : [t0 , T ] ⊆ � → u (j

k)

( .) ∈ u

(12)

where, x = = xi

( x1 , x2 ,..., xn ) ∈ � nN

,u ( x ( ) , x ( ) ,..., x ( ) ) ∈ �= i

1

i

2

i

N

N

= uj ( u1 , u 2 ,..., u m ) ∈ � m( N +1) and

( u ( ) , u ( ) ,..., u ( ) ) ∈ � 0 j

1 j

N j

N +1

.

Considering the existence of single delay constants r and q on the state and control variables over the interval [t0-r,T] and [t0-q,T], respectively, such that r=Ď…δ and q=wδ; the discretized delay state and delay control variables are hereby represented in the equations below:  x h ( t k − v ) =Ď• ( t k − v ) : k ≤ v;k =0,1, 2,..., Ď…, t ∈ [ t 0 − r, t 0 ] , x ( t 0 ) =x 0 ( known state ) f xh =   x ( t k − v ) =x k − v : k > v;k =( v + 1) , ( v + 2 ) ,..., N, t ∈ [ t 0 ,T ] ( Unknown state )

(13)

and u h ( t k − w ) = 0,1, 2,..., w, t ∈ [ t 0 − q, t 0 ] ( known control ) Ďˆ ( t k − w ) : k ≤ w;k = f uh =  = > = + u t u : k w; k w 1) , ( w + 2 ) ,..., N, t ∈ [ t 0 ,T ] Unknown state. ( ) (  k − w k−w

(14)

For the case of multiple delay constants rj∊R (for j=1, 2,..., d) and ql∊R (for l=1, 2,..., e) on the state and control variables respectively, it is assumed that the numbers are monotonically increasing; that is, (rj<rj+1) and (ql<ql+1). Therefore, the multiple state and control variables for eqns. (13) and (14) for j=1,2,...,d and l=1,2,...,e are represented in eqns. (15) and (16):

(

xk −n j

and

)

 x hk − v = Ď• t k−vj : k − v j ≤ k = 0,1, 2,..., v j [ t 0 − r, t 0 ] ( known ) j   =  xk − v j : k - v j > 0;k =( v j + 1) ,.., N, t ∈ [ t 0 ,T ] ( unknown ) (15)  x ( t 0 ) = x 0 given  

(

)

u hk − w = Ďˆ t k − w l : k − w l < 0;k = 0,1, 2,,..., ( w l − 1) , t ∈ [ t 0 − q, t 0 ] , ( known ) l u k − wl =  − ≼ = u : k w 0fork w , w  k − w l l l ( l + 1) ,..., N, t ∈ [ t 0 ,T ] , ( unknown )

(16)

r= rd= max{r j }dj 1 = and q= q= max{q l }le 1 are the largest where, = e delays on the state and control variables respectively. In other words, the discretized delay state and control variables are represented in the form below: Volume 8 • Issue 3 • 1000445


Citation: Olotu O, Dawodu KA, Yusuf TT (2019) Algorithm for Solving Multi-Delay Optimal Control Problems Using Modified Alternating Direction Method of Multipliers. J Appl Computat Math 8: 445.

Page 3 of 9

( x(

1− v j

= xˆ

(

= uˆ

u

)

( − wl )

,x

,u

(2−v j )

(1− wl )

,...., x

,...., u

( N −v j )

( N − wl )

for all j )∈ℝ = nN

)

∈ℝ

m ( N +1)

problem then yielded the recurrence relation below:

1, 2,..., d ,

for = all l 1, 2,...,e,

(k −vj )

u

(k −vj )

( x(

k−vj 1

(u

)

(k − wl )

1

(k −vj )

, x2

(k −vj )

,..., x n

(k − wl )

, u2

) ∈ ℝ , for any j,

(k − wl )

,..., un

δ

k)

(x

d

∑α

+ 4σ x ( (k − j)

k +1)

+x

+ (σ − I x ) x (

( k +1− j )

+x

(k + 2− j)

k + 2)

+ γ u(

δ

(17)

n

)∈ℝ

m

, for any l,

(18)

3

k)

+ 4γ u (

) − 3 ∑ β (u e

j l 1 =j 1 =

where, for any arbitrary k (for k = 0,1,2,....,N), the discretized delay state and control vectors are concatenated giving, x

+ I x ) x(

l

(k −l )

k + 2)

+u

=

( k +1 − l )

+ u(

k + 2−l )

(24)

)

where, Aδ Bδ σ = ∈ ℝ p× n , γ = ∈ ℝ p× m , I x ∈ ℝ p× n and k = 0,1,..., N − 2

3 3 Setting k=0 in eqn. (24) yielded:

(σ + I x ) x( ) + 4σ x( ) + (σ − I x ) x( ) + γ u ( ) + 4γ u ( ) + γ u ( ) = while x(t) and u(t) are estimated within their respective delay intervals by the given delay functions ϕ(t) and ψ(t) respectively such that ϕ(t) ≃ δ d δ e − ∑ α j x( − j ) + x(1− j ) + x( 2 − j ) − ∑ βl u ( − l ) + u (1− l ) + u ( 2 − l ) ). (t−k) and ψ(t) ≃ ψ(t−k) where tk=t0-kh for k=1,2,...,vd(≥ w= e 3 j 1= 3l1 0

1

2

(

0

)

1

2

(

)

Discretization of the objective function

Expanding and collecting like-terms of x and u yielded

In the discretization of the continuous-time objective function, the third order one-third Simpson rule in eqn. (19) below was used:

(4σ + α1 ) x (1) + (σ − I x ) x (2) + [γ + ( β1 + β 2 )]u (0) + (4γ + β1 )u (1) + γ u (2) = −[σ + I x + 3 3 3

  T δ f ( x0 ) + 2 ∑ f ( x2 k ) + 4∑ f ( x2 k −1 ) + f ( xN )  (19) ∫t0 F ( x, t ))dt  3 =   k 1= k 1   N −1 2

N 2

Then the discretized objective function of the optimal control problem in eqn. (6) using eqn. (19) is expressed below as:   N N −1  δ T 1δ 2δ 2 T 4δ 2 T x Px + x Px + x Px min J ( x, u )  min  x0T Px0 + ∑ 2k 2k 3 ∑ 2 k −1 2 k −1 N N  x ,u x 2 3 3 K 1= 3 = k 1      x T Px  

6

X 0T Px0 ∈ ℝ and

1 1

2

n

1

N

nN

0

1

(20)

the objective functional stated below as:

P

δ 3Q 0  2δ 0 Q  3               0 

          4δ  Q     3        4δ Q 0    3  δ  Q   0 3 

(21)

k

δ

 f k + 4 f k +1 + f k + 2  . 3

f k = Ax ( ) + Bu ( ) + ∑ α j x k

k

(k −v j )

e

+ ∑ βl u (

k − wl )

δ

(α d − 2 + α d −1 +

3

3

δ

3

3

( β e −1 + β e )u (1− e ) −

δ 3

(25)

3

β eu ( − e ) .

Setting k=1 in eqn. (24) yielded: [σ + I x +

δ

δ

3

δ

(α1 + α 2 )]x (1) + (4σ + α1 ) x (2) + (σ − I x ) x (3) + ( β1 + β 2 + β3 )u (0) + [γ + 3 3

( β1 + β 2 )]u (1) + (4γ +

δ

3

δ δ β1 )u (2) + γ u3 = − (α1 + α 2 + α 3 ) x (0) − (α 2 + α 3 + α 4 ) x ( −1) − 

δ

δ

δ

(26)

δ

− ( β 3 + β 4 + β )u ( −2) −  − ( β e − 2 + β e −1 + β e )u (3− e ) − ( β e −1 + β e )u (2 − e ) − β eu (1− e ) . 3 3 3 3

δ 3

d

e(≤ d )

j =1

l =1

[∑α j ( x ( d − 4 − j ) + x ( d − 3− j ) + x ( d − 2 − j ) ) +

∑ β (u l

(e − 4−l )

+ u ( e − 3− l ) + u ( e − 2 − l ) )]

Expanding and collecting the like-terms yielded: δ 3

(α d − 5 + α d − 4 + α d − 3 ) x (1) +

α d −2− s ) x

(s)

+ +

δ 3

δ

3

δ 3

δ 3

(α d − 6 + α d − 5 + α d − 4 ) x (2) +  +

(α1 + α 2 + α 3 ) x

( d − 5)

δ

δ 3

(α d − 4 − s + α d − 3− s +

δ

+ [ (α1 + α 2 ) + σ + I x ]x ( d − 4) + [ α1 + 4σ ]x ( d − 3) + 3 3

( β e − 4 + β e − 3 + β e − 2 )u (0) +

δ

3

( β e − 5 + β e − 4 + β e − 3 )u (1) +

( β e − 4 − s + β e − 3− s + β e − 2 − s )u ( s ) +  +

δ 3

δ

3

(βe −6 + βe −5 +

δ

( β1 + β 2 + β 3 )u ( e − 5) + [ ( β1 + β 2 ) + 3

(27)

δ δ δ γ ]u ( e − 4) + [ β1 + 4γ ]u ( e − 3) + γ u ( e − 2) = − (α d − 4 + α d − 3 + α d − 2 ) x (0) − (α d − 3 + α d − 2 + 3

3

3

δ δ δ δ α d −1 ) x−1 − (α d − 2 + α d −1 + α d ) x ( −2) − (α d −1 + α d ) x ( −3) − α d x ( −4) − ( β e − 3 + β e − 2 + 3

3

3

3

3

Setting k=d=e in eqn. (24) yielded:

(23)

=j 1 =l 1

Adapting eqns. (22) and (23) to eqn. (7) of the optimal control

J Appl Computat Math, an open access journal ISSN: 2168-9679

3

3

( β e − 2 + β e −1 + β e )u (2 − e ) −

3

(22)

,

δ

3

= x ( t ) f ( t, x, u ) implies that f ( t, x, u ) ≈ f ( t k , x k , = u k , xˆ k , uˆ k ) f k where d

(α 2 + α 3 + α 4 ) x ( −2) −  −

3

δ δ δ β e −1 )u ( −1) − ( β e − 2 + β e −1 + β e )u ( −2) − ( β e −1 + β e )u ( −3) − β eu ( −4)

1 Simpson’s rule in eqn. (22) was used for the discretization of 3 the multi-delay constraints: x( ) + =

−

β e − 4 )u (2) +  +

The

k + 2)

δ

δ δ δ δ α d ) x (2 − d ) − (α d −1 + α d ) x (1− d ) − α d x ( − d ) − ( β1 + β 2 + β3 )u ( −1) − ( β 2 + β3 + β 4 )u ( −2) −

(σ − I x ) x ( d − 2) +

0

Discretization of the constraints

x(

(α1 + α 2 + α 3 ) x ( −1) −

(σ + I x ) x ( d − 4) + 4σ x ( d − 3) + (σ − I x ) x ( d − 2) + γ u ( d − 4) + 4γ u ( d − 3) + γ u ( d − 3) =

respectively. However, the block-diagonal coefficient matrices m N +1 × m N +1 P ∈ ℝ nN × nN and Q ∈ℝ ( ) ( ) are the block-matrix operators of  4δ   0   3 P 0   2δ  0 P     3        and Q =     0    δ   P    0 3    

3

Setting k=d-4=e-4 in eqn. (24) yielded:

m N +1)

N m

1 1

δ

δ

3 3 h δ δ δ − (α d − 2 + α d −1 + α d ) x (3− d ) − (α d −1 + α d ) x (2 − d ) − α d x (1− d ) − ( β 2 + β3 + β 4 )u ( −1) 3 3 3 3

(u( ) , u( ) ,..., u ( ) ) ∈ ℝ (

and = u

δ

(α1 + α 2 )]x (0) −

3

3

the concatenated state and control variables

( x( ) , x( ) ,..., x( ) ) ∈ ℝ

are= x

δ

δ

  N N −1  δ T 1  δ T 2δ 2 T 4δ 2 T + min u0 Qu0 + u2 k Qx2 k + u2 k −1Qx2 k −1 + u N Qu N  ∑ ∑ u 2 3 3 3 3 K 1= k 1 =      u T Qu   1 1  min x T Px + u T Qu + R x ,u 2 2

δ where, = R

δ

δ 3

(α d −1 + α d ) x (1) +

α d − 2 ) x (4) +  +

δ 3

δ 3

(α d − 2 + α d −1 + α d ) x (2) +

δ 3

(α d − 3 + α d − 2 + α d −1 ) x (3) +

(α d − s + α d +1− s + α d + 2 − s ) x ( s ) +  +

δ 3

δ 3

(α d − 4 + α d − 3 +

δ

(α1 + α 2 + α 3 ) x ( d −1) + [ (α1 + α 2 ) + 3

δ δ δ δ σ + I x ]x ( d ) + [ α1 + 4σ ]xd +1 + (σ − I x ) x ( d + 2) + β eu (0) + ( β e −1 + β e )u (1) + ( β e − 2 + β e −1 + β e )u (2) + + +

δ

δ 3

δ 3

3

( β e − 3 + β e − 2 + β e −1 )u (3) +

δ 3

3

( β e − s + β e +1− s + β e + 2 − s )u ( s ) +  +

δ

3

( β e − 4 + β e − 3 + β e − 2 )u (4) +

δ

δ 3

3

(28)

( β e − 5 + β e − 4 + β e − 3 )u (5)

δ

( β1 + β 2 + β 3 )u ( e −1) + [ ( β1 + β 2 ) + γ ]u ( e ) 3 3

+ ( β1 + 4γ )u ( e +1) + γ u ( e + 2) = α d x (0) . 3 3

Setting k=d+t(1≤t<N−2−d) in eqn. (24) yielded:

Volume 8 • Issue 3 • 1000445


Citation: Olotu O, Dawodu KA, Yusuf TT (2019) Algorithm for Solving Multi-Delay Optimal Control Problems Using Modified Alternating Direction Method of Multipliers. J Appl Computat Math 8: 445.

Page 4 of 9 δ 3

α d x (t ) +

+ +

δ 3

δ 3

(α d −1 + α d ) x (1+ t ) +

δ

(α d − 2 + α d −1 + α d ) x (2 + t ) +

3

δ

(α d − s + α d +1− s + α d + 2 − s ) x ( s + t ) +  +

3

δ 3

(α d − 3 + α d − 2 + α d −1 ) x (3 + t )

δ

(α1 + α 2 + α 3 ) x ( d + t −1) + [ (α1 + α 2 ) + 3

δ δ δ δ σ + I x ]x ( d + t ) [ α1 + 4σ ]x ( d + t +1) + (σ − I x ) x ( d + t + 2) + β eu (t ) + ( β e −1 + β e )u ( t +1) + ( β e − 2 3

+ βe −1 + βe )u ( t + 2) + +

δ

3

δ 3

( βe − 3 + βe − 2 + βe −1 )u ( t + 3) +  +

δ

δ 3

3

3

(29)

( β e − s + β e +1− s + β e + 2 − s )u ( s + t ) +

δ

( β1 + β 2 + β3 )u ( e + t −1) + [ ( β1 + β 2 ) + γ ]u ( e + t ) + ( β1 + 4γ )u ( e + t +1) + γ u( e + t + 2) = 0 3 3

3

Setting k=N-2 in eqn. (24) yielded: δ 3

αd x( N − 2−d ) +

α d −1 ) x

( N +1− d )

δ 3

(α d −1 + α d ) x ( N −1− d ) +

++

δ

δ 3

+ β e )u

+

δ

3

( N − 2)

+ [ α1 + 4σ ]x 3

( β e − 2 + β e −1 + β e )u

+ ( β1 + 4γ )u 3

( N −1)

(α d − 2 + α d −1 + α d ) x ( N − d ) +

δ

+ β e +1− s + β e + 2 − s )u ( N + s − 2 − e ) +  +

δ

3

(α d − s + α d +1− s + α d + 2 − s ) x

+[ (α1 + α 2 ) + σ + I x ]x 3 ( N −1− e )

δ

+γu

(N )

( N −e)

δ

+

δ

3

( N + s −2−d )

( N −1)

++

+ (σ − I x ) x

δ 3

(N )

δ 3

(α d − 3 + α d − 2 +

(α1 + α 2 + α 3 ) x

+

( β e − 3 + β e − 2 + β e −1 )u

δ 3

( N +1− e )

+ +

δ

δ 3

δ 3

( β e −1

(30)

(βe − s

( β1 + β 2 + β 3 )u ( N − 3) + [ ( β1 + β 2 ) + γ ]u ( N − 2) 3 3

=0

The combination of all the equations for various values of k forms the linear inequality below: Ax + Bu ≤ Ex h + Fu h = C

(31)

where A is a multi-diagonal matrix with p(N-1) rows and nN columns (i.e., A ∈ ℝ p ( N −1)× nN ); B is also a diagonal matrix with p(N-1) p ( N − 1) rows and m( N + 1) columns (i.e., B ∈ ℝ p ( N −1)× m ( N +1) ), x is the concatenated vector of the state variable with dimension nN × 1 while u is the concatenated row vector of the control variable with dimension m(N+1) × 1. The dimensions of the matrices E , x h , F and u h are p ( N − 1) × n(d + 1), n(d + 1) × 1, p ( N − 1) × me and me × 1 respectively. The respective structures of the concatenated vectormatrices are represented below: δ   (4σ + 3 α1 ) (σ − I x )    (σ + I x + α 0 )   α1     Ax =   αd        0     (γ      Bu =         

+

δ 3

β 0 ) (4γ +

δ

β1 )

γ

0     

  0    

0

βe 

  

0

βe  β2

0   

α d  α1

3    

β1 β2

  0       β 2 β1 

  (1)   x    x (2)       x (3)             (d )  0 0 , x             0     x( N −1)  δ (σ + I x + α 0 ) (4σ + α1 ) (σ − I x )  x ( N )   3      

0

 0   u (0)    (1)     u  0     0 0   (e)    u  0 0          ,  .  δ δ γ (γ + β 0 ) (4γ + β1 ) 0   ( N −1)  3 3  u    (N)  δ δ β1 (γ + β 0 ) (4γ + β1 ) γ   u  3 3     

 x (0)  −α s −α d −1 −α d   ( −1)     −(σ + I x + α 0 ) −α1 x  −α1 −α s −α d −1 −α d 0   ( −2)     x   −α s −α d −1 −α d 0          −α s −α d −1 −α d 0          −α d −1 −α d 0         Ex h =  −α d −1 −α d 0          −α d 0 0   (2 − d )      and   x   0 0        (1− d )    x 0         ( − d )   x  0 0 0         

J Appl Computat Math, an open access journal ISSN: 2168-9679

 − β2 − β e −1  − β e −1 − β e 0 − β e −1 − β e 0  − βe 0           

 u (0)  − β e   ( −1)   x 0   ( −2)  u           ,          (1− d )   u    ( −d )  u  0     

where the delay coefficients α k and β k , for k=0,1,2,…,d are defined below as:

( N − 3)

β eu ( N − 2 − e ) +

 − β1   − β2    −β Fu h =  e −1 − βe   0     0

δ  3 (α k +1 + α k + 2 )  δ  3 (α k + α k +1 + α k + 2 ) αk =   δ (α + α ) k +1 3 k δ  αk  3 and

k =0 k = 1, 2,..., d − 2

(32)

k = d −1 k=d

δ k =0  3 ( β k +1 + β k + 2 )  δ (β + β + β ) k = 1, 2,..., e − 2 k k +1 k +2  (33) βk =  3 δ (β + β ) k = e −1 k +1 3 k δ  βk k =e  3 However, the entries [aij ] ∈ A,[bij ] ∈ B ,[eij ] ∈ E and [ fij ] ∈ F of the various matrix structures are described below as follows: (σ + I x + α 0 )  (σ − I ) x  [aij ] = (σ + I x + α 0 )  α  k  0

i = j 1 ≤ i ≤ N −1 j = i +1 1 ≤ i ≤ N −1 (34) j = i −1 2 ≤ i ≤ N −1 j = i − 1 − k 2 + k ≤ i ≤ N − 1; k = 1, 2,...,d elsewhere

δ  ( γ + 3 β0 )  δ  [bij ] = (4 γ + β1 ) 3   βk  0 

i = j 1 ≤ i ≤ N −1 j = i +1 1 ≤ i ≤ N −1 j = i − k 1 + k ≤ i ≤ N − 1; k = 1, 2,...,e elsewhere

 −( σ + I x + α 0 )  [eij ] =  − α k  0 

 − βk [fij ] =   0

(35)

i = j = 1 1 ≤ i ≤ N − 1; j = k + 2 − i 1 ≤ i ≤ k + 1; k = 1, 2,...,d (36) elsewhere

j = i + k − 1 1 ≤ i ≤ k; k = 1, 2,...,e elsewhere

(37)

Analysis of matrix operators The well-posessness and optimal parameter selections are dependent on the nature and properties of the discretized matrices P , Q, A and B . It is then imperative that P and Q be real symmetric Volume 8 • Issue 3 • 1000445


Citation: Olotu O, Dawodu KA, Yusuf TT (2019) Algorithm for Solving Multi-Delay Optimal Control Problems Using Modified Alternating Direction Method of Multipliers. J Appl Computat Math 8: 445.

Page 5 of 9 and positive definite to avoid ill-conditioning during the convergence of the algorithm. Theorem 3.1 (Sylvester's criterion): A real, symmetric matrix is positive definite if and only if all the principal minors are positive definite [11]. Corollary 3.1: Given a matrix P = diag[ p1 , p2 , pn ] with pi’s the principal diagonal entries and zeros elsewhere. Then the matrix P is positive definite if and only if all the pi’s are strictly positive (i.e., no zero principal diagonal entry). Proof: Let M j , j = 1, 2,, n be the principal minors of the real, symmetric matrix P. Then, M 1 = p1 = p1 > 0 M2 =

p1 0

0 = p1 p2 > 0 p2   pn −1 0

In the derivation of the ADMM algorithm, the optimality conditions were then applied to eqn. (40) above through the sequential minimization of x and u . The update of the state variable x is of the form: (42)

Then at optimum, ∇ x Lρ ( x , u k , z k , v k ) = 0 such that x = x k +1 yielded:

0 n  = p1 p2  pn = ∏ p j > 0 ∈ ℝ 0 j =1 pn

(

x k +1 = −ρ P + ρA T A

Since all the Mj’s are positive, then P is strictly positive definite. Considering the matrices P and Q , all the principal diagonal entries p,2p and 4p are strictly positive and as such they are positive definite.

The modified ADMM formulation The original ADMM formulation was modified to accommodate both the state and control variables before imposing the Karush-KhunTucker (KKT) optimality conditions. It was later accelerated using the Gauss-Seidel accelerator variant to speed up the rate of convergence of the M-ADMM algorithm. The re-formulated compact convex quadratic optimization problem is then stated thus: 1 T 1 x Px + u T Qu + R s.t Ax + B ≤ C 2 2

 ∂L p ( x , u k , λ k ) ∇ x Lρ ( x , u k , λ k ) = = 0 ∂x   ∂L p ( x k +1 , u , λ k )  k +1 k (41) = 0 KKT ⇔ ∇u L= x u λ , , ( ) ρ ∂u   k +1 k +1 k +1 k ∇ L x k +1 , u k +1 , λ = ∂L p ( x , u , λ ) = λ − λ ) λ ρ(  ∂λ ρ 

∇ x Lρ ( x , u , z , v) = Px + ρ AT Ax + ρ AT [ Bu − C + z + v]

 p1 0 0  Mn =   1≤ j ≤ n 0 

The following Karush-Kuhn-Tucker (KKT) optimality conditions in eqn. (4) below were imposed on the Augmented Lagrangian for the derivation of the ADMM algorithm.

)

−1

A T ( Bu k − C + z k + v k ) x − update

Likewise, the update of the control variable

u

is of the form:

∇u Lρ ( x , u , z , v) = Ru + ρ B Bu + ρ B [ Ax − C + z + v] T

Then at optimum, ∇u Lρ ( x yielded:

(

u k +1 = −ρ Q + ρBT B

)

−1

T

k +1

(43)

(44)

k +1 , u , z , v ) = 0 such that u = u k

k

(

BT Ax k +1 − C + z k + v k

)

u − update (45)

The update of the slack variable is of the form:

∇ z Lρ ( x , u , z , v) = ρ ( Ax + Bu − C + z + v) Then at optimum, ∇ z Lρ ( x yielded:

k +1

,u

k +1

k

, z , v ) = 0 such that

(46)

z = z k +1

(38)

(47) z k +1 =max{0, −(Ax k +1 + Bu k +1 − C + v k )} z − update Introducing the over-relaxation factor α ∈ [1.5,1.8] into the augmented lagrangian, in the sense of Nesterov [20], by replacing

x ∈ ℝ nN ,u ∈ ℝ m ( N +1) , P ∈ ℝ nN × nN ,Q ∈ ℝ m ( N +1)× m ( N +1) , A ∈ ℝ p ( N −1)× nN , B ∈ ℝ p ( N −1)× m ( N +1)

k +1 k k +1 k k k Ax k +1 with h = α Ax − (1 − α )( Bu − C + z ) in eqn. (45) yielded:

min x ,u

Where,

and C ∈ ℝ p ( N −1) and P and Q are real, symmetric and positive-definite. The associated augmented Lagrangian of eqn. (38) is then given as: 1 T 1 T T minLρ ( x , u , z , λ ) = min{ x Px + u Qu + R + l+ ( z ) + λ ( Ax + Bu − C + z ) x ,u x ,u 2 2 +

ρ

2

(39)

2 2

 Ax + Bu − C + z  } s.t. z ≥ 0,

where λ is the Lagrange multiplier, ρ>0 is the penalty parameter, ||.||2 is the euclidean (spectral) norm of a vector (matrix) argument, z is the introduced slack vector and l+ ( z ) is the indicator function for the non-negative orthants defined as l+ ( z ) = 0 for z≥0 and l+ ( z ) = +∞ otherwise. The scaled augmented Lagrangian is of the form: 1 T 1 T minLρ ( x , u , z , v) = min{ x Px + u Qu + R + l+ ( z ) + x ,u x ,u 2 2 +

ρ

2

 Ax + Bu − C + z + v 22 } s.t. z ≥ 0,

where v=λ/ρ is the scaled dual variable.

J Appl Computat Math, an open access journal ISSN: 2168-9679

u k +1 = − ρ (Q + ρ B T B ) −1 B T [α ( Ax k +1 − C + z k ) − (1 − α ) Bu k + v k ], (48)

while eqn. (47) yielded: z k +1 = max{0, − h k +1 − Bu k +1 + C − v k }.

Upon expansion and re-arrangement of eqn. (49) yielded the overrelaxed z − update z k +1 = max{0, −α ( Ax k +1 + Bu k +1 − C ) − (1 − α ) B (u k +1 − u k ) + (1 − α ) z k − v k } = max{0, −α ( Ax k +1 + Bu k +1 − C ) − (1 − α )[ B (u k +1 − u k ) − z k ] − v k }

(50)

Updating the dual variable v requires that v k +1 = v k + Ax k +1 + Bu k +1 − C + z k +1

(40)

(49)

(51)

When h k +1 was substituted in place of Ax k +1 in eqn. (51) yielded:

v k +1 = v k + h k +1 + Bu k +1 − C + z k +1

(52)

Upon expansion and re-arrangement of eqn. (52) yielded:

Volume 8 • Issue 3 • 1000445


Citation: Olotu O, Dawodu KA, Yusuf TT (2019) Algorithm for Solving Multi-Delay Optimal Control Problems Using Modified Alternating Direction Method of Multipliers. J Appl Computat Math 8: 445.

Page 6 of 9 v k +1 = v k + α ( Ax k +1 + Bu k +1 + z k +1 − C ) + (1 − α ) B(u k +1 − u k ) +(1 − α )( z k +1 − z k )

(53)

The over-relaxed M-ADMM computes the new iterates as follows:  x k +1 =   k +1 u =  z k +1 =  1  vk += 

( ) −ρ ( Q + ρB B ) B α ( Ax − C + z ) − (1 − α ) Bu + v  , max {0, −α ( Ax + Bu − C ) − (1 − α )  B ( u − u ) − z  − v } v + α ( Ax + Bu + z − C ) + (1 − α ) B ( u − u ) −ρ P + ρA T A A T ( Bu k − C + z k + v k ) x − update T

−1

T

k +1

k

k +1

k +1

k

k

k +1

k +1

k +1

k +1

k

k

k +1

k

k

(54)

k

In the development of the M-ADMM algorithm, the augmented Lagrangian functional(ALF) will be deployed because the penalty term associated with the ALF in order to improve the convergence of the M-ADMM algorithm. In literature, the convergence of ADMM algorithm to a solution, in general convex optimization problems, is guaranteed provided the solution exists. Hence the need to carry out spectral analyses regarding the spectrum, symmetry, consistency and positive-definiteness of the matrix operators to ascertain that they were well-posed for the algorithm. However, the limit of the proposed M-ADMM iterates which satisfied the set of first-order optimality conditions produced a certificate of either primal or dual feasibility or both as illustrated below. In the derivation of the convergence residues, the objective function of the multi-delay optimal control problem was assumed to be pk at the kth iteration(cycle), which converges to the optimal value p*(pk→p*) for large values of k (k→∞). Suppose the objective function is a closed, proper, convex and sub-differentiable function of f and g expressed on the form pk=f k+g k, then the residuals generated at each iteration, known as the dual residual, converges to zero. In the same light, the primal residual r k = Ax k + Bu k − C of the constraint at the kth iteration approaches zero as the algorithm approaches optimality. The derivation of the residuals is presented in theorem 5.1 below: k k k Theorem 5.1: Let p = f ( x ) + g (u ) be the k-th iterate value of the closed, proper, convex and sub-differentiable objective functions f and g such that it converges to its optimal objective value p*. Given the constraint Ax + Bu ≤ C and multiplier λ, then there exists a dual

residual d k +1 = ρ AT [ B (u k +1 − u k ) + (v k +1 − v k )] that converges to zero for a given penalty parameter ρ. k

k

Proof: Given the objective function p = f ( x ) + g (u ) and linear inequality constraint Ax + Bu ≤ C , the associated Lagrangian with slack z is stated thus: Lρ ( x , u , z , λ ) = f ( x ) + g (u ) + λ T ( Ax + Bu − C + z ) +

ρ 2

 Ax + Bu − C + z 22 ,

(55)

Applying the optimality conditions (KKT) to obtain ∂f ( x ) + A λ + ρ ( A Ax + A Bu − A C + 2 A z ) = 0 T

k

T

T

T

T

+1 ∂f ( x k +1 ) + AT λ k + ρ AT ( Ax k +1 + Buk − C + z k +1) −

ρ A Bu

− ρA z T

k +1

+ ρ A Bu + ρ A z = 0, T

k

T

k

Therefore, J Appl Computat Math, an open access journal ISSN: 2168-9679

(57)

d k +1 = ρ AT [ B (u k +1 − u k ) + ( z k +1 − z k )] → 0 which completes the proof. Q.E.D.

Theorem

5.2:

Given

a

convex

quadratic

programming

1 (QP) problem min y T Py + qT y such that Ay ≤ b , where 2 n× n n P ∈ ℝ , y ∈ ℝ , A ∈ ℝ m× n , b ∈ ℝ m , then the optimal step-size for the

QP is

ρ * =  λmin ( AP −1 AT ) λmax ( AP −1 AT ) 

 and the convergence factor

ξ R* =

−1

(58)

λmax ( AP −1 AT ) − λmin ( AP −1 AT ) λmax ( AP −1 AT )

(59)

λmax ( AP −1 AT ) + λmin ( AP −1 AT ) λmax ( AP −1 AT )

−1 T for α ∈ (0, 2], λmin ( AP −1 AT ) and λmax ( AP A ) are the minimum and −1 T maximum eigenvalues of the matrix ( AP A ) respectively, Ghadimi et al. [10].

The result of optimal parameter selection (stepsize) of the convex optimization problem stated above by Ghadimi et al. [10] can be extended to a convex optimal control problem of the form, 1 (60) Min x T Px + u T Qu + R s.t Ax + Bu ≤ C 2 nN ×1 where x ∈ ℝ , u ∈ ℝ m ( N +1)× m ( N +1) ,

P ∈ ℝ nN × nN , Q ∈ ℝ m ( N +1)× m ( N +1) , A ∈ ℝ p ( N −1)× nN , R ∈ ℝ, B ∈ ℝ p ( N −1)× m ( N +1)

and C ∈ ℝ p ( N −1)×1 . Therefore, eqn. (60) above can be re-formulated into a convex optimal control problem of the form: 1 ˆˆ+R Min wˆ T Pw 2 where,

s.t

ˆ ˆ ≤C Aw

(61)

wˆ = ( x , u ) ∈ ℝ p ( N −1)×( nN + mN + m ) , Aˆ =  A, B  ∈ ℝ p ( N −1)×( nN + mN + m ) ,

[c | c]P 0ˆ  ( nN + mN + m )×( nN + mN + m ) Pˆ =  ∈ℝ T Q   0̂

and

(

)

(

)

−1

(62)

and the convergence factor is then,

where the primal residual is given below as: r k +1 = ( Ax k +1 + Bu k +1 − C + z k +1 ).

Since at the ADMM, the update ∂f ( x k +1 ) + AT λ k +1 → 0 , then its dual residual

ˆ ˆ −1 Aˆ T λ ˆ ˆ −1 ˆ T  ρ * =  λmin AP max AP A 

r k +1

k +1

λ k +1

0̂ ∈ ℝ nN × m ( N +1) . Then, the optimal stepsize for the concatenation of the DOCP, by theorem (5.2) above, is:

∂f ( x k +1 ) + AT λ k + ρ AT ( Ax k +1 + Bu k − C + z k ) = 0

T

ρ AT [ B (u k +1 − u k ) + ( z k +1 − z k )] = ∂f ( x k +1 ) + AT [ λ k + ρ r k +1]

The convergence of the primal-dual feasibility to zero in eqn. (57) is a clear indication that the algorithm is superlinearly convergent.

Convergence analysis of the modified ADMM

k

∂f ( x k +1 ) + AT λ k + ρ AT r k +1 − ρ AT B (u k +1 − u k ) − ρ AT ( z k +1 − z k ) = 0

(56)

ξ = * Rˆ

( ˆˆ ( AP

) Aˆ ) +

( ˆˆ ( AP

) ( ˆˆ Aˆ ) λ ( AP

) Aˆ )

ˆ ˆ −1 Aˆ T − λ ˆ ˆ −1 ˆ T λ ˆ ˆ −1 ˆ T λmax AP min AP A max AP A

λmax

−1

T

λmin

−1

T

max

−1

T

(63)

Volume 8 • Issue 3 • 1000445


Citation: Olotu O, Dawodu KA, Yusuf TT (2019) Algorithm for Solving Multi-Delay Optimal Control Problems Using Modified Alternating Direction Method of Multipliers. J Appl Computat Math 8: 445.

Page 7 of 9 for α ∈ (1, 2],0 < ξ R*ˆ < 1 and P̂ is symmetric and positive definite. Dual > 0 residuals, selected as The primal ε Prim > 0 and dual ε termination (stopping) criterion for the convergence of the ADMM, k +1 Dual . However, were so small such that || r k +1 ||2 ≤ ε Prim and || d ||2 ≤ ε the choices of our tolerances depend on the relative and absolute criteria on account that the  2 norms are in ℝ n and ℝ p , respectively. Usually in practice or literature, the following values, ε rel = 10−3 and ε abs = 10−4 , are used as reasonable stopping criteria for the ADMM algorithms. Stated below are basic computations in literature as in Boyd et al. [7].

ε Prim = pε abs + ε rel .max{|| Ax k +1 ||2 ,|| Bu k +1 ||2 ,|| z k +1 ||2 ,|| C ||2 } (64)

ε Dual = nε abs + ε rel . || ρ u |2

(65)

The formulated algorithm: Algorithm 3.3: Modified-ADMM For multiple DOCP,

Step 2 : Set k = 0 Step 3 : Compute x k +1, i = − ρ ( P + ρ AT A) −1 AT ( Bu k , i − C + z k , i + v k , i ) eqn. (43)

Step 4 : Stop if || x k , i +1 − x k , i ||≤ Tol. Step 5 : Compute u k , i +1 = − ρ (Q + ρ BT B ) −1 BT [α ( Ax k +1, i − C + z k , i ) − (1 − α ) Bu k , i + v k , i ]

Step 6 : Stop if || u Step 7 : Output x

k +1

−u

,u

k +1

eqn. (48)

||≤ Tol.

and go to step 9 otherwise

Step 8 : Repeat steps 3 & 5 for i = 1, 2, until steps 4 & 6 are satisfied Step 9 : Compute z k +1 = Max{0, −α ( Ax k +1 + Bu k +1 − C ) + (1 − α )[ B (u k +1 − u k ) − z k ] − v k } eqn. (50)

v k +1 = v k + α ( Ax k +1 + Bu k +1 − C + z k +1 ) + (1 − α ) B (u k +1 − u k ) + (1 − α )( z k +1 − z k )

eqn. (53)

k +1 k +1 k +1 k Prim Step 10 : Stop if || r ||2 = (Ax + Bu − C + z ) ≤ ε and eqn. (56)

. || d k +1 ||2 = ρA T [B(u k +1 − u k ) + (z k +1 − z k )] ≤ ε Dual else go to step 12 eqn. (57) k +1 k +1 Step 11 : Output x , u and Compute

ρ * (optimal steplength) eqn. (58) ξ R* (Convergence profile) eqn. (59) Step 12 : Set k = k + 1 and go to Step 2.

Results and Discussion The multi-delay optimal control problem below was considered for the implementation of the algorithm: MinJ ( x, u ) =

1 0.5 2 [2 x1 + x1 x2 + x22 + u12 + u1u2 + u22 ]dt 2 ∫0

J Appl Computat Math, an open access journal ISSN: 2168-9679

x2 (t − 0.2) − x1 (t − 0.3) + u1 (t − 0.1) + u2 (t − 0.1) + u1 (t − 0.3) + 2u2 (t − 0.2),

x2 (t ) = x1 (t ) − u1 (t ) + 2u2 (t ) − x1 (t − 0.1) + x1 (t − 0.2) + 2 x2 (t − 0.2) − x1 (t − 0.3) + u2 (t − 0.1) + u1 (t − 0.2) + 3u2 (t − 0.2),

x(0) = (0,0);

0 ≤ t ≤ 0.5, 2

x(t ) = (2t + 1, t + 1);

− 0.3 ≤ t ≤ 0,

u (t ) = (2, 2 + t );

− 0.2 ≤ t ≤ 0,

2 where x(t ) ∈ ℝ and u (t ) ∈ ℝ 2

In adopting the M-ADMM algorithm, the coefficient matrices are defined below as: 4 1 2 1 2 1  1 3  1 1 P= ,Q =  ,A= ,B =  , α1 =      , 1 2 1 2 1 0  −1 2   −1 0 

2 1 

Step 1 : Initialize x 0,0 , u 0,0 , λ 0,0 ,V 0,0

k ,i

x1 (t ) = 2 x1 (t ) + x2 (t ) + u1 (t ) + 3u2 (t ) + x1 (t − 0.1) + x2 (t − 0.1) + 2 x1 (t − 0.2) +

 −1 0 

1 1

1 2 

α2 =   , α 3 =  −1 0  , β1 = 0 1 and β 2 = 1 3  . 1 2       

Prim Dual Step 0 : Input A, B,C , P,Q, R, ρ ,α ,Tol.,ε ,ε

k , i +1

s.t

The coefficient matrices, P and Q , of the quadratic functional are symmetric, invertible (non-singular) and positive definite (i.e., P  0 and Q  0 ) since their respective eigenvalues (λ1 = 1.585, λ2 = 4.4142) and (λ1 = 1, λ2 = 3) are all positive. This ensures that the operators are well-posed for the algorithm. However, the algorithm runs for various numbers of iterations plotted against 29 values of the relaxation parameters α evenly spaced between 0.85 and 1.13. At each iteration (cycle), the penalty parameter ρ was kept fixed at its optimum ρ * = 0.2743 , based on the step-size heuristically computed from eqn. (58). As shown in Figure 1A, 5 values of the dual and primal tolerances 10-k for k=4,5,..8 geometrically spaced between 10-4 and 10-8. Each line on the graph corresponds to a fixed value of the dual and primal tolerance (10-k), where the relaxation factor runs over the entire chosen feasible values α∊[0,2]. Clearly, each line converges at the point α * = 1.05 , which demonstrates the optimum over-relaxation factor for reducing values of α, though few values of k were chosen to keep the plot manageable and the cycle of the algorithm not exceeding 200 iterations Figure 1B illustrates the convergence of the M-ADMM iterations for the Quadratic program at the optimum step-size ρ* and over-relaxation factor α ∈ [0, 2] . It demonstrates the responses of the dual and primal residuals for various iteration counters at the optimum relaxation factor α * = 1.05 and at the optimum penalty parameter of ρ * = 0.2743 . When comparing the convergence of the dual and primal residuals at start, the dual residual rk is far lower than that the primal residual dk which gives it high propensity to satisfy the feasibility conditions than that of the primal. However, the rate of convergence of the primal residual towards zero for large iteration ( || d k ||22 → 0 as k → ∞ ) is faster than that of the dual as demonstrated in Figure 1B. Figure 2A demonstrates the responses of the state-control trajectories and the objective values to increasing iterations at chosen tolerance, fixed relaxation factor and optimal penalty parameter ( 10− k ,α * , ρ * ) = (10−8 ,1.05,0.2743 ). However, it was observed that the objective values increase with increasing values of the state and control * variables until the optimum objective value p = 0.93 was arrived * * * w = ( x , u ) = (2.8058,0.2356) . at with the optimum trajectories Moreover, Figure 2B is the feedback law which demonstrates the degree of relationship between the state x(t) and control u(t) trajectories within the specified time domain [0,0.5]. The control (feedback) law is a time-variant measure unlike the proportional control system where the feedback law is time-invariant (real constant). Volume 8 • Issue 3 • 1000445


Citation: Olotu O, Dawodu KA, Yusuf TT (2019) Algorithm for Solving Multi-Delay Optimal Control Problems Using Modified Alternating Direction Method of Multipliers. J Appl Computat Math 8: 445.

Page 8 of 9

(b) Effects of Primal-Dual on convergence (a) Response of Algorithm to Relaxation factor Figure 1: The operators for algorithm.

(a) Optimal state-control trajectories

(b) Feedback law

Figure 2: The optimum parameters.

Conclusion and Future Work The solution to optimal control problem with multiple constant delays using the modified alternating direction method have been established. The convergence of the algorithm was discovered to be at a superlinear rate and provides an explicit expressions for the optimum convergence parameter. We also considered the over-relaxation of the M-ADMM, for various values of the penalty parameters, for which the algorithm is guaranteed to convergence at its optimum. Numerical example was given to establish the effectiveness and general performance of the algorithm in terms of its lower iteration cycle and time in obtaining the optimal solution to the multi-delay control problem. In the future, the algorithm can be extended to other classes of control problems with objective functional of the Bolza-type constrained by bounded or mixed inequality constraints.

2. Bakke VL (1981) Optimal Fields for Problems with Delays. Opt Theo and Appl 33: 567-593. 3. Bashier EBM, Patider KC (2017) Optimal Control of an Epidemiological Model with Multiple Time Delay. Journal of Applied Mathematics and Computation 292 :47-56. 4. Bertsekas DP (1982) Constrained Optimization and Lagrange Multiplier methods (1stedn), Academic press, Inc., Belmont, USA. 5. Boley D (2013) Local Linear Convergence of the Alternating Direction Method of Multipliers on Quadratic or Linear Programs. SIAM Journal on Optimization 23: 2183-2207. 6. Boyd S, Vandenberghe L (2004) Convex Optimization, New York, USA, Cambridge University Press. 7. Boyd S, Parikh N,Chu E, Peleato B, Eckstein J (2011) Distributed optimization and statistical learning via the alternating direction method of multipliers. Foundations and Trends in Machine Learning 3: 11-22.

References

8. Chyung DH, Lee EB (1966) Linear Optimal Systems with Time Delays. SIAM J Control 4: 548-575.

1. Angell TS, Kirsch A (1990) On the necessary conditions for optimal control of retarded systems. Appl Math and Optimization 22:117-145.

9. Colonius F, Hinrichsen D (1978) Optimal Control of Functional Differential Systems. SIAM Journal on Control and Optimization 16: 861-879.

J Appl Computat Math, an open access journal ISSN: 2168-9679

Volume 8 • Issue 3 • 1000445


Citation: Olotu O, Dawodu KA, Yusuf TT (2019) Algorithm for Solving Multi-Delay Optimal Control Problems Using Modified Alternating Direction Method of Multipliers. J Appl Computat Math 8: 445.

Page 9 of 9 10. Ghadimi E, Teixeira A, Shames I, Johansson M (2014) Optimal Parameter Selection for the Alternating Direction Method of Multipliers (ADMM): Quadratic Problems. IEEE Transactions on Automatic Control 60(3): 644-658. 11. Gilbert GT (1991) Positive Definite Matrices and Sylvester’s Criterion. The American Mathematical Monthly 98: 44-46. 12. Goldfarb D, Schneiberg K (2012) Fast Alternating Linearization Methods for Minimizing the Sum of Two Convex Functions. Mathematical Programming 141: 349-382. 13. Gollman LD, Kern D, Maurer H (2008) Optimal Control Problems with Delays in State and Control Variables Subject to Mixed Control-State Constraints. Optimal Control Applications and Methods 30: 341-365. 14. Gollman LD, Maurer H (2014) Theory and Applications of Optimal Control Problems with Multiple Delays. Journal of Industrial and Management Optimization 10: 413-441.

Delays. In: Balakrishnan AV, Neustadt LW (eds.), Mathematical Theory of Control; Academic Press, New York, USA. 19. Laarabi H, Abta A, Hahaf K (2015) Optimal Control of a Delayed SIRS Epidemic Model with Vaccination and Treatment. Acta Biotheor 63: 87-97. 20. Nesterov YE (1983) A method for Solving the Convex Programming Problem with Convergence Rate O(1/k2 ). Dokl Akad Nauk SSSR 269: 543-547. 21. O’Donoghue B, Stathopoulos G, Boyd S (2013) A Splitting Method for Optimal Control. IEEE Trans. Control Systems Technology 21: 2432-2442. 22. Olotu O, Dawodu KA, Adebowale K (2013) On the Discretized Algorithm for Optimal Proportional Control Constrained by Delay Differential Equation. Journal of Mathematical Theory and Modelling by IISTE, USA 3, pp: 157-169. 23. Rihan FA, Anwar MN (2012) Qualitative Analysis of Delayed SIR Epidemic Model with a Saturated Incidence rate. Intl J Diff Equation 408637:1-13.

15. Harastivilli GL (1961)The Principle in the Theory of Procces Involving Delay. Dokl Akad, Nauk SSSR 136, pp: 39-42.

24. Soliman MA (1972) Optimal Feedback Control for Linear-Quadratic Systems having Time Delays. Int J Control 15: 609-627.

16. Kadkhodaie M, Christakopoulou K, Sanjabi M, Banerjee A (2015) Accelerated alternating direction method of multipliers. In: Proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining, ACM, pp: 497-506.

25. Xue Y, Duan X (2011) Dynamic Analysis of an SIR Epidemic Model with Nonlinear Incidence rate and Double Delays. Intl J Inf Syst Sci Computer Inf 7: 92-102.

17. Kharatishvili GL (1961) The maximum principle in the theory of optimal processes involving delay. Doklady Akademii Nauk USSR 136: 39-42. 18. Kharastishvilli GL (1967) A Maximum Principle in External Problems with

J Appl Computat Math, an open access journal ISSN: 2168-9679

26. Yang J, Zhang Y (2011) An Alternating Direction Algorithms for l1 Problems in Compressive Sensing. SIAM J Sci Comput 33: 250-278. 27. Zhang X, Berger M, Osher M (2011) A Unified Primal-Dual Algorithm Framework Based on Bregman Iteration. Journal of Scientific Computing 1: 20-46.

Volume 8 • Issue 3 • 1000445


Issuu converts static files into: digital portfolios, online yearbooks, online catalogs, digital photo albums and more. Sign up and create your flipbook.