International Mathematical Forum, 2, 2007, no. 30, 1477 - 1498 Classical and Relaxed Optimization Methods for Optimal Control Problems I. Chryssoverghi, J. Coletsos and B. Kokkinis Department of Mathematics, School of Appl. Mathematics and Physics National Technical University of Athens Zografou Campus, 15780 Athens, Greece [email protected] Abstract. We consider an optimal control problem for systems governed by nonlinear ordinary differential equations, with control and state constraints, including pointwise state constraints. The problem is formulated in the classical and in the relaxed form. Various necessary/sufficient conditions for optimality are first given for both problems. For the numerical solution of these problems, we then propose a penalized gradient projection method generating classical controls, and a penalized conditional descent method generating relaxed controls. Using also relaxation theory, we study the behavior in the limit of sequences constructed by these methods. Finally, numerical examples are given. Keywords: Optimal control, nonlinear systems, state constraints, gradient projection method, conditional descent method, penalty method, relaxed controls. Mathematics Subject Classification: 49M 1 Introduction We consider an optimal control problem for systems governed by nonlinear ordinary differential equations, with control and state constraints, including pointwise state constraints. The problem is formulated in the classical form, and also in the relaxed form using Young measures. Various necessary/sufficient conditions for optimality are first given for both problems. For the numerical solution of these problems, we then propose a penalized gradient projection method generating classical controls, and a penalized conditional descent method generating relaxed controls. Under appropriate assumptions, we prove that relaxed (resp. strong classical) limits of subsequences (resp. sequences) constructed by the classical method are admissible and weakly extremal relaxed (resp. classical) for the relaxed (resp. classical) problem, and that relaxed limits of subsequences of controls constructed by the relaxed method are admissible and strongly extremal for the relaxed problem. Finally, several numerical examples are given. For classical and relaxed optimization and 1478 I. Chryssoverghi, J. Coletsos and B. Kokkinis approximation methods applied to optimal control problems, see e.g. [2-9], [11-12], [14], and the references therein. 2 Classical and relaxed optimal control problems Consider the following optimal control problem. The state equation is given by y '(t ) = f (t , y (t ), w(t )) , t ∈ I := [0, T ] , y (0) = y 0 , where y (t ) ∈ IR d , the constraints on the control w are w(t ) ∈ U , for t ∈ I , where U is a compact subset of IR d ' , the constraints on the state y := yw are T G1 ( w) = g1 ( y (T )) + ∫ g1 (t , y (t ), w(t ))dt = 0 , 0 T G2 ( w) = g 2 ( y (T )) + ∫ g 2 (t , y (t ), w(t ))dt ≤ 0 , 0 G3 ( w)( s ) = g3 ( s, y ( s )) ≤ 0 , for s ∈ I , where the vector functions gl , gl take values in IR ml , l = 1, 2 , and g3 in IR m3 , and the cost functional is T G0 ( w) = g 0 ( y (T )) + ∫ g 0 (t , y (t ), w(t ))dt . 0 Defining the set of classical controls W = {w : I → U w measurable} ⊂ L2 ( I , IR d ' ) , the classical optimal control problem is to minimize G0 ( w) subject to w ∈ W and to the above state constraints. It is well known that, even if the set U is convex, the classical problem may have no solutions. The existence of such a solution is usually proved under strong, often unrealistic for nonlinear systems, convexity assumptions (such as the Cesari property). Reformulated in the so-called relaxed form, the problem is convexified in some sense and has a solution in a larger space under weaker assumptions. Next, we define the set of relaxed controls (Young measures; for the relevant theory, see [13], [10]) by R = {r : I → M 1 (U ) r weakly measurable} ⊂ L∞w ( I , M (U )) ≡ L1 ( I , C (U )) * , where M (U ) (resp. M 1 (U ) ) is the set of Radon (resp. probability) measures on U . The set W (resp. R ) is endowed with the relative strong (resp. weak star) topology, and R is convex, metrizable and compact. If each classical control w(⋅) is identified with its associated Dirac relaxed control r (⋅) := δ w(⋅) , then W may be also considered as a subset of R , and W is thus dense in R . For a given φ ∈ L1 ( I ; C (U ; IR n )) (or equivalently φ ∈ B( I ,U ; IR n ) , where B is the set of Caratheodory functions in the sense of Warga [13]) and r ∈ L∞w ( I , M (U )) (in particular, for r ∈ R ), we shall use for simplicity the notation φ ( x, r (t )) := ∫ φ (t , u )r (t )(du ), U and φ (t , r (t )) is thus linear (under convex combinations, for r ∈ R ) in r . A sequence (rk ) converges to r ∈ R in R if and only if Optimal Control Problems 1479 lim ∫ φ (t , rk (t ))dt = ∫ φ (t , r (t ))dt , k →∞ I I for every φ ∈ L ( I ; C (U ; IR )) , or φ ∈ B( I , U ; IR n ) , or φ ∈ C ( I × U ; IR n ) . The relaxed optimal control problem is then defined by replacing w by r (with the above notation) and W by R in the classical problem. 1 n p p We define the norms x = ( ∑ xi2 )1/ 2 , x 1 = ∑ xi , x i =1 and denote by ⋅ L2 , ⋅ L1 , ⋅ L∞ , ⋅ i =1 L∞ ∞ = max xi , in IR p , i =1,..., p the corresponding usual norms in L2 ( I ) p , L1 ( I ) p , L∞ ( I ) p , C ( I ) p , respectively. We denote by M ( I ) ≡ C ( I ) * the set of finite regular measures on I , and by ⋅ * the norm in M ( I ) defined by μ * = ∫ μ (dt ) I (with μ = μ if μ is positive). The order relations between vectors, functions or vector functions, are defined componentwise and/or pointwise. We suppose that the function f is defined on I × IR d × U , measurable for y , u fixed, continuous for t fixed, and satisfies f (t , y, u ) ≤ ψ (t )(1 + y ), for every (t , y, u ) ∈ I × IR d × U , with ψ ∈ L1 ( I ) , f (t , y1 , u ) − f (t , y2 , u ) ≤ L y1 − y2 , for every (t , y1 , y2 , u ) ∈ I × IR 2 d × U . The following result is standard (see [13]). Theorem 2.1 For every relaxed (or classical, as W ⊂ R ) control r ∈ R , the state equation has a unique absolutely continuous solution y := yr . Moreover, there exists a constant b such that yr ∞ ≤ b for every r ∈ R . Let B denote the closed ball in IR d with center 0 and radius b (see Theorem 2.1). We suppose now in addition that the functions gl , l = 0,1, 2 , are defined on I × B × U , measurable for fixed y , u , continuous for fixed t , and satisfy g l (t , y , u ) ≤ ζ l (t ), for every (t , y, u ) ∈ I × B × U , with ζ l ∈ L1 ( I ) , the function g3 is continuous on I × B , and that the functions gl are continuous on B . The results of the following theorem are proved in [13]. Theorem 2.2 The mappings y → yw , from L2 to C ( I ) d , y → yr , from R to C ( I ) d , and Gl : W or R → R ml , l = 0,1, 2 , G3 : W or R → C ( I ) m3 , are continuous. If the relaxed problem is feasible, then it has a solution. Note that in the classical problem we have y '(t ) ∈ f (t , y (t ),U ) (velocity set), while in the relaxed problem y '(t ) ∈ co(f (t , y (t ),U )) . Since W ⊂ R , we have in general cR := min G0 ( r ) ≤ inf G0 ( w) := cW , constraints on r constraints on w where the equality holds, in particular, if there are no state constraints, as W is dense in R . Since usually approximation methods slightly violate the state constraints, approximating an optimal relaxed control by a relaxed or a classical control, hence the 1480 I. Chryssoverghi, J. Coletsos and B. Kokkinis relaxed optimal cost cR , is not a drawback in practice (see [13], p. 248). Note also that approximating sequences of classical controls may converge to relaxed ones. In order to state the necessary conditions for optimality, we suppose in addition that the functions f , gl , f y , f u , gly , glu , l = 0,1, 2 , are defined on I × B '× U ' , where B ' (resp. U ' ) is a open set containing B (resp. U ), measurable on I for fixed ( y, u ) ∈ B × U , continuous on B × U for fixed t ∈ I , and satisfy f y (t , y, u ) ≤ ξ (t ) , f u (t , y, u ) ≤ η (t ) , gly (t , y , u ) ≤ ζ l1 (t ), g lu (t , y , u ) ≤ ζ l 2 (t ), for every (t , y, u ) ∈ I × B × U , with ξ ,η , ζ l1 , ζ l 2 ∈ L1 ( I ) , and that the functions gly are defined on B ' and continuous on B . The two following theorems can be proved by using the techniques of [13]. Theorem 2.3 (i) If U is convex, the directional derivative of the mapping Gl : W → R ml , for l = 0,1, 2 , is given by G ( w + α ( w '− w)) − Gl ( w) DGl ( w, w '− w) = lim+ l α α →0 T = ∫ [ zl (t ) f u (t , y (t ), w(t )) + glu (t , y (t ), w(t ))][ w '(t ) − w(t )]dt , for w, w ' ∈ W , 0 where y := yw , and the adjoint state zl := zl w , a row vector function ( l = 0 ), or a matrix function ( l = 1, 2 ), is the solution of the classical linear adjoint equation zl '(t ) = − zl (t ) f y (t , y (t ), w(t )) − gly (t , y (t ), w(t )) , t ∈ I , zl (T ) = gly ( y (T )) , with y := yw , the controls being regarded here as classical. The directional derivative of G3 : W → C ( I ) m3 is given by the matrix function DG3 ( w, w '− w)( s) s = g 3 y ( s, y ( s )) Z ( s ) −1 ∫ Z (t ) f u (t , y (t ), w(t ))[ w '(t ) − w(t )]dt , s ∈ I , 0 where the matrix function Z := Z w satisfies the fundamental matrix equation Z '(t ) = − Z (t ) f y (t , y (t ), w(t )) , t ∈ I , Z (T ) = E ( E identity matrix). (ii) The directional derivative of the mapping Gl : R → IR ml , for l = 0,1, 2 , is given by G (r + α (r '− r )) − Gl (r ) DGl (r , r '− r ) := lim+ l α →0 α = ∫ [ zl (t ) f (t , y (t ), r '(t ) − r (t )) + gl (t , y (t ), r '(t ) − r (t ))]dt , for r, r ' ∈ R , I where y = yr , and the relaxed adjoint zl := zl r is the solution of the relaxed linear adjoint equation zl '(t ) = − zl (t ) f y (t , y (t ), r (t )) − gly (t , y (t ), r (t )) , t ∈ I , zl (T ) = g ly ( y (T )) , with y := yr . The directional derivative of G3 : R → C ( I ) m3 is given by 1481 Optimal Control Problems s DG3 ( r, r '− r )( s ) = g3 y ( s, y ( s )) Z ( s ) −1 ∫ Z (t ) f u (t , y (t ), r '(t ) − r (t ))dt , s ∈ I , 0 where Z := Z r is defined as in (i), but with w replaced by r . (iii) The following mappings are continuous z a zw , from W to C ( I ) d , z a zr , from R to C ( I ) d , ( w, w ') a DGl ( w, w '− w) , l = 0,1, 2,3 , from W × W to IR ml , l = 0,1, 2 , C ( I ) m3 , ( r, r ') a DGl ( r, r '− r ) ), l = 0,1, 2,3 , from R × R to IR ml , l = 0,1, 2 , C ( I ) m3 . In the notations of DG , it is understood, depending on the arguments, w or r , that the directional derivative is taken in the corresponding space, W or R , on which G is defined. The following theorem gives various necessary conditions for optimality (the weak relaxed minimum principle is proved similarly to [7]). Theorem 2.4 (i) We suppose that U is convex. If w ∈ W is optimal for the classical problem, then w is weakly extremal classical, i.e. there exist multipliers λ0 ∈ IR , λ1 ∈ IR m1 , λ2 ∈ IR m2 , λ3 ∈ [C ( I ) m3 ]* ≡ M ( I ) m3 , with λ0 ≥ 0 , λ2 ≥ 0 , λ3 ≥ 0 , 2 ∑ l =0 m3 λl + λ3 * = 1 , where λ3 * = ∑ λ j 3 j =1 M1 , such that 2 ∑ λ DG (w, w '− w) + ∫ l l =0 l 2 T 0 λ3 (ds ) DG3 ( w, w '− w)( s) = ∑ λl ∫ [ zl (t ) f u (t , y (t ), w(t )) + glu (t , y (t ), w(t ))][ w '(t ) − w(t )]dt T 0 l =0 T s + ∫ λ3 ( ds ) g 3 y ( s, y ( s )) Z ( s ) −1 ∫ Z (t ) f u (t , y (t ), w(t ))[ w '(t ) − w(t )]dt 0 0 2 = ∫ {∑ λl [ zl (t ) f u (t , y (t ), w(t )) + glu (t , y (t ), w(t ))] T 0 + l =0 ( ∫ λ (ds) g T t 3 3y ) ( s, y ( s )) Z ( s ) −1 Z (t ) fu (t , y (t ), w(t ))}[ w '(t ) − w(t )]dt ≥ 0 , for every w ' ∈ W , and λ2G2 ( w) = 0 , ∫ T 0 λ3 ( ds )G3 ( w)( s ) = 0 (classical transversality conditions). The above inequality condition is equivalent to the pointwise weak classical minimum principle 2 {∑ ( λl [ zl (t ) f u (t , y (t ), w(t )) + glu (t , y (t ), w(t ))] l =0 T +[ ∫ λ3 ( ds ) g 3 y ( s, y ( s )) Z ( s ) −1 ]Z (t ) f u (t , y (t ), w(t ))} w(t ) t 2 = min{{∑ λl [ zl (t ) f u (t , y (t ), w(t )) + glu (t , y (t ), w(t ))] u∈U T l =0 +[ ∫ λ3 ( ds ) g 3 y ( s, y ( s )) Z ( s ) −1 ]Z (t ) f u (t , y (t ), w(t ))} u} , for a.a. t ∈ I . t 1482 I. Chryssoverghi, J. Coletsos and B. Kokkinis (ii) If r ∈ R is optimal for either the relaxed or the classical problem, then r is strongly extremal relaxed, i.e. there exist multipliers as in (i), such that 2 ∑ λ DG (r, r '− r ) + ∫ l l 0 l =0 2 T λ3 (ds ) DG3 ( r, r '− r )( s ) = ∑ λl ∫ [ zl (t ) f (t , y (t ), r '(t ) − r (t )) + gl (t , y (t ), r '(t ) − r (t ))]dt T 0 l =0 T s + ∫ λ3 ( ds ) g 3 y ( s, y ( s )) Z ( s ) −1 ∫ Z (t ) f (t , y (t ), r '(t ) − r (t ))dt 0 0 2 = ∫ {∑ λl [ zl (t ) f (t , y (t ), r '(t ) − r (t )) + gl (t , y (t ), r '(t ) − r (t ))] T 0 l =0 T +[ ∫ λ3 ( ds ) g 3 y ( s, y ( s )) Z ( s ) −1 ]Z (t ) f (t , y (t ), r '(t ) − r (t ))}dt ≥ 0 , t for every r ' ∈ R, and λ2G2 (r ) = 0 , ∫ T 0 λ3 ( ds )G3 ( r )( s ) = 0 (relaxed transversality conditions). The above inequality condition is equivalent to the pointwise strong relaxed minimum principle 2 ∑ λ [ z (t ) f (t, y (t ), r(t )) + g (t, y(t ), r(t ))] l l l l =0 T +[ ∫ λ3 ( ds ) g 3 y ( s, y ( s )) Z ( s ) −1 ]Z (t ) f (t , y (t ), r (t )) t 2 = min{∑ λl [ zl (t ) f (t , y (t ), u ) + gl (t , y (t ), u )] u∈U l =0 T +[ ∫ λ3 ( ds ) g 3 y ( s, y ( s )) Z ( s ) −1 ]Z (t ) f (t , y (t ), u )} , for a.a. t ∈ I . t If in addition U is convex, then this minimum principle implies the pointwise weak relaxed minimum principle 2 {∑ λl [ zl (t ) f u (t , y (t ), r (t )) + glu (t , y (t ), r (t ))] l =0 T +[ ∫ λ3 ( ds ) g 3 y ( s, y ( s )) Z ( s ) −1 ]Z (t ) f u (t , y (t ), r (t ))} r (t ) t 2 = min{{∑ λl [ zl (t ) f u (t , y (t ), r (t )) + glu (t , y (t ), r (t ))] φ l =0 T +[ ∫ λ3 ( ds ) g 3 y ( s, y ( s )) Z ( s ) −1 ]Z (t ) f u (t , y (t ), r (t ))} φ (t , r (t ))} , for a.a. t ∈ I , t where the minimum is taken over the set B( I ,U ;U ) of Caratheodory functions (see [13]) φ : I × U → U , which in turn implies the global weak relaxed condition 2 ∫ {∑ λ [ z (t ) f (t, y(t ), r(t )) + g T 0 l l u lu (t , y (t ), r (t ))] l =0 T +[ ∫ λ3 ( ds ) g 3 y ( s, y ( s )) Z ( s ) −1 ]Z (t ) f u (t , y (t ), r (t ))}[φ (t , r (t )) − r (t )]dt ≥ 0 , t for every φ ∈ B( I ,U ;U ) . A control r satisfying this condition and the above relaxed transversality conditions is called weakly extremal relaxed. 1483 Optimal Control Problems The following theorem gives sufficient conditions for optimality. Theorem 2.5 With the derivatives in u omitted (resp. included) in our last assumptions, we suppose in addition that the data are such that G0 , G2 , G3 are convex and that G1 is affine. If r ∈ R (resp. w ∈ W , with U convex) is admissible and strongly extremal relaxed (resp. weakly extremal classical) for the relaxed (resp. classical) problem, with λ0 > 0 , then r is optimal for this problem. Proof. (Relaxed case, the classical case is similar) The assumptions imply that the functional 2 G (r ) := ∑ λl Gl (r ) + ∫ λ3 (ds )G3 (r )( s ) l =0 I is convex. The necessary inequality condition of Theorem 2.4 is then satisfied if and only if r minimizes G on R . Suppose now that r does not minimize G0 , in which case there exists r ' ∈ R satisfying the state constraints and such that G0 (r ') < G0 (r ) . Using the state constraints and the relaxed transversality conditions, we obtain 2 G (r ') := ∑ λl Gl (r ') + ∫ λ3 (ds )G3 (r ')( s ) l =0 I 2 ≤ λ0G0 (r ') < λ0G0 (r ) = ∑ λl Gl (r ) + ∫ λ3 (ds )G3 (r )( s ) = G (r ) , l =0 I i.e. r does not minimize G , a contradiction. 3 Classical and relaxed optimization methods Let ( M lm ) , l = 1, 2,3 , be nonnegative increasing sequences such that M lm → ∞ as m → ∞ , γ > 0 , b, c ∈ (0,1) , and ( β k ) , (ζ k ) positive sequences, with ( β m ) decreasing and converging to zero, and ζ k ≤ 1 . Define first the penalized functionals on W m1 m2 2 1 G m ( w) := G0 ( w) + {M 1m ∑ G1 j ( w) + M 2m ∑ [max(0, G2 j ( w))]2 2 j =1 j =1 m3 + M 3m ∑ ∫ [max(0, G3 j ( w)( s))]2 ds} . j =1 I It can be easily shown that the directional derivative of G m is given by DG m ( w, w '− w) = DG0 ( w, w '− w) m1 m2 j =1 j =1 + M 1m ∑ G1 j ( w) DG1 j ( w, w '− w) + M 2m ∑ max(0, G2 j ( w)) DG2 j ( w, w '− w) m3 + M 3m ∑ ∫ max(0, G3 j ( w)( s )) DG3nj ( w, w '− w)( s )ds . j =1 I The classical penalized gradient projection method is described by the following Algorithm, where U is assumed to be convex. 1484 I. Chryssoverghi, J. Coletsos and B. Kokkinis Algorithm 1 Step 1. Set k := 0 , m := 1 , and choose an initial control w01 ∈ W . Step 2. Find vkm ∈ W such that ek := DG m ( wkm , vkm − wkm ) + (γ / 2) vkm − wkm 2 L2 2 m k L2 = min[ DG m ( wkm , v '− wkm ) + (γ / 2) v '− w v '∈W ], and set d k := DG m ( wkm , vkm − wkm ) . Step 3. If d k ≤ β m , set wm := wkm , v m := vkm , em := ek , d m := d k , m := m + 1 , and then go to Step 2. Step 4. (Modified Armijo step search) Find the lowest integer value s (positive or not), say s , such that α = c sζ k ∈ (0,1] and α satisfies the inequality G m ( wkm + α k (vkm − wkm )) − G m ( wkm ) ≤ α k bd k , and then set α := c s ζ k . Step 5. Set wkm+1 := wkm + α k (vkm − wkm ) , k := k + 1 , and go to Step 2. One can see by “completing the square” that Step 2 amounts to finding the projection vkl of the function m2 ukm (t ) := wkm (t ) − (1/ γ ){∑ λlm [ zlkm (t ) fu (t , ykm (t ), wkm (t )) + glu (t , ykm (t ), wkm (t ))] + ( ∫ λ (ds) g T t m 3 l =0 3y ) ( s, ykm ( s )) Z ( s ) −1 Z (t ) f u (t , ykm (t ), wkm (t ))} , onto W , which in turn reduces to finding the corresponding pointwise projection onto U for a.a. t ∈ I . On the other hand, by the definition of the directional derivative and since b, c ∈ (0,1) , clearly the Armijo step α k in Step 4 can be found for every k . The parameter γ is chosen experimentally to yield a good rate of convergence. A (strongly or weakly, classical or relaxed) extremal control is called abnormal if there exist multipliers as in the corresponding optimality conditions, with λ0 = 0 . A control is admissible and abnormal extremal in rather exceptional situations (see [13]). With wm as defined in Step 3, define the sequences of multipliers λ1m := M 1mG1 ( wm ), λ2m := M 2mmax(0, G2 ( wm )), λ3m ( s ) := M 3mmax(0, G3 ( wm )( s )) = M 3mmax(0, g3 ( s, y m ( s ))) , where max denotes a vector of maximum values. Theorem 3.1 We suppose here that U is convex. (i) In the presence of state constraints, if the whole sequence ( wkm ( k ) ) generated by Algorithm 1 converges to some w ∈ W in L2 strongly and the sequences (λlm ) , l = 1, 2,3 , (λ3m ) in L1 ( I )m3 , are bounded, then w is admissible and weakly extremal classical for the classical problem. In the absence of state constraints, if a subsequence 1485 Optimal Control Problems ( wk ) k∈K (no index m ) converges to some w ∈ W in L2 strongly, then w is weakly extremal classical for the classical problem. (ii) In the presence of state constraints, if a subsequence ( wm ) m∈M (regarded as a sequence in R ) of the sequence generated by Algorithm 1 in Step 3 converges to some r in R and the sequences ( λlnm ) , l = 1, 2,3 , ( λ3nm ) in L1 ( I )m3 , are bounded, then r is admissible and weakly extremal relaxed for the relaxed problem. In the absence of state constraints, if a subsequence ( wk ) k∈K (no index m ) converges to some r in R , then r is weakly extremal relaxed for the relaxed problem. (iii) In any of the convergence cases (i) or (ii) with state constraints, suppose that the classical, or relaxed, problem has no admissible, abnormal extremal, controls. If the limit control is admissible, then the sequences of multipliers are bounded, and this control is extremal as above. Proof. (State constraints present) We shall first show that m → ∞ in the Algorithm. Suppose on the contrary that m remains constant after a finite number of iterations in k , and so we drop here the index m . Consider case (i). Let us show that then d k → 0 . Since wk → w in L2 strongly, using also Proposition 2.1 in [1], we can deduce that uk → u in L2 strongly, where 2 u (t ) := w(t ) − (1/ γ ){∑ λl [ zl (t ) fu (t , y (t ), w(t )) + glu (t , y (t ), w(t ))] + ( ∫ λ (ds) g l =0 T t 3 3y ) ( s, y ( s )) Z ( s ) −1 Z (t ) fu (t , y (t ), w(t ))} . Since vk is the projection of uk onto W , we have also vk → v in L2 strongly, where v is the projection of u onto W . Clearly, by Step 2, d k ≤ ek ≤ 0 for every k , hence, by Theorem 2.3 2 e := lim ek = DG ( w, v − w) + (γ / 2) v − w ≤ 0, k →∞ d := lim d k = DG ( w, v − w) ≤ lim ek = e ≤ 0. k →∞ k →∞ Suppose that d < 0 . The function Φ(α ) := G ( w + α (v − w)) is continuous on [0,1] , and since the directional derivative DG ( w, v − w) is linear in v − w , Φ is differentiable on (0,1) and has derivative Φ '(α ) = DG ( w + α (v − w), v − w). By the mean value theorem, for each α ∈ (0,1] there exists α '(α ) ∈ (0, α ) such that G ( wk + α (vk − wk )) − G ( wk ) = α DG ( wk + α '(α )(vk − wk ), vk − wk ) . Therefore, by Theorem 2.3 G ( wk + α ( vk − wk )) − G ( wk ) = α ( d + ε kα ), for α ∈ [0,1] , where ε kα → 0 as k → ∞ , k ∈ K , and α → 0+ . Now, we have d k = d + ηk , where ηk → 0 as k → ∞ , and since b ∈ (0,1) d + ε kα ≤ b(d + ηk ) = bd k , for α ∈ [0, α ] , k ≥ k , for some α > 0 and k . Hence G ( wk + α (vk − wk )) − G ( wk ) ≤ α bd k , for α ∈ [0, α ] , k ≥ k . It follows from the choice of the Armijo step α k in Step 4 that α k ≥ cα , for k ≥ k . Hence 1486 I. Chryssoverghi, J. Coletsos and B. Kokkinis G ( wk +1 ) − G ( wk ) = G ( wk + α k (vk − wk )) − G ( wk ) ≤ α k bd k ≤ cα bd k ≤ cα bd / 2, for k ≥ k . This contradicts the fact that G ( wk ) → G ( w) , by Theorem 2.2. Therefore, we must have d = e = 0 , d k → 0 , and ek → 0 . But Step 3 then implies that m → ∞ , which is a contradiction. Therefore, m → ∞ in Algorithm 1. Let us show now that d k → 0 in case (ii). Since R is compact, let ( wk ) k∈K , (vk ) k∈K be subsequences, regarded as sequences in R , of the sequences generated by Algorithm 1 that converge to some r% ∈ R , r ' ∈ R , respectively. Then, by the continuity of the relaxed state and adjoint operators (Theorems 2.2-3), using Proposition 2.1 in [1], and since d k ≤ ek ≤ 0 , we have 2 e := lim ek = lim [ DG ( wk , vk − wk ) + (γ / 2) vk − wk ] k →∞ , k ∈K k →∞ , k∈K = d + (γ / 2) ∫ [r '( x) − r% ( x)]2 dt ≤ 0 , Ω d = lim d k = lim DG ( wk , vk − wk ) ≤ lim ek = e ≤ 0 , k →∞ , k∈K k →∞ , k∈K k →∞ , k∈K with 2 d := ∫ {∑ λl [ zl (t ) fu (t , y (t ), r% (t )) + glu (t , y (t ), r% (t ))] T 0 + l =0 ( ∫ λ (ds) g T t 3 3y ) ( s, y ( s )) Z ( s ) −1 Z (t ) fu (t , y (t ), r% (t ))}[r '(t ) − r% (t )]dt , where λ0 := 1 , y := yr% , z := zr% . Note that d k is, but d is not, a directional derivative in this case. Suppose that d < 0 . Since we have also, for α '(α ) ∈ (0, α ) lim + DG ( wk + α '(α )(vk − wk ), vk − wk ) = d , k →∞ , k∈K , α → 0 we get as above G ( wk +1 ) − G ( wk ) ≤ cα bd / 2, for k ≥ k , k ∈ K , for some k . Since the whole sequence (G ( wk )) is non-increasing by Steps 4 and 5, it follows that G ( wk ) → −∞ as k → ∞ , k ∈ K , which leads to a contradiction, as above. Therefore, d = e = 0 , d k → 0 , ek → 0 , k ∈ K , and this holds also for the whole sequences by the uniqueness of the limit 0. We conclude as above that l → ∞ in Algorithm 1. (i) Suppose that the sequences (λlm ) are bounded and, up to subsequences, that λlm → λl , l = 1, 2 , and λ3m → λ3 in M ( I )m weak star. By Theorem 2.2, since 3 wm → w in L2 strongly, we have 0 = lim m →∞ 0 = lim m →∞ 0 = lim m →∞ λ1m M m 1 λ2m M m 2 λ3m M = lim G1 ( wm ) = G1 ( w), m →∞ = lim[max(0, G1 ( wm ))] = max(0, G2 ( w)), L1 m 3 m →∞ 1 m →∞ M m 3 = lim = ∫ max(0, G3 ( w)( s ))ds , I ∫λ I m 3 ( s )ds = lim ∫ max(0, G3 ( wm )( s ))ds m →∞ I 1487 Optimal Control Problems which show that w is admissible. Now, let any w ' ∈ W and let ( w ' ∈ W ) be a subsequence converging to w ' in L2 strongly. By Steps 2 and 3, we have 2 DG m ( wm , w '− wm ) + (γ / 2) w '− wm L2 = DG0 ( wm , w '− wm ) + λ1m DG1 ( wm , w '− wm ) + λ2m DG2 ( wm , w '− wm ) + ∫ λ3m ( s ) DG3 ( wm , w '− wm )( s )ds + (γ / 2) w '− wm I 2 L2 ≥ dm . By Theorem 2.3, the derivatives DGl , l = 0,1, 2,3 ( DG3 with values in C ( I ) m3 ) are continuous and λ3m → λ3 in M ( I ) m3 weak star. Since also wm → w in L2 strongly, we can pass to the limit in the above inequality and obtain DG0 ( w, w '− w) + λ1 DG1 ( w, w '− w) + λ2 DG2 ( w, w '− w) T 2 0 L2 + ∫ λ3 (ds ) DG3 ( w, w '− w)( s ) + (γ / 2) w '− w ≥ 0, which holds for every w ' ∈ W . Replacing w ' by w + θ ( w '− w) , θ ∈ (0,1] , dividing by θ , and taking the limit as θ → 0 , we obtain the same inequality, but without the term containing the norm. On the other hand, by the definition of λ2m and Theorem 2.2, if G2 j ( w) < 0 , for some index j ∈ {1,..., m2 } , then λ2mj = 0 for m sufficiently large, hence λ2 j = 0 , which shows that λ2G2 ( w) = 0 . Now, since w is admissible, for each j = 1,..., m3 fixed, we have G3 j ( w) ≤ 0 , i.e. g3 j ( s, y ( s )) ≤ 0 , s ∈ I , with y = y w . Let ε > 0 be given, and define the sets Sε j = {s ∈ I − ε ≤ g3 j ( s, y ( s )) ≤ 0,} , S 'ε j = {s ∈ I g3 j ( s, y ( s )) ≤ −ε } . Let m ' be such that g3 j ( s, y m ( s )) − g3 j ( s, y ( s )) ≤ ε , s ∈ I , for m ≥ m ' , By the definition of λ3m , we have λ3mj ( s ) = 0 , s ∈ S 'ε j , for m ≥ m ' . It follows that η mj := ∫ T 0 ≤ 2ε λ3mj λ3mj ( s) g3 j ( s), y m ( s))ds = ≤ 2ε λ3mj L1 ( Sε j ) L1 ( I ) ∫ Sε j λ3mj ( s ) g3 j ( s ), y m ( s ))ds ≤ 2cε , for m ≥ m ' . Therefore, by the involved weak star and uniform convergences 0 = lim η mj = m →∞ ∫ T 0 λ3 j (ds) g3 j ( s, y ( s)) , j = 1,..., m3 , or equivalently (since λ3 ≥ 0 and g3 ( s, y ( s )) ≤ 0 , in the limit) ∫ T 0 λ3 ( ds ) g3 ( s, y ( s )) = 0 . On the other hand, since λ3m ≥ 0 and λ3m → λ3 weak star, we have 2 2 m3 1 + ∑ λlm + ∫ λ3m ( s )ds = 1 + ∑ λlm + ∫ {∑ [1⋅ λ3mj ( s )]}ds l =0 I l =0 T 0 j =1 1488 I. Chryssoverghi, J. Coletsos and B. Kokkinis 2 m3 2 j =1 l =0 → 1 + ∑ λl + ∫ {∑ [1 ⋅ λ j 3 ( ds )]} = 1 + ∑ λl + λ3 * = a ≠ 0 . T 0 l =0 We clearly have also λ0 = 1 , λ2 ≥ 0 , and since λ3m ≥ 0 and λ3m → λ3 in M ( I ) m3 weak star, we have also λ3 ≥ 0 . Dividing all λl by a , w is thus weakly extremal classical. (ii) Let ( wm ) be a subsequence (same notation) of the sequence generated in Step 3 that converges in R to some r in R . The admissibility of r is proved as in (i). Suppose as in (i) that λlm → λl , l = 1, 2 , and λ3m → λ3 in M ( I ) m3 weak star. By Step 2, we have DG m ( wm , w '− wm ) + (γ / 2) w '− wm 2 L2 = DG0 ( w , w '− w ) + λ DG1 ( w , w '− wm ) + λ2m DG2 ( wm , w 'n − wm ) m nm m 1 m + ∫ λ3m ( s ) DG3 ( wm , w '− wm )( s )ds + (γ / 2) w '− wm I 2 L2 ≥ dm , for every w ' ∈ W , which can be written 2 ∑λ ∫ m l l =0 T 0 [ zlm (t ) f u (t , y m (t ), wm (t )) + glu (t , y m (t ), wm (t ))][ w '(t ) − wm (t )]dt T + ∫ λ3m (ds ) g3 y ( s, y m ( s )) Z ( s ) −1 0 ( ∫ Z (t) f (t, y (t), w (t))[w '(t) − w (t)]dt ) s m m m u 0 2 T + (γ / 2) ∫ w '(t ) − wm (t ) dt ≥ d m , 0 for every w ' ∈ W . Choosing now any Caratheodory function φ ∈ B( I × U ;U ) and setting w '(t ) := φ (t , w(t )) in this inequality, by the continuity of the relaxed state and adjoint operators (Theorems 2.2-3) and using Proposition 2.1 in [1] (also for each fixed s - the integrals ∫ s 0 are equicontinuous and converge for each s , hence uniformly), and the convergences λ3m → λ3 in M ( I ) m3 weak star and wm → r in R , we can pass to the limit in this inequality and obtain 2 ∑λ ∫ l l =0 T 0 [ zl (t ) f u (t , y (t ), r (t )) + glu (t , y (t ), r (t ))][φ (t , r (t )) − r (t )]dt T + ∫ λ3 ( ds ) g3 y ( s, y ( s )) Z ( s ) −1 0 T (∫ Z (t) f (t, y(t), r(t))[φ (t, r(t)) − r(t)]dt ) s 0 u +(γ / 2) ∫ φ (t , r (t )) − r (t ) dt ≥ 0 , for every such φ , 0 2 which implies, by a θ -argument similar to (i), the same inequality, but without the last integral term, and with multipliers as in the optimality conditions. The transversality conditions are proved similarly to (i). (iii) In any of the above cases (i) or (ii), suppose that the limit control is admissible and that the classical, or relaxed, problem has no admissible, abnormal extremal, controls. Suppose that the multipliers are not all bounded. Then, dividing the corresponding inequality resulting from Step 2 by the greatest multiplier norm and passing to the limit for a subsequence, we readily see that we obtain an extremality inequality where the first multiplier is zero, and that the limit control is abnormal extremal, a contradiction. Therefore, the sequences of multipliers are bounded, and by (i) or (ii), this limit control is extremal as above. 1489 Optimal Control Problems The proofs in the absence of state constraints are similar. If the additional assumptions of Theorem 2.4 are also satisfied (classical case), then Algorithm 1 computes optimal classical controls in case (i). Next, we define the penalized functionals on R m1 m2 2 1 G m (r ) = G0 (r ) + {M 1m ∑ G1 j (r ) + M 2m ∑ [max(0, G2 j (r ))]2 2 j =1 j =1 m3 + M 3m ∑ ∫ [max(0, G3 j (r )( s ))]2 ds} . j =1 I The directional derivative of G m is given by DG m (r , r '− r ) = DG0 (r , r '− r ) +M +M m 1 m1 ∑G 1j j =1 (r ) DG1 j (r , r '− r ) + M m3 m 3 ∑ ∫ max(0, G j =1 3j I m 2 m2 ∑ max(0, G j =1 2j (r )) DG2 j (r , r '− r ) (r )( s )) DG3nj (r , r '− r )( s )ds . The relaxed penalized conditional descent method is described by the following Algorithm, where U is not necessarily convex. Algorithm 2 Step 1. Set k := 0 , m := 1 , and choose an initial control r01 ∈ R . Step 2. Find rk m ∈ R such that d k := DG m (rkm , rk m − rkm ) = min DG m (rkm , r '− rkm ). r '∈R m Step 3. If d k ≤ β , set r := r , r := rk m , d m := d k , m := m + 1 , and go to Step 2. m m m k Step 4. Find the lowest integer value s (positive or not), say s , such that α ( s ) = c sζ k ∈ (0,1] and α ( s ) satisfies the inequality G m (rkm + α ( s )(rk m − rkm )) − G m (rkm ) ≤ α ( s )bd k , and then set α k := α ( s ) . Step 5. Choose any rkm+1 ∈ R such that G l (rkl+1 ) ≤ G l (rkl + α k (rkl − rkl )), set k := k + 1 , and go to Step 2. With r m as defined in Step 3, define the sequences of multipliers λ1m := M 1mG1 (r m ), λ2m := M 2mmax(0, G2 (r m )), λ3m ( s) := M 3mmax(0, G3 (r m )( s )) = M 3mmax(0, g3 ( s, y m ( s))) . Theorem 3.2 We suppose that the derivatives in u are excluded in the last assumptions of Section 2. (i) In the presence of state constraints, if a subsequence (r l )l∈L of the sequence generated by Algorithm 2 in Step 3 converges to some r in R and the sequences (λlm ) , l = 1, 2,3 , (λ3m ) in L1 ( I )m3 , are bounded, then r is admissible and strongly 1490 I. Chryssoverghi, J. Coletsos and B. Kokkinis extremal relaxed for the relaxed problem. In the absence of state constraints, if a subsequence ( rk ) k∈K (no index m ) converges to some r in R , then r is strongly extremal relaxed for the relaxed problem. (ii) In case (i) with state constraints, suppose that the relaxed problem has no admissible, abnormal extremal, controls. If r is admissible, then the sequences (λmm ) are bounded and r is also strongly extremal relaxed for the relaxed problem. Proof. (State constraints present) Suppose by contradiction, as in the proof of Theorem 3.1, that m remains constant after a finite number of iterations in k , and so we drop here the index m . Since R is compact, let (rk )k∈K , (rk ) k∈K be subsequences of the sequences generated by Algorithm 2 that converge to some r% ∈ R , r% ∈ R , respectively. By Theorem 2.3, we have d := lim d k = lim DG (rk , rk − rk ) = DG (r%, r% − r% ) ≤ 0. k →∞ , k∈K k →∞ , k∈K Suppose that d < 0 . The function Φ(α ) := G (r + α (r '− r )) is continuous on [0,1] , and since the directional derivative DG (r , r '− r ) is linear in r '− r , Φ is differentiable on (0,1) and has derivative Φ '(α ) = DG (r + α (r '− r ), r '− r ). By the mean value theorem, for each α ∈ (0,1] there exists α '(α ) ∈ (0, α ) such that G (rk + α (rk − rk )) − G (rk ) = α DG (rk + α '(α )(rk − rk ), rk − rk ) . Therefore, by Theorem 2.3 G (rk + α (rk − rk )) − G (rk ) = α (d + ε kα ), for α ∈ [0,1] , where ε kα → 0 as k → ∞ , k ∈ K , and α → 0+ . We then show, similarly to Theorem 3.1, that d = 0 , hence d k → 0 , which leads also to a contradiction. Therefore, m → ∞ in Algorithm 2. (i) Let (r m ) be a subsequence (same notation) of the sequence generated in Step 3 of Algorithm 2, that converges to some r ∈ R as m → ∞ . The admissibility of r is shown similarly to Theorem 3.1. Suppose that the sequences (λmm ) are bounded and, up to subsequences, that λmm → λm . Now, by Steps 2 and 3 we have, for any r ' ∈ R 2 DG m (r m , r '− r m ) = ∑ λlm DGl (r m , r '− r m ) + ∫ λ3m ( s ) DG3 (r m , r '− r m )( s )ds ≥ d m . I l =0 Since d m ≤ β m → 0 by Step 3, using the above convergences and Theorem 2.3, we can pass to the limit and obtain 2 ∑ λ DG (r , r '− r ) + ∫ λ (s) DG (r , r '− r )(s)ds ≥ 0, l =0 l l I 3 3 for every r ' ∈ R , with multipliers λl as in the optimality conditions. Finally, we find in the limit the transversality and other conditions similarly to Theorem 3.1. Therefore, r is also weakly extremal relaxed for the relaxed problem. The proof in the absence of state constraints is similar. (ii) The proof is similar to that of Theorem 3.1 (iii). If the additional assumptions of Theorem 2.4 are also satisfied (relaxed case), then Algorithm 2 computes optimal relaxed controls. 1491 Optimal Control Problems One can see (using Filippov’s selection theorem, see [13]) that a classical (measurable) control rk m in Step 2 can be found for every k by appropriately minimizing H (t , y m (t ), z m (t ), u ) on U , for each t ∈ I (in practice, this is trivially done for discrete, e.g. piecewise constant, controls). Algorithm 2 can be implemented as follows. Suppose for simplicity that there are no state constraints. We first choose the initial discrete control in Step 1 to be of Gamkrelidze type, i.e. equal for each t to a convex combination of d + 1 Dirac measures at d + 1 points of U , where d is the dimension of the system. Suppose, by induction, that the control rkm computed in Algorithm 2 is of Gamkrelidze type. Since the control rk m in Step 2 is chosen to be classical, i.e. piecewise Dirac, the control r%km := rkm + α (rk m − rkm ) in Step 5 is piecewise equal to a convex combination of d + 2 Dirac measures. Using now a known property of convex hulls of finite vector sets, we can construct a Gamkrelidze control rkm+1 equivalent to r%km , i.e. such that f (t , y% km (t ), rkm+1 (t )) = f (t , y% km (t ), r%km (t )) ∈ IR d , t ∈ I , where y% km corresponds to r%km , by selecting only d + 1 appropriate points in U among the d + 2 ones defining r%km , for each t . The control rkm+1 yields then the same state, values of functionals and G m as r%km . Therefore, the constructed control rkm is of Gamkrelidze type for every k . In practice, by choosing in Algorithms 1 and 2 moderately growing sequences m ( M l ) and a sequence ( β m ) relatively fast converging to zero, the resulting sequences of multipliers (λlm ) are often kept bounded. One can choose a fixed ζ k := ζ ∈ (0,1] in Step 4; a usually faster and adaptive procedure is to set ζ 0 := 1 , and then ζ k := α k −1 , for k ≥ 1 . Gamkrelidze Formulation Approach When directly applied to nonconvex optimal control problems whose solutions are non-classical relaxed controls, the classical method yields often very poor convergence (due to highly oscillating extremal controls). For this reason, we describe here an alternative approach, assuming that U is convex, following [7], that uses the Gamkrelidze formulation. For simplicity, we consider only the case without state constraints. Consider the following relaxed problem, with state equation y '(t ) = f (t , y (t ), r (t )) , t ∈ I , y (0) = y 0 , control constraint r ∈ R , and cost functional G (r ) := g ( y (T )) . For each t fixed, the vector f (t , y (t ), r (t )) in IR d belongs to the convex hull of the set f (t , y (t ), U ) ⊂ IR d . Hence, we can write d +1 f (t , y (t ), r (t )) = ∑ v j (t ) f (t , y (t ), w j (t )) , with 0 ≤ v j (t ) ≤ 1 , j =1 d +1 ∑ v (t ) = 1 , j =1 j and by Filippov’s selection theorem, we can suppose that these functions v j , w j are measurable. Therefore, the control r yields the same state y as the Gamkrelidze 1492 I. Chryssoverghi, J. Coletsos and B. Kokkinis d +1 control rG := ∑ v j (t )δ w j (t ) . Conversely, every such a control rG is clearly a relaxed j =1 control r that yields the same state. Therefore, the above relaxed control problem is equivalent to the following extended classical problem, with state equation d +1 y '(t ) = ∑ v j (t ) f (t , y (t ), w j (t )) in I , y (0) = y 0 , j =1 classical controls v = ( v j ) , w = ( w j ) , convex control constraints d +1 ∑ v (t ) = 1 , 0 ≤ v (t ) ≤ 1 , w (t ) ∈ U , j j j =1 j j = 1,..., d + 1 , and cost functional G ( v, w ) := g ( y (T )) . We can therefore apply the classical method (Algorithm 1) to this problem. The main disadvantage of this approach is that the dimension of the control space is rapidly increased; it can be successfully applied for relatively small dimensions d , d ' . In the general case, i.e. if U is not convex, one can use Algorithm 2 to solve such highly nonconvex problems. Finally, Gamkrelidze relaxed controls (in practice discrete ones) computed as above, or by Algorithm 2, can then be approximated, and simulated, by classical controls using a standard procedure, see [2], [4]. 4 Numerical examples Let I := [0,1] . Example 1. Define the reference state and control y1 (t ) = e − t , y2 (t ) = e −2t , y3 (t ) = e −3t , w(t ) = min(1, −1 + 2.5t ) , and consider the following optimal control problem, with state equations y1 ' = − y1 + y3 − e −3t + sin y1 − sin y1 + w1 − w , y2 ' = y1 − 2 y2 − e − t + w2 − w, y3 ' = y2 − 3 y3 − e −2t + w3 − w, t∈I , y1 (0) = y2 (0) = y3 (0) = 1, control constraint set U = [−1,1] , and cost functional 3 G0 ( w) := 0.5∫ {∑ [( yi − yi ) 2 + ( wi − w) 2 ]}dt. 1 0 i =1 The optimal control and state are clearly w*:= ( w, w, w) and y*:= ( y1 , y2 , y3 ) . Algorithm 1, without penalties, was applied to this example, using the midpoint scheme (step size 0.002) for solving the differential equations, with piecewise constant classical controls, and γ = 0.5 . After 15 iterations in k , we obtained G0 ( wK ) = 1.366 ⋅10−12 , d k = −4.979 ⋅10−15 , ε k = 4.316 ⋅10−7 , ζ k = 4.789 ⋅10−7 , Optimal Control Problems 1493 where d k was defined in Step 2 of Algorithm 1, ε k is the max-error for the state at the nodes, and ζ k the max-error for the control at the midpoints of the intervals. Figure 1 shows the first component of the last computed control. Example 2. With the same state equations, cost, and parameters as in Example 1, but with constraint set U = [−0.75,1] , additional pointwise state constraints G3 ( w)(t ) = 0.8 − 0.4t − y1 (t ) ≤ 0 , t ∈ I , and applying here the penalized Algorithm 1, we obtained after 85 iterations in k G0 ( wk ) = 3.765340220 ⋅10−3 , d k = −4.502 ⋅10−6 , the maximum state constraint violation θ k := max[max(0, 0.8 − 0.4ti − y1i ,k )] = 1.444 ⋅10−5 , i and the first control and state components shown in Figures 3 and 4. Example 3. Consider the nonconvex problem, with state equations y1 ' = − y1 + w, y2 ' = 0.5( y1 − y ) 2 − w2 , t ∈ I , y1 (0) = 1, y2 (0) = 0, where y (t ) = e − t , control constraint set U = [−1,1] , and cost G ( w) = y2 (1) . The unique optimal relaxed control is clearly r * (t ) = (δ −1 + δ 1 ) / 2 ( δ −1 , δ1 Dirac measures), with optimal state y* = y and optimal cost G ( r*) = −1 . Note that the optimal relaxed cost can be approximated as closely as desired with a classical control, but cannot be attained for such a control. Since here the set f (t , y ,U ) is a continuous arc, hence a connected set, the Gamkrelidze formulation involves only three controls v, u, w 1 y1 ' = − y1 + vu + (1 − v) w, y2 ' = ( y1 − y ) 2 − vu 2 − (1 − v) w2 , t ∈ I , 2 y1 (0) = 1, y2 (0) = 0, with v ∈ [0,1] and u, w ∈ [ −1,1] . Applying to this problem Algorithm 1 without penalties, we obtained after 15 iterations the control vk ≈ 0.5 with max-error ≈ 3 ⋅10−6 at the midpoints, the controls uk = −1 , wk = 1 exactly, the optimal state with maxerror ≈ 3.8 ⋅10−6 at the nodes, the approximate cost G (vk , uk , wk ) = −0.999999999997 , and d k = −1.134 ⋅10−6 . Example 4. Consider the following problem, with state equations y1 ' = − y2 + w1 , y2 ' = − y1 + w2 , t ∈ [0,0.5) , y1 ' = − y2 + w1 − t + 0.5 , y2 ' = − y1 + w2 − t + 0.5 , t ∈ [0.5,1] , y1 (0) = y2 (0) = 1 , nonconvex control constraint set U = {(u1 , 0) ∈ IR 2 0 ≤ u1 ≤ 1} ∪ {(0, u2 ) ∈ IR 2 0 ≤ u2 ≤ 1} , and cost functional 1 G0 ( w) = 0.5∫ [( y1 − y ) 2 + ( y2 − y ) 2 ]dt , 0 −t where y = e . Clearly, the unique optimal relaxed control is r * (t ) = δ (0,0) , t ∈ [0,0.5) (classical) 1494 I. Chryssoverghi, J. Coletsos and B. Kokkinis r * (t ) = 2(1 − t )δ (0,0) + (t − 0.5)δ (1,0) + (t − 0.5)δ (0,1) , t ∈ [0.5,1] (non-classical), where δ denotes Dirac measures, which yields the optimal state ( y1*, y2 *) = ( y , y ) , and cost G0 ( r*) = 0 . Algorithm 2, without penalties, was applied here, using also the midpoint scheme (step size 0.002), here with piecewise constant relaxed controls. After 120 iterations in k , we obtained G0 (rk ) = 2.552 ⋅10−6 , d k = −3.393 ⋅10−5 . Figures 4 and 5 show the last relaxed control probability functions ⎧1, t ∈ [0, 0.5) p0 (t ) = rk (t )({(0, 0)}) ≈ ⎨ ⎩2(1 − t ), t ∈ [0.5,1) ⎧0, t ∈ [0, 0.5) p1 (t ) = rk (t )({(1, 0)}) ≈ ⎨ ⎩t − 0.5, t ∈ [0.5,1) We also obtained p2 (t ) = rk (t )({(0,1)}) = 1 − p0 (t ) − p1 (t ) . Example 5. Consider the following problem, with state equation y ' = − y + w , t ∈ I , y (0) = 1 , control constraint set U = [−1,1] , state constraints G1 ( w) = y (1) − 0.5 = 0 , G3 ( w)( s ) = 0.3 − y ( s ) ≤ 0 , s ∈ I , and nonconvex cost functional 1 G0 ( w) = ∫ (0.5 y 2 − w2 )dt . 0 Writing the solution of the state equation in closed form via the fundamental solution, first forward with initial condition y (0) = 1 and then backward with terminal condition y (1) = 0.5 , we can see that the unique optimal relaxed control and state are ⎧δ −1 , t ∈ [0, ρ ) (classical) ⎪ r * (t ) = ⎨0.35δ −1 + 0.65δ 1 , t ∈ [ ρ , σ ) (non-classical) ⎪δ , t ∈ [σ ,1] (classical) ⎩ 1 ⎧−1 + 2e − t , t ∈ [0, ρ ) ⎪ y * (t ) = ⎨0.3, t ∈ [ ρ , σ ) ⎪1 − 0.5e1−t , t ∈ [σ ,1] ⎩ where ρ = − ln 0.65 ≈ 0.43 is such that −1 + 2e− ρ = 0.3 , and σ = 1 − ln1.4 ≈ 0.66 such that 1 − 0.5e1−σ = 0.3 . The optimal cost is G0 (r*) ≈ −0.868398913624 . Applying here the penalized Algorithm 2, we obtained after 200 iterations in k the results G0 (rkm ) = −0.868200567558 , G1 (rkm ) = −9.732 ⋅10−6 , max[max(0, 0.3 − y1mi,k )] = 2.962 ⋅10−4 , d k = −1.236 ⋅10−3 . i Figure 6 shows the last relaxed control probability function ⎧0, t ∈ [0, ρ ) ⎪ m p1 (t ) = rk (t )({(1)}) ≈ ⎨0.65, t ∈ [ ρ , σ ) ⎪1, t ∈ [σ ,1] ⎩ and we also obtained p0 (t ) = rkm (t )({−1}) = 1 − p1 (t ) . Figure 7 shows the last state. 1495 Optimal Control Problems W1 1 0.5 1 t -1 Figure 1. Example 1: Last control (1st component) W1 1 0.5 1 - 0.75 Figure 2. Example 2: Last control (1st component) t 1496 I. Chryssoverghi, J. Coletsos and B. Kokkinis Y1 1 0.5 0.4 0.5 1 t Figure 3. Example 2: Last state (1st component) P0 1 0.5 1 t Figure 4. Example 4: Last relaxed control probability p0 P1 1 0.5 0.5 1 t Figure 5. Example 4: Last relaxed control probability p1 1497 Optimal Control Problems P1 1 0.65 с у 1 t Figure 6. Example 5: Last relaxed control probability p1 Y 1 0.5 0.3 с у 1 t Figure 7. Example 5: Last state References [1] I. Chryssoverghi, Nonconvex optimal control of nonlinear monotone parabolic systems, Systems and Control Letters, 8 (1986), 55-62. [2] I. Chryssoverghi, A. Bacopoulos, Discrete approximation of relaxed optimal control problems, Journal of Optimization Theory and Applications, 65, 3 (1990), 395-407. [3] I. Chryssoverghi, A. Bacopoulos, J. Coletsos and B. Kokkinis, Discrete approximation of nonconvex hyperbolic optimal control problems with state constraints, Control and Cybernetics, 27, 1 (1998), 29-50. [4] I. Chryssoverghi, J. Coletsos and B. Kokkinis, Discrete relaxed method for semilinear parabolic optimal control problems, Control and Cybernetics, 28, 2 (1999), 157-176. [5] I. Chryssoverghi, J., Coletsos, B., Kokkinis, Approximate relaxed descent method for optimal control problems, Control and Cybernetics, 30, 4 (2001), 385-404. 1498 I. Chryssoverghi, J. Coletsos and B. Kokkinis [6] I. Chryssoverghi, Discrete gradient projection method with general Runge-Kutta schemes and control parameterizations for optimal control problems, Control and Cybernetics, 34, (2005), 425-451. [7] I. Chryssoverghi, Discretization methods for semilinear parabolic optimal control problems, International Journal of Numerical Analysis and Modeling, 3 (2006), 437458. [8] A.L. Dontchev, W.W. Hager and V.M. Veliov, Second-order Runge-Kutta approximations in control constrained optimal control, SIAM Journal on Numerical Analysis, 38, 1 (2000), 202-226. [9] T. Roubíček, A Convergent computational method for constrained optimal relaxed control problems, Journal of Optimization Theory and Applications, 69 (1991), 589603. [10] T. Roubíček, Relaxation in Optimization Theory and Variational Calculus, Walter de Gruyter, Berlin (1997). [11] V.M. Veliov, On the time-discretization of control systems, SIAM Journal on Control and Optimization, 35, 5 (1997), 1470-1486. [12] A. Schwartz and E. Polak, Consistent approximations for optimal control problems based on Runge-Kutta integration, SIAM Journal on Control and Optimization, 34, 4 (1996), 1235-1269. [13] J. Warga, Optimal Control of Differential and Functional Equations, Academic Press, New York (1972). [14] J. Warga, Steepest Descent with Relaxed Controls, SIAM Journal on Control and Optimization, 15 (1977) 674-682. Received: May 18, 2006
© Copyright 2026 Paperzz