Aachen Institute for Advanced Study in Computational Engineering Science Preprint: AICES-2011/09-1 06/September/2011 Towards an Adaptive One Shot Method for Optimal Control Problems L. Kaland, J. C. De Los Reyes and N. R. Gauger Financial support from the Deutsche Forschungsgemeinschaft (German Research Association) through grant GSC 111 is gratefully acknowledged. ©L. Kaland, J. C. De Los Reyes and N. R. Gauger 2011. All rights reserved List of AICES technical reports: http://www.aices.rwth-aachen.de/preprints Towards an Adaptive One Shot Method for Optimal Control Problems Lena Kaland ∗ Juan Carlos De Los Reyes Nicolas R. Gauger † ‡ Abstract We consider two model problems for the One Shot approach in Hilbert spaces, the stationary solid fuel ignition model and the stationary viscous Burgers equation. For these two problems we present a convergence analysis for the One Shot method, where we evaluate one step in the state, adjoint and design equation simultaneously. We experience mesh independence in the numerics, which motivates an adaptive One Shot approach. 1 Introduction We study optimal control problems of the following type min f (y, u) s.t. e(y, u) = 0, (P) where e(y, u) displays a partial differential equation with a state vector y ∈ Y and the control u ∈ U . Y and U are Hilbert spaces and f : Y × U → is the objective function. These kind of problems are studied in many textbooks (see e.g. [14] ) and can be solved numerically in numerous ways. One essential difference in treating the optimization problem is to first discretize the problem and then use a finite dimensional optimization method, or to apply an optimization method in a Hilbert space setting and finally discretize. For both approaches the optimality system plays an important role. By the assumption that the discretized state and adjoint equation can be solved iteratively, Ta’asan suggested in a multigrid framework for elliptic PDEs in [13] to solve the corresponding optimality system simultaneously, i.e. one step in the state equation, one step in the adjoint equation and finally one control update. This so-called One Shot Method was employed in many applications, especially aerodynamics, see e.g. [11] and [12] and references therein. In [5] and [6] a convergence proof for the finite dimensional setting and a suitable preconditioner for the control update is given. To further improve the One Shot approach the question of adaptivity and therefore the analysis of the method in a Hilbert space setting arises. We present a convergence result in Hilbert spaces following the ideas of the proof in [5]. ∗ RWTH Aachen, [email protected] Quito, [email protected] ‡ RWTH Aachen, [email protected] † EPN 1 In this context we study two distributed optimal control problems. On the one hand we consider a tracking type functional and the following partial differential equation ∆y + δ ey + u = 0 y=0 in Ω, on Γ. This equation serves for instance as a model for combustion and is often called the steady state solid fuel ignition model. Additionally, due to the strong nonlinearity, it presents a good example for an optimal control problem with nonlinear partial differential equations. The main difficulty is to handle this nonlinearity. In fact just by changing the sign and working with −ey instead, one could easily apply the standard theory for semilinear elliptic equations ([14]). As a second example and simple model for convection-diffusion problems we consider the stationary viscous Burgers equation given by −νy !! + yy ! = u y(0) = 0, y(1) = 0 in (0, 1), in one space dimension. The instationary analogue was introduced by Burgers in [1]. The objective function is again chosen to be a tracking type functional. In two space dimensions the Burgers equation is given by −ν∆y + (y · ∇)y = u, y|Γ = 0. The paper is structured as follows. In section 2 we introduce the setting of both model problems and answer the question about the existence of an optimal solution. In section 3 we state the general form of the One Shot Method. We continue in section 4 with the convergence analysis for the two model problems. Here we adopt the approach in [5] and show that the iterates of the One Shot method serve as a descent direction for a doubly augmented Lagrange function. Finally in section 5 we show numerical results and observe mesh independence. We extend the One Shot method algorithmically by an additional adaptive step. According to an error indicator the mesh is refined or coarsened and the solution is interpolated to the new mesh. Notation. Throughout this paper the norm & · & without any indices denotes the operator norm. Additionally we will not differ between an operator and its Riesz representative. g ∗ ∈ L(Y ∗ , X ∗ ) denotes the dual of some operator g ∈ L(X, Y ) with given Banach spaces X and Y . If not stated otherwise C will be a generic constant. 2 2.1 Problem statement The solid fuel ignition model We study (P) with the tracking type cost functional ! ! µ 1 |y − zd |2 dx + |u|2 dx f (y, u) := 2 Ω 2 Ω 2 (1) and the partial differential equation ∆y + δ ey + u = 0 y=0 in Ω, on Γ. (2) Here Ω is an open, bounded domain with Lipschitz boundary Γ := ∂Ω, y ∈ Y = H01 = H01 (Ω), u ∈ U = L2 = L2 (Ω). Moreover, we assume that zd ∈ L2 (Ω) and δ, µ ∈ + . Finally the optimal control problem reads min f (y, u) s.t. (y, u) satisfies (2). (P1) The existence and uniqueness of the uncontrolled problem of (2) is a complicated task and depends on the domain Ω and the positive constant δ. More precisely the problem may admit multiple solutions, exactly one solution or no solution at all. For some values of δ, for which the uncontrolled case did not have a solution, the controlled equation (2) may admit control-state solution pairs. We can assume the existence for adequate Ω and δ (see [2]), but because of the nonlinearity we need to state an appropriate concept of a solution. Definition 1. For u ∈ L2 (Ω) a function y ∈ H01 (Ω) is called a solution of (2) if ey ∈ L1 (Ω) and it satisfies (2) in the distributional sense. The optimal control of the solid fuel ignition model was studied by several authors. To handle the exponential term in [8] the term ∇y is included in the cost functional. This technique is also applied in [2], here the authors alternatively include the exponential term. The optimal control of the instationary solid fuel ignition model is handled e.g. in [9], [10] or [8]. In this paper we bound the nonlinearity explicitely to get existence of an solution for problem (P1). Theorem 1. If ey(x) ≤ M holds for a.e. x ∈ Ω, y(x) ∈ (3) ∗ ∗ , then (P) admits a solution (y , u ) ∈ H01 × L2 . Proof. There exists (y, u) ∈ H01 × L2 satisfying (2), i.e. the feasible set is nonempty. Since the cost functional is bounded from below, there exists a minimizing sequence (yn , un ) ∈ H01 × L2 with lim f (yn , un ) = inf f (y, u). n→∞ u Since f (y, u) → ∞ for &u&L2 → ∞, {un } is bounded and also relatively weakly sequentially compact because L2 is reflexive. Therefore there exists a subsequence that converges weakly to a limit u∗ in L2 . Now we need the convergence of a subsequence of {yn } in H01 . From (3) we get that {eyn } is bounded in L2 and a subsequence converges weakly to an element w∗ . In the following we denote the subsequence of {(yn , un , eyn )} with the same symbol. Since (yn , un ) is a solution to (2) for every n, yn solves the linear boundary value problem −∆yn = Rn in Ω yn = 0 on Γ with Rn = eyn + un . {Rn } converges weakly to w∗ + u∗ in L2 . Additionally Rn )→ yn is a linear and continuous mapping from L2 to H01 and therefore 3 weakly sequentially continuous. That is why we obtain the weak convergence of {yn } to y ∗ in H01 . At all we obtain yn $ y ∗ in H01 , un $ u∗ in L2 and eyn $ w∗ in L2 . We still need to show that (y ∗ , u∗ ) is a solution of (2). H01 is compactly embedded in L2 and it follows yn → y ∗ in L2 . Using the assumption (3), the nonlinearity is Lipschitz for y(x) ∈ (−∞, ln M ], i.e. ∗ &eyn − ey &L2 ≤ M &yn − y ∗ &L2 ∗ and therefore eyn → ey in L2 . For every v ∈ H01 the expression (yn , v)H01 − δ(eyn , v)L2 = (un , v)L2 converges to ∗ (y ∗ , v)H01 − δ(ey , v)L2 = (u∗ , v)L2 . Because f is weakly lower semicontinuous it follows that (y ∗ , u∗ ) is optimal. We can reformulate the state equation (2) as a fixed point equation y = G(y, u) := (−∆)−1 (δ ey + u) (4) and assume because of (3) a contraction factor of ρ, i.e. &Gy (y, u)& ≤ ρ < 1. (5) We need this strong assumption for the subsequent analysis and are going to verify it in section 5. Formally we can derive the optimality system for problem (P1) as ∆y + δ ey + u = 0, ∆p + δ pey + y − zd = 0, y|Γ = 0 p|Γ = 0 (6a) (6b) µu + p = 0 a.e.in Ω (6c) Note that due to (5) there exists a solution of the linearized fixed point equation x = Gy (y, u)x. Consequently there exists an unique adjoint p that solves (6b). 2.2 The stationary viscous Burgers equation As a second model problem we consider again the tracking type cost functional (1) with the stationary Burgers equation. In the two-dimensional case it is given by −ν∆y + (y · ∇)y = u y=0 in Ω, on Γ. (7) In one space dimension it reduces to: −νy !! + yy ! = u in Ω, y=0 on Γ. 4 (8) Here ν > 0 denotes the viscosity parameter and u ∈ L2 (Ω). Originally Burgers introduced the equation yt + yy ! = νy !! (9) in [1] as the simplest model for fluid flow and ever since it was studied analytically and numerically (see e.g. [7]). The optimal control problem reads min f (y, u) s.t. (y, u) satisfies (7) resp. (8). (P2) In [15] the one-dimensional stationary and instationary case was considered in the context of optimal control. Therefore we refer to [15] for the aspects of existence and uniqueness of the Burgers equation and the existence of an optimal solution of (P2). Additionally in [4] the optimal control of (8) with pointwise control constraints was studied. In the following we set without loss of generality the viscosity parameter ν = 1. Formally we find an adjoint p ∈ H01 (Ω) to the optimal solution (y, u) ∈ H01 (Ω) × L2 (Ω), so that the following Lemma holds. Lemma 1. The optimality system for (P2) is given in the two dimensional case by −∆y + (y · ∇)y − u = 0, T −∆p − (y · ∇)p − div(y)p + (∇y) p − y + zd = 0, µu + p = 0 y|Γ = 0 (10a) p|Γ = 0 (10b) a.e. in Ω. (10c) Proof. See Remark 4. In one dimension this reduces to the system: −y !! + yy ! − u = 0, !! ! −p − yp − y + zd = 0, µu + p = 0 y|Γ = 0 p|Γ = 0 a.e. in Ω. In the subsequent analysis we deal with fixed point operators. To obtain a numerical stable operator we choose G(y, u) := (−∆ + y · ∇)−1 (u). (12) Several times we will need the derivatives of G(y, u). We obtain the derivative with respect to u directly by Gu (y, u)h := (−∆ + y · ∇)−1 (h). (13) For the derivative with respect to y we need the following lemma Lemma 2. Let f be the inverse operator f : x )→ x−1 . Then the derivative of f at x0 in direction h is given by −1 f ! (x0 )h = x−1 0 (−h)x0 . 5 Proof. By definition and with a Neumann series we obtain f (x0 + th) − f (x0 ) = (x0 + th)−1 − x−1 0 −1 = [(I + thx−1 − x−1 0 )x0 ] 0 ∞ " −1 n = x−1 (−thx−1 0 0 ) − x0 = n=0 −1 −tx0 hx−1 0 + O(t2 ) Deviding by t and t → 0 finishes the proof. For the more general form f (g(y)) we obtain fy (g(y))w = f ! (g(y))[g ! (y)w] It follows with g(y) = −∆ + y · ∇ and Lemma 2 that Gy (y, u)w = (−∆ + y · ∇)−1 (−(w · ∇)(−∆ + y · ∇)−1 (u)) = (−∆ + y · ∇)−1 (−(w · ∇)G(y, u)). (14) Here again we need the general assumption &Gy (y, u)& ≤ ρ < 1 (15) for the fixed point equation to converge. Notice that in a stationary point (14) yields Gy (y, u)w = (−∆ + y · ∇)−1 (−(w · ∇)y). 3 General setting of the One Shot Method To explain the idea of the One Shot Method we consider the general optimization problem min f (y, u) s.t. y = G(y, u), (16) with f (y, u) : Y × U → , G : Y × U → Y and Y and U appropriate Hilbert spaces. For the optimality conditions let us define the Lagrange function L(y, p, u) := f (y, u)− < ϕ, y − G(y, u) >Y ∗ ,Y . (17) Here ϕ is the problem specific adjoint operator as explained in the following. For a given partial differential equation e(y, u) it is possible to reformulate it in terms of a fixed point operator G(y, u) e(y, u) = F (y, u)[y − G(y, u)] with F (y, u) : Y → Y ∗ and provided that F (y, u)−1 exists . Now < p, e(y, u) >Y,Y ∗ =< F (y, u)∗ p, y − G(y, u) >Y ∗ ,Y . Finally we define ϕ := F (y, u)∗ p. Therefore the Lagrangian (17) is equivalent to the standard Lagrangian. As a first example consider the solid fuel ignition model with e(y, u) = −∆y − δ ey − u = −∆(y − (−∆)−1 (δ ey + u)) 6 With G(y, u) as defined in (4) it follows F (y, u) = F = −∆ and ϕ := −∆p. For the Burgers equation with e(y, u) = −∆y + (y · ∇)y − u = (−∆ + y · ∇)(y − (−∆ + y · ∇)−1 (u)) and (12) it follows F (y, u) = F (y) = (−∆+y·∇) and ϕ := (−∆−y·∇−div(y))p. Because the operator F (y, u) depends in our examples at most on y we will write F (y) in the following. With the help of the Lagrangian (17) we obtain the optimality conditions by differentiation. For the state equation if follows Lp (y, p, u)v = − < F (y)∗ v, y − G(y, u) >Y ∗ ,Y = − < v, F (y)[y − G(y, u)] >Y ∗ ,Y . For the adjoint equation we need to take the dependencies on the state into account. Therefore Ly (y, p, u)w = fy (y, u)w− < Fy (y)∗ w p, y − G(y, u) >Y ∗ ,Y − < F (y)∗ p, w − Gy (y, u)w >Y ∗ ,Y = fy (y, u)w− < p, Fy (y)w(y − G(y, u)) >Y,Y ∗ + < F (y)∗ p, Gy (y, u)w >Y ∗ ,Y − < F (y)∗ p, w >Y ∗ ,Y If we now define the operator Ny (y, p, u) by < F (y)∗ Ny (y, p, u), w >Y ∗ ,Y :=< fy (y, u), w >Y ∗ ,Y (18) − < p, Fy (y)w(y − G(y, u)) >Y,Y ∗ + < F (y)∗ p, Gy (y, u)w >Y ∗ ,Y it follows that < Ly (y, p, u), w >Y ∗ ,Y =< F (y)∗ Ny (y, p, u), w >Y ∗ ,Y − < F (y)∗ p, w >Y ∗ ,Y . Finally the KKT system for problem (16) reads F (y)[y − G(y, u)] = 0 in Y ∗ , (19a) ∗ (19b) ∗ F (y) [Ny (y, p, u) − p] = 0 in Y , Lu (y, p, u) = 0 in U. (19c) Ny is the iteration operator for the adjoint problem. We use this notation to be consistent with [5], where it was introduced as the derivative of the so-called shifted Lagrangian. We note that we define it here directly in a linearized manner. A standard optimization technique would solve (19a) for some initial u = u0 to get y, calculate p with these values using (19b) and finally update u using some optimization strategy involving (19c). Following the approach in [5] we instead introduce the iterative One Shot method yk+1 = G(yk , uk ) pk+1 = Ny (yk , uk , pk ) (20a) (20b) uk+1 = uk − Bk−1 Lu (yk , uk , pk ), (20c) with some preconditioning operator Bk , that needs to be determined. Notice that (20a), (20b) and (20c) operate fully independently in each iteration step. 7 Remark 1. If the contraction factor of the primal iteration is given by &Gy (y, u)& ≤ ρ < 1 it is obvious by (18) that the contraction factor of the dual iteration is the same in the case if Fy (y) = 0, for example for the solid fuel ignition model. Now we are able to formulate the following algorithm: Algorithm 1 One Shot Method 1: 2: 3: 4: 5: 6: 7: Choose u0 , y0 , p0 , tol > 0 for k = 0, 1, 2, . . . do if res < tol then STOP end if calculate yk+1 , pk+1 and uk+1 using (20a), (20b) and (20c) end for 4 Convergence Analysis For the convergence analysis of the One Shot Method we can exploit the special structure of the two model problems. Especially for the solid fuel ignition model the operator F (y) has a simple structure, in detail F (y) = −∆. Here the dual product reduces to the H01 scalar product. For the Burgers equation F (y) also depends on y and we cannot do this simplification. 4.1 The solid fuel ignition model For the subsequent analysis we consider problem (P1) again. Because G(y, u) = (−∆)−1 (δ ey + u), we can define the Lagrangian for p ∈ H01 (Ω) and ϕ := −∆p according to (17) as L(y, p, u) := f (y, u) − (p, y − G(y, u))H01 . (21) For this problem we can define the shifted Lagrangian directly by N (y, p, u) := f (y, u) + (p, G(y, u))H01 . Note that p = Ny (y, p, u) is the adjoint iteration. Remark 2. Notice that since &Gy (y, u)& ≤ ρ < 1 holds, the adjoint equation has the same contraction factor ρ, see Remark 1. In fact, from Ny (y, p, u) = fy (y, u) + Gy (y, u)∗ p we get Nyp (y, p, u) = Gy (y, u)∗ and therefore &Nyp (y, p, u)& = &Gy (y, u)∗ & = &Gy (y, u)& ≤ ρ < 1. 8 According to (20) the One Shot iteration can be written as yk+1 = G(yk , uk ) = (−∆)−1 (δ eyk + uk ) −1 yk pk+1 = Ny (yk , pk , uk ) = (−∆) (δ pk e + yk − zd ) 1 1 uk+1 = uk − Nu (yk , pk , uk ) = uk − (µuk + pk ). γ γ (22a) (22b) (22c) We get the iterative method (22) directly from the optimality system (6) by introducing the preconditioning constant γ > 0. How this constant can be set, will be determined in the subsequent analysis. For the convergence analysis we introduce the augmented Lagrangian α β &G(y, u) − y&2H 1 + &Ny (y, p, u) − p&2H 1 0 0 2 2 1 µ = &y − zd &2L2 + &u&2L2 − (p, y − G(y, u))H01 2 2 β α + &G(y, u) − y&2H 1 + &Ny (y, p, u) − p&2H 1 . 0 0 2 2 La (y, p, u) := L(y, p, u) + (23) The idea is to show that the augmented Lagrangian is an exact penalty function. That means that every local minimizer of problem (16) is also a local minimizer of the augmented Lagrangian. We want to find conditions on α and β, so that this is fulfilled. Lemma 3. If there exist constants α > 0 and β > 0 such that the following condition is fulfilled αβ(1 − ρ)2 > 1 + β&Nyy (y, p, u)& (24) then a point is a stationary point of La as defined in (23) if and only if it is a root of s as defined in (25). Proof. We start by deriving the gradient of La . For v, w ∈ H01 and h ∈ L2 we get Lay (y, p, u)w = (y − zd , w)L2 − (p, w)H01 + (p, (−∆)−1 (ey w))H01 + α((−∆)−1 (ey + u) − y, (−∆)−1 (ey w) − w)H01 + β((−∆)−1 (pey + y − zd ) − p, (−∆)−1 (pey w + w))H01 = (Ny (y, p, u) − p, w)H01 + α(G(y, u) − y, (Gy (y, u) − I)w)H01 + β(Ny (y, p, u) − p, Nyy (y, p, u)w)H01 Lap (y, p, u)v = ((−∆)−1 (ey + u) − y, v)H01 + β((−∆)−1 (pey + y − zd ) − p, (−∆)−1 (vey ) − v)H01 = (G(y, u) − y, v)H01 + β(Ny (y, p, u) − p, (Nyp (y, p, u) − I)v)H01 Lau (y, p, u)h = µ(u, h)L2 + (p, h)L2 + α((−∆)−1 (ey + u) − y, (−∆)−1 h)H01 = Nu (y, p, u)h + α(G(y, u) − y, Gu (y, u)h)H01 . 9 Since we aim to find a descent direction later we write this in the form −Lay (y, p, u)w = −(s2 , w)H01 + α(s1 , (I − Gy (y, u))w)H01 − β(s2 , Nyy (y, p, u)w)H01 −Lap (y, p, u)v = −(s1 , v)H01 + β(s2 , (I − Nyp (y, p, u))v)H01 −Lau (y, p, u)h = γ(s3 , h)L2 − α(s1 , Gu (y, u)h)H01 with G(y, u) − y s1 s = s2 = Ny (y, p, u) − p . − γ1 Nu (y, p, u) s3 (25) If s = 0, it follows immediately that −∇La = 0. For the other direction we interpret −∇La = 0 as the weak formulation of a system of partial differential equations ∆s2 − αs1 ey − α∆s1 − βs2 pey − βs2 = 0 ∆s1 − β∆s2 − βs2 ey = 0 γs3 − αs1 = 0. If the related bilinear form is coercive for all (w, v, h) ∈ H01 × H01 × L2 , then there exists a unique solution due to Lax-Milgram and therefore s = 0. In −Lau (y, p, u)h = 0, s3 just depends on s1 and we can restrict ourselves to solving the first two equations. From −Lap (y, p, u)v = 0 it follows (s1 , v)H01 = β(s2 , (I − Nyp (y, p, u))v)H01 and therefore (s1 , (I − Gy (y, u))w)H01 = β(s2 , (I − Nyp (y, p, u))(I − Gy (y, u))w)H01 = β((I − Nyp (y, p, u))∗ s2 , (I − Gy (y, u))w)H01 = β((I − Gy (y, u))s2 , (I − Gy (y, u))w)H01 . We plug this into −Lay (y, p, u)w = 0 and obtain the bilinear form a(w, v) = −(v, w)H01 + αβ((I − Gy (y, u))v, (I − Gy (y, u))w)H01 − β(v, Nyy (y, p, u)w)H01 = 0. For coercivity we need a(w, w) ≥ C&w&2H 1 . With 0 a(w, w) = −&w&2H 1 + αβ&(I − Gy (y, u))w&2H 1 − β(w, Nyy (y, p, u)w)H01 0 0 ≥ −&w&2H 1 + αβ&(I − Gy (y, u))w&2H 1 − β&Nyy (y, p, u)&&w&2H 1 0 0 0 and &(I − Gy (y, u))w&H01 ≥ &w&H01 − &Gy (y, u)w&H01 ≥ (1 − ρ)&w&H01 it follows a(w, w) ≥ −&w&2H 1 + αβ(1 − ρ)2 &w&2H 1 − β&Nyy (y, p, u)&&w&2H 1 0 0 0 = (αβ(1 − ρ)2 − 1 − β&Nyy (y, p, u)&)&w&2H 1 . 0 10 Therefore, from the assumption αβ(1 − ρ)2 − 1 − β&Nyy (y, p, u)& > 0 we obtain positivity. We still need to show, that the stationary point is a local minimizer of La . Since (y ∗ , u∗ ) is a minimizer of (P1) and we assume that the sufficient second order condition is fulfilled, there exist constants C > 0 and ε > 0 such that for (y, u) ∈ H01 × L2 with &y − y ∗ &2H 1 + &u − u∗ &2L2 ≤ ε it holds 0 &u − u∗ &2L2 ≤ C(f (y, u) − f (y ∗ , u∗ )) = C(L(y, p, u) + (p, G(y, u) − y)H01 − L(y ∗ , p∗ , u∗ )) = C(La (y, p, u) − La (y ∗ , p∗ , u∗ ) + (p, G(y, u) − y)H01 − α β &G(y, u) − y&2H 1 − &Ny (y, p, u) − p&2H 1 ) 0 0 2 2 Under the assumption (p, G(y, u) − y)H01 ≤ β α &G(y, u) − y&2H 1 + &Ny (y, p, u) − p&2H 1 0 0 2 2 (26) this results in &u − u∗ &2L2 ≤ C(La (y, p, u) − La (y ∗ , p∗ , u∗ )). (27) La (y ∗ , p∗ , u∗ ) ≤ La (y, p, u), (28) Therefore for any (y, p, u) in a neighborhood of (y ∗ , p∗ , u∗ ). α and β must be big enough to fulfill (26). So far we know that any minimizer of problem (P1) is also a minimizer of the augmented Lagrangian. In the next step we show that the One Shot iterate is a descent direction for La and therefore the One Shot Method converges to the right minimum. Lemma 4. If there exist constants α > 0 and β > 0 such that the following conditions are fulfilled α(1 − ρ) − α2 β &Gu &2 > 1 + &Nyy &, 2µ 2 β β(1 − ρ) > 1 + &Nyy &, 2 µ γ> , 2 (29a) (29b) (29c) then the One Shot iterate yields descent on La . Proof. We follow the proof of Lemma with the difference that for the descent direction s (−∇La , s) ≥ C ! s!2 , 11 with a constant C > 0 and the norm !s!2 := &s1 &2H 1 + &s2 &2H 1 + &s3 &2L2 needs 0 0 to be fulfilled. Hence for the bilinear form B(w, v, h) := −2(w, v)H01 + α(w, w − Gy w)H01 − β(v, Nyy w)H01 + β(v, v − Gy v)H01 + γ(h, h)L2 − α(w, Gu h)H01 with (w, v, h) ∈ H01 × H01 × L2 , we need to verify the existence of a constant C > 0 such that B(w, v, h) ≥ C(&w&2H 1 + &v&2H 1 + &h&2L2 ). 0 0 After applying Young’s inequality ab ≤ a2 νb2 + 2ν 2 with the parameter ν chosen as 1, ρ, &Nyy & and combination with (5) it follows that α&Gu &2 µ respectively and in α1 2 β α ρ&w&2H 1 − ρ &w&2H 1 − &Nyy &&v&2H 1 0 0 0 2 2ρ 2 1 β β + β&v&2H 1 − ρ&v&2H 1 − ρ2 &v&2H 1 0 0 0 2 2ρ B(w, v, h) ≥ −&w&2H 1 − &v&2H 1 + α&w&2H 1 − 0 − 0 0 β 1 &Nyy &2 &w&2H 1 0 2 &Nyy & µ α α α&Gu &2 &w&2H 1 − &Gu &2 &h&2L2 0 2 µ 2 α&Gu &2 β α2 = &w&2H 1 (α(1 − ρ) − &Gu &2 − 1 − &Nyy &) 0 2µ 2 β + &v&2H 1 (β(1 − ρ) − 1 − &Nyy &) 0 2 µ 2 + &h&L2 (γ − ). 2 + γ&h&2L2 − Therefore, from the assumption α(1 − ρ) − β α2 &Gu &2 − 1 − &Nyy & > 0, 2µ 2 β β(1 − ρ) − 1 − &Nyy & > 0, 2 µ γ − > 0, 2 we obtain positivity. Remark 3. From (29a) and (29b) we get estimates for α and β. From (29a) for example β α α[(1 − ρ) − &Gu &2 ] > 1 + &Nyy & > 1 > 0 (31) 2µ 2 it follows with α, β > 0 that (1 − ρ) − 2µ α &Gu &2 > 0 ⇒ α < (1 − ρ). 2µ &Gu &2 12 Additionally we get a lower bound, because by (31) it follows (1 − ρ) − α 1 &Gu &2 > . 2µ α 2ρµ α The case α < 1 yields (1 − ρ) − 2µ &Gu &2 > 1 and therefore α < − &G 2 , which u& results in a contradiction. Therefore we have α > 1. By the following lemma we even obtain a result stating the sufficiently decrease in a neighborhood of the minimizer. Lemma 5. Let the slightly stronger conditions than (29) β α2 &Gu &2 > 1 + CNyy 2µ 2 β β(1 − ρ) > 1 + CNyy 2 µ γ> 2 α(1 − ρ) − (32) and (26) be satisfied for a constant CNyy with &Nyy (y, p, u)& < CNyy , (33) that is independent of any iterates (yk , pk , uk ). Then the angle condition holds in a neighborhood U(y ∗ , u∗ ) of the optimum (y ∗ , u∗ ), i.e. (−∇La , s) ≥ C > 0. &∇La & ! s! (34) Proof. The proof is based on the inequality (29a) and (29b) that in general may depend on the iterates. Since the constant in the angle condition needs to be independent of the iterates, it is necessary to find suitable approximations of the critical terms &Gu (y, u)& and &Nyy (y, p, u)&. By definition this is fulfilled for the norm &Gu (y, u)& := &Gu (y, u)[h]&H01 = sup &h&L2 =1 sup &h&L2 =1 &(−∆)−1 h&H01 . (35) For &Nyy (y, p, u)& := &Nyy (y, p, u)[v, w]& sup &v&H 1 =&w&H 1 =1 0 = 0 | sup &v&H 1 =&w&H 1 =1 0 ≤ |1 + ! 0 ! vw dx + Ω ! pey vw dx| Ω pey dx| Ω it follows with a generic constant C that &Nyy (y, p, u)& ≤ C&p&L2 &ey &L2 . The term &ey &L2 is bounded due to (3) and with (6c) we get &Nyy (y, p, u)& ≤ C&u&L2 . 13 (36) For a given starting iterate (y0 , p0 , u0 ) we consider the level set N0 := {(y, p, u) ∈ H01 × H01 × L2 : La (y, p, u) ≤ La (y0 , p0 , u0 )}. (37) From (27) it follows that &u − u∗ &2L2 ≤ C(La (y0 , p0 , u0 ) − La (y ∗ , p∗ , u∗ )) for (y, p, u) ∈ N0 , particularly &u&2L2 ≤ C(La (y0 , p0 , u0 ) − La (y ∗ , p∗ , u∗ ) + &u∗ &2L2 ) and finally since the right hand side is independent of any iterates &Nyy (y, p, u)& ≤ CNyy . For the angle condition we get with the same arguments as in the proof of Lemma 4 and (33) that α2 β &Gu &2 − 1 − CNyy ) 0 2µ 2 β µ + &s2 &2H 1 (β(1 − ρ) − 1 − CNyy ) + (γ − )&s3 &2L2 , 0 2 2 (−∇La , s) ≥ &s1 &2H 1 (α(1 − ρ) − i.e. with (32) we receive (−∇La , s) ≥ δ∗ ! s!2 , but here the constant δ∗ is independent of any iterates. Additionally we get by definition &Lay & = sup &w&H 1 =1 |Lay w| = 0 ≤ sup &w&H 1 =1 0 sup &w&H 1 =1 |(s2 , w)H01 + α(s1 , (Gy − I)w)H01 + β(s2 , Nyy w)H01 | 0 ' ( &s2 &H01 &w&H01 + α(1 + ρ)&s1 &H01 &w&H01 + β&s2 &H01 &Nyy &&w&H01 = &s2 &H01 + α(1 + ρ)&s1 &H01 + β&Nyy &&s2 &H01 &Lap & = sup &v&H 1 =1 |Lap v| = 0 sup &v&H 1 =1 |(s1 , v)H01 + β(s2 , (Gy − I)v)H01 | 0 ≤ &s1 &H01 + β(1 + ρ)&s2 &H01 &Lau & = sup &h&L2 =1 |Lau h| = sup &h&H 1 =1 | − γ(s3 , h)L2 + α(s1 , Gu h)H01 | 0 ≤ γ&s3 &L2 + α&Gu &&s1 &H01 Therefore with (33) it follows * ) &∇La & ≤ &s1 &H01 [α(1 + ρ) + 1 + α&Gu &] + &s2 &H01 1 + βCNyy + β(1 + ρ) + γ&s3 &L2 ≤ C ! s!, with a constant C > 0. Finally we get δ∗ ! s!2 (−∇La , s) ≥ > 0. &∇La & ! s! C ! s ! !s! 14 4.2 The stationary viscous Burgers equation Theoretically it would be possible to do the same analysis for the Burgers equation as for the solid fuel ignition model by using F (y) = −∆. But this choice does not describe a numerical stable method and therefore we choose the One Shot iteration with F (y) = (−∆ + y · ∇) as follows: yk+1 = (−∆ + yk · ∇)−1 (uk ) (38a) −1 pk+1 = (−∆ − yk · ∇ − div(yk )) 1 uk+1 = uk − (µuk + pk ). γ T (yk − zd − (∇yk ) pk ) (38b) (38c) For the convergence analysis we define the Lagrangian as in the general setting to be L(y, p, u) := f (y, u)− < ϕ, y − G(y, u) >H −1 ,H01 (39) with the operator ϕ := −∆p − (y · ∇)p − div(y)p for the adjoint p ∈ H01 . Remark 4. One can obtain the optimality system in a standard way by differentiating the Lagrangian defined above. With the following relationships ! ! ((v · ∇)y)p dx = (∇y)T p v dx, Ω ! ((y · ∇)v)p dx = Ω Ω ! " yi ∂i vj pj dx = − ! " yi ∂i (pj )vj dx − ! ! " ∂i (yi )pj vj dx Ω i,j Ω i,j =− ∂i (yi pj )vj dx Ω i,j Ω i,j =− ! " (y · ∇)pv + div(y)pv dx Ω and (14) we obtain Ly (y, p, u)w = (y − zd , w)L2 − < ϕy w, y − G(y, u) >H −1 ,H01 − < ϕ, w − Gy (y, u)w >H −1 ,H01 ! = (y − zd , w)L2 − (−(w · ∇) − div(w))p(y − G(y, u)) dx ! − (−∆ − (y · ∇) − div(y))p(w − (−∆ + y · ∇)−1 (−(w · ∇)G(y, u))) dx ! = (y − zd , w)L2 − (w · ∇)y p − (w · ∇)G(y, u) p dx ! − p(−∆ + y · ∇)(w − (−∆ + y · ∇)−1 (−(w · ∇)G(y, u))) dx ! = (y − zd , w)L2 + (w · ∇)G(y, u) p − (∇y)T pw dx ! − p(−∆w + y · ∇w + (w · ∇)G(y, u)) ! = (y − zd , w)L2 − (−∆ − (y · ∇) − div(y))pw + (∇y)T pw dx 15 Finally Ly (y, p, u)w = ! (−∆ − (y · ∇) − div(y))[(−∆ − (y · ∇) − div(y))−1 (y − zd − (∇y)T p) − p]w dx Similarly we can compute the derivative with respect to the adjoint and the control. ! Lp (y, p, u)v = − (−∆ − (y · ∇) − div(y))v (y − G(y, u)) dx ! = − v (−∆ + y · ∇)(y − G(y, u)) dx ! = − v(−∆y + (y · ∇)y − u) dx. This yields the state equation. And finally ! Lu (y, p, u)h = µ(u, h)L2 + (−∆ − (y · ∇) − div(y))p Gu (y, u)h dx ! = µ(u, h)L2 + p(−∆ + y · ∇)Gu (y, u)h dx ! = (µu + p)h dx. The analysis is more complex in this case, because the operator F (y) depends on y. Again we can define the augmented Lagrangian as La (y, p, u) := L(y, p, u) + β α &G(y, u) − y&2H 1 + &Ny (y, p, u) − p&2H 1 0 0 2 2 For the derivative we obtain with the help of G(y, u) − y s1 s = s2 = Ny (y, p, u) − p − γ1 Lu (y, p, u) s3 Lay (y, p, u)w = Ly (y, p, u)w + α(G(y, u) − y, (Gy (y, u) − I)w)H01 + β(Ny (y, p, u) − p, Nyy (y, p, u)w)H01 ! = (−∆ − (y · ∇) − div(y))s2 w dx + α(s1 , (Gy (y, u) − I)w)H01 + β(s2 , Nyy (y, p, u)w)H01 Lap (y, p, u)v = Lp (y, p, u)v + β(Ny (y, p, u) − p, (Nyp (y, p, u) − I)v)H01 ! = v (−∆ + y · ∇)s1 dx + β(s2 , (Nyp (y, p, u) − I)v)H01 Lau (y, p, u)h = Lu (y, p, u)h + α(G(y, u) − y, Gu (y, u)h)H01 = −γ(s3 , h)L2 + α(s1 , Gu (y, u)h)H01 . 16 We proceed in the same way as in subsection 4.1. For s = 0 it follows ∇La = 0. In addition if the related bilinear form to −∇La = 0 is coercive, we obtain s = 0. The bilinear form reads ! B(w, v, h) := −2 v(−∆ + y · ∇)w dx + α(w, w − Gy w)H01 − β(v, Nyy w)H01 + β(v, v − Nyp v)H01 + γ(h, h)L2 − α(w, Gu h)H01 . For the subsequent approximations we need to assume that &Nyp & ≤ ρ < 1, since the contraction factor of the adjoint equation is not given as for the solid fuel ignition model, see Remark 1. Compared to the bilinear form in subsection 4.1 there is the additional term ! −2 (y · ∇)wv dx. Because of the continuous injection H01 +→ L4 we obtain ! −2 (y · ∇)wv dx ≥ −&y&2H 1 &w&2H 1 − C∗2 &v&2H 1 0 0 0 with a constant C∗ . Therefore β α2 &Gu &2 − 1 − &Nyy & − &y&2H 1 ) 0 0 2µ 2 β + &v&2H 1 (β(1 − ρ) − 1 − &Nyy & − C∗2 ) 0 2 µ 2 + &h&L2 (γ − ) 2 B(w, v, h) ≥ &w&2H 1 (α(1 − ρ) − and finally Lemma 6. If there exist constants α > 0 and β > 0 such that the following conditions are fulfilled α(1 − ρ) − α2 β &Gu &2 > 1 + &Nyy & + &y&2H 1 , 0 2µ 2 β β(1 − ρ) > 1 + &Nyy & + C∗2 , 2 µ γ> , 2 (40a) (40b) (40c) and < F (y)∗ p, G(y, u)−y >H −1 ,H01 ≤ β α &G(y, u)−y&2H 1 + &Ny (y, p, u)−p&2H 1 (41) 0 0 2 2 then the One Shot Method converges to a local minimum of problem (P2). (41) is the analogue to (27) for the general setting. You need this condition to obtain a local minimum of the augmented Lagrangian. 17 5 Numerical Results In this section we test the One Shot Method for the two example problems, the solid fuel ignition model and the viscous Burgers equation. 5.1 5.1.1 The solid fuel ignition model Discretization Let Ω = [0, 1]2 be subdivided into shape-regular meshes Th consisting of elements K. The optimization problem is solved by discretizing the One Shot Method (22) with continuous piecewise linear finite elements. Let V h := {v h ∈ C0 (Ω)| v h |K ∈ P1 (K), K ∈ Th }, then we consider y h , ph , uh ∈ V h as the finite element approximations to the primal y, the dual p and the control u. Therefore in every iteration step of the One Shot Method we solve the finite element problem h h a(yk+1 , v h ) = (δ eyk + uhk , v h ), v h ∈ V h a(phk+1 , v h ) = h (δ phk eyk uhk+1 = uhk − + ykh h (42a) h − zd , v ), v ∈ V h 1 (µuhk + phk ). γ Here a(·, ·) denotes the bilinear form ! a(u, v) = ∇u · ∇v dx, Ω (42b) (42c) u, v ∈ H01 and (·, ·) the standard L2 scalar product. 5.1.2 Optimization We update y h , ph , uh according to (42). As a stopping criteria we verify that the residual of the coupled iteration in the L2 -norm is less than some given tolerance tol, i.e. h res = &yk+1 − ykh &L2 + &phk+1 − phk &L2 + &uhk+1 − uhk &L2 ≤ tol. All the subsequent results are obtained with the help of the finite element library deal.II. The linear problem in each iteration step is solved with a conjugated gradient method. We first set u0 ≡ 1, y0 ≡ 1, p0 ≡ 0, tol = 10−3 and zd (x1 , x2 ) = x1 x2 . The numerical results for µ = γ = 0.005 and δ = 1 can be seen in Figure 1 on page 19. The choice of δ has a huge effect on the solutions. The control for various δ and µ = γ = 0.01 is shown in Figure 2 on page 20. 18 1 0.2 0.18 0.16 0.14 0.12 0.1 0.08 0.06 0.04 0.02 0 0.8 0.6 0.4 0.2 0 1 1 0.8 0 0.8 0.6 0.2 0.4 0 0.4 0.6 0.8 0.6 0.2 0.2 0.4 0.4 0.6 1 0 (a) Desired state 0.8 0.2 1 0 (b) State 3 0 2.5 -0.002 -0.004 2 -0.006 1.5 -0.008 1 -0.01 0.5 -0.012 0 -0.014 1 1 0.8 0 0.8 0.6 0.2 0.4 0 0.4 0.6 0.8 0.2 0.6 0.2 0.4 1 0 (c) Control 0.4 0.6 0.8 0.2 1 0 (d) Dual Figure 1: Numerical Solution for µ = γ = 0.005, tol = 10−3 and δ = 1 Additionally the method seems not to be robust with respect to the parameters δ, µ and the desired state zd . The biggest regularization parameter that we obtain results for is µ = 0.005. In [2] the authors experienced similar problems. They solved the discretized optimality system with a multigrid One Shot approach. The number of iterations for a complete One Shot optimization are displayed in Tables 1 to 3 for various regularization parameters µ, preconditioners γ and degrees of freedom. Here δ is fixed to 1. ∞ indicates divergence of the One Shot Method or that the maximal number of iterations was reached. # Dofs 81 289 1089 4225 γ = 0.04 ∞ ∞ ∞ ∞ γ = 0.05 ∞ ∞ ∞ ∞ γ = 0.06 19 19 19 19 γ = 0.08 9 9 9 9 γ = 0.1 8 8 8 8 γ = 0.2 10 10 10 10 γ = 0.5 24 24 24 24 Table 1: Number of iterations for Solid Fuel Ignition Model with µ = 0.1, δ = 1 and tol = 10−3 19 γ=1 44 44 44 44 1.6 1.4 1.2 1 0.8 0.6 0.4 0.2 0 1.4 1.2 1 0.8 0.6 0.4 0.2 0 -0.2 1 1 0.8 0 0.8 0.6 0.2 0.4 0 0.4 0.6 0.8 0.6 0.2 0.2 0.4 0.4 0.6 1 0 (a) δ = 1 0.8 0.2 1 0 (b) δ = 2 1.2 1 0.8 0.6 0.4 0.2 0 -0.2 -0.4 -0.6 -0.8 1 0.8 0.6 0.4 0.2 0 -0.2 1 1 0.8 0 0.8 0.6 0.2 0.4 0 0.4 0.6 0.8 0.6 0.2 0.2 0.4 0.4 0.6 1 0 (c) δ = 3 0.8 0.2 1 0 (d) δ = 4 0.5 0.6 0.4 0.2 0 -0.2 -0.4 -0.6 -0.8 -1 -1.2 -1.4 0 -0.5 -1 -1.5 -2 -2.5 1 1 0.8 0 0.8 0.6 0.2 0.4 0 0.4 0.6 0.8 0.2 1 0 (e) δ = 5 0.6 0.2 0.4 0.4 0.6 0.8 0.2 1 0 (f) δ = 6 Figure 2: Control for µ = γ = 0.01 and tol = 10−3 20 #Dofs 81 289 1089 4225 γ = 0.004 ∞ ∞ ∞ ∞ γ = 0.005 ∞ ∞ ∞ ∞ γ = 0.007 56 57 58 58 γ = 0.008 29 30 30 30 γ = 0.01 21 21 21 21 γ = 0.02 16 16 16 16 Table 2: Number of iterations for Solid Fuel Ignition Model with µ = 0.01, δ = 1 and tol = 10−3 #Dofs 81 289 1089 4225 γ = 0.0025 ∞ ∞ ∞ ∞ γ = 0.004 96 102 104 104 γ = 0.005 40 40 40 40 γ = 0.01 27 27 27 27 γ = 0.1 70 70 70 70 γ=1 271 271 271 271 Table 3: Number of iterations for Solid Fuel Ignition Model with µ = 0.005, δ = 1 and tol = 10−3 For all choices of µ and γ the number of iterations seems to be independent of the number of grid points. The best results are obtained for γ close to µ, more precisely for µ = 0.1 the best results are obtained for γ = µ, but for µ = 0.01 and µ = 0.005 they are obtained for γ = 2µ. In all three cases the numerical results reflect the condition (40c) in Lemma 4. The method diverges if the condition is violated, i.e. γ ≤ µ/2. 5.1.3 Adaptive scheme The numerical mesh independence results above encourage the idea of an adaptive One Shot strategy. Therefore we extend (42) by an additional adaptive step, i.e. we refine or coarsen the grid according to an error estimator and interpolate the solutions to the new mesh. For the refinement process we chose a residual type error estimator for the cells and the faces with respect to the state and adjoint state. 1 " hE & [n · ∇y]E &2E 2 E∈∂K 1 " := h2K &∆p + δ pey + y − zd &2K + hE & [n · ∇p]E &2E 2 2 ηK,y := h2K &∆y + δ ey + u&2K + 2 ηK,p E∈∂K 2 2 2 ηK := ηK,y + ηK,p Overall the Adapted One Shot Method is stated in Algorithm 2 on page 23. The 40% of the cells with the largest error indicator are refined, those 30% with the smallest are coarsened. This error estimater may not be optimal for the problem but serves as a good test. The final adaptively generated grids are plotted for µ = γ = 0.1, δ = 1 in Figure 3 on page 22 and δ = 2 in Figure 4 on page 22 respectively. 21 γ = 0.1 40 40 40 40 γ=1 181 180 180 179 Figure 3: Adaptively generated grid for µ = γ = 0.1 and δ = 1 Figure 4: Adaptively generated grid for µ = γ = 0.1 and δ = 2 22 Algorithm 2 Adapted One Shot Method 1: 2: 3: 4: 5: 6: 7: 8: 9: 10: 11: Choose u0 , y0 , p0 , tol > 0 for k = 0, 1, 2, . . . do h calculate yk+1 , phk+1 and uhk+1 using (42) if res < tol1 then calculate ηK for all K and refine or coarsen the mesh interpolate the solution vectors to the new mesh end if if res < tol2 then STOP end if end for 5.2 The stationary viscous Burgers equation For the discretization of the Burgers equation a SUPG scheme was implemented using deal.II. For the 1D case we solve in every step of the One Shot Method the following problem !! ! −νyk+1 + yk yk+1 = uk −νp!!k+1 − yk p!k+1 = yk − zd 1 uk+1 = uk − (µuk + pk ). γ (43) As a first example we choose zd = 1, µ = 0.1, ν = 0.1 and ν = 0.01. The solutions for the second case can be seen in Figure 5 on page 24. Preconditioning the linear system in every iteration step, that is solved using the GMRES method, seems to be very important for the Burgers equation. While for the first example with µ = 0.1 and ν = 0.1 the SSOR preconditioner in deal.II worked fine, the preconditioner had to be improved for the solution with ν = 0.01. deal.II offers a wrap around to the PETSc library, which itself offers various solvers and preconditioners. Therefore whenever possible we solved the system using deal.II, but for comparison in this case we also computed the solutions with PETSc using an ILU preconditioner. Otherwise in case the deal.II solver failed, we switched to PETSc completely. Tables 4 and 5 show the number of iterations for various degrees of freedom and preconditioners γ with ν = 0.1. In the first case the deal.II solver was used and in the second case a PETSc solver was used. #Dofs 33 65 129 257 γ = 0.1 ∞ ∞ ∞ ∞ γ = 0.5 ∞ ∞ ∞ ∞ γ = 0.8 90 90 90 90 γ=1 65 65 65 65 γ = 1.2 65 65 65 65 γ=2 98 98 98 98 γ=3 137 136 136 136 γ=5 205 205 205 205 Table 4: Number of iterations for 1D Burgers equation with µ = 0.1, ν = 0.1, zd = 1, without PETSc 23 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 0 0.2 0.4 0.6 0.8 1 (a) State 2.5 0 2 -0.05 1.5 -0.1 1 -0.15 0.5 -0.2 0 -0.25 0 0.2 0.4 0.6 0.8 1 0 0.2 (b) Control 0.4 0.6 0.8 1 (c) Dual Figure 5: Solution to 1D Burgers equation for µ = 0.1 and ν = 0.01 #Dofs 33 65 129 257 γ = 0.1 ∞ ∞ ∞ ∞ γ = 0.5 ∞ ∞ ∞ ∞ γ = 0.8 85 88 91 95 γ=1 62 64 66 69 γ = 1.2 61 63 66 70 γ=2 91 96 101 107 γ=3 126 133 141 150 γ=5 186 199 212 227 Table 5: Number of iterations for 1D Burgers equation with µ = 0.1, ν = 0.1, zd = 1, with PETSc The mesh independence can be clearly seen in the case of the deal.II solver. On the contrary PETSc seems to destroy the mesh independent behaviour. For ν = 0.01 the results can be seen in Table 6. A small increase in the number of iterations can be observed, which again might be the influence of PETSc. As a second example we chose µ = 0.1, ν = 0.1 and zd = sin(13x). The solutions are displayed in Figure 6 on page 25, the number of iterations in Tables 7 and 8. Again the difference in the mesh independent behaviour between deal.II and PETSc solver is clear. 24 #Dofs 33 65 129 257 γ=1 ∞ ∞ ∞ ∞ γ=2 162 159 161 164 γ = 2.5 121 123 127 134 γ=3 128 134 142 150 γ=4 158 166 176 187 γ=5 187 197 209 224 Table 6: Number of iterations for 1D Burgers equation with µ = 0.1, ν = 0.01, zd = 1, with PETSc 1 0.04 0.8 0.03 0.6 0.02 0.01 0.4 0 0.2 -0.01 0 -0.02 -0.2 -0.03 -0.4 -0.04 -0.6 -0.05 -0.8 -0.06 -1 0 0.2 0.4 0.6 0.8 1 -0.07 0 0.2 (a) Desired state 0.4 0.6 0.8 1 0.6 0.8 1 (b) State 0.6 0.08 0.4 0.06 0.2 0.04 0 0.02 -0.2 0 -0.4 -0.02 -0.6 -0.04 -0.8 -0.06 0 0.2 0.4 0.6 0.8 1 0 0.2 (c) Control 0.4 (d) Dual Figure 6: Solution to 1D Burgers equation for µ = 0.1, ν = 0.1 and zd = sin(13x) #Dofs 33 65 129 257 γ=1 ∞ ∞ ∞ ∞ γ=2 205 210 210 210 γ = 2.5 134 134 134 134 γ=3 155 156 156 156 γ=4 197 197 197 197 γ=5 235 235 235 235 Table 7: Number of iterations for 1D Burgers equation with µ = 0.1, ν = 0.1, zd = sin(13x), without PETSc 25 #Dofs 33 65 129 257 γ=1 ∞ ∞ ∞ ∞ γ=2 190 200 215 225 γ = 2.5 122 129 137 145 γ=3 141 150 159 169 γ=4 177 189 202 215 γ=5 211 226 242 258 Table 8: Number of iterations for 1D Burgers equation with µ = 0.1, ν = 0.1, zd = sin(13x), with PETSc As a next step the Burgers equation in two space dimensions is solved. In every iteration step the following linear system needs to be solved: (−∆ + yk · ∇)yk+1 = uk (−∆ − yk · ∇ − div(yk ))pk+1 = yk − zd − (∇yk )T pk 1 uk+1 = uk − (µuk + pk ). γ The results for the 2D Burgers equation with µ = 0.1 and ν = 0.1 are shown in Tables 9 and 10. Here the solver is more sensitive to changes in degrees of freedom compared to the one-dimensional case. But at least for the best value of γ = 1.2 the deal.II solver seems to behave mesh independent. #Dofs 882 1922 3362 5202 γ = 0.6 ∞ ∞ ∞ ∞ γ = 0.7 ∞ 170 142 130 γ = 0.8 129 98 88 84 γ=1 69 59 56 55 γ = 1.2 55 52 52 51 γ=2 74 73 73 73 Table 9: Number of iterations for the 2D Burgers equation with µ = 0.1 and ν = 0.1, without PETSc #Dofs 882 1922 3362 5202 γ = 0.6 ∞ ∞ ∞ ∞ γ = 0.7 335 224 199 189 γ = 0.8 166 136 128 125 γ=1 94 86 85 86 γ = 1.2 87 90 93 95 γ=2 134 141 146 150 Table 10: Number of iterations for the 2D Burgers equation with µ = 0.1 and ν = 0.1, with PETSc We add adaptation steps as in Algorithm 2. After 3 adaptive steps with 40% of the cells with the largest gradient jumps being refined and 8% of the cells with the smallest being coarsened, the grid exhibits a fine region near the boundary. The solution on this adaptively generated grid is illustrated in Figure 7. 26 1 1 v1 0.4 0.2 0 0.2 0.4 x 0.6 0.8 0.85 0.8 0.75 0.7 0.65 0.6 0.55 0.5 0.45 0.4 0.35 0.3 0.25 0.2 0.15 0.1 0.05 0.8 0.6 y y 0.6 0 v2 0.85 0.8 0.75 0.7 0.65 0.6 0.55 0.5 0.45 0.4 0.35 0.3 0.25 0.2 0.15 0.1 0.05 0.8 0.4 0.2 0 1 (a) First component of Primal 0 0.2 0.4 x 0.6 0.8 1 (b) Second component of Primal 1 1 v1 v2 2.4 2.2 2 1.8 1.6 1.4 1.2 1 0.8 0.6 0.4 0.2 y 0.6 0.4 0.6 0.4 0.2 0 2.4 2.2 2 1.8 1.6 1.4 1.2 1 0.8 0.6 0.4 0.2 0.8 y 0.8 0.2 0 0.2 0.4 x 0.6 0.8 0 1 (c) First component of Control 0 0.2 0.4 x 0.6 0.8 1 (d) Second component of Control 1 1 v1 v2 -0.02 -0.04 -0.06 -0.08 -0.1 -0.12 -0.14 -0.16 -0.18 -0.2 -0.22 -0.24 y 0.6 0.4 0.6 0.4 0.2 0 -0.02 -0.04 -0.06 -0.08 -0.1 -0.12 -0.14 -0.16 -0.18 -0.2 -0.22 -0.24 0.8 y 0.8 0.2 0 0.2 0.4 x 0.6 0.8 0 1 (e) First component of Dual 0 0.2 0.4 x 0.6 0.8 1 (f) Second component of Dual Figure 7: Solution to 2D Burgers equation on adaptively refined grid for µ = 0.1, ν = 0.1 and zd = 1 27 Additionally the One Shot Method is tested for µ = 0.01 and ν = 0.1. The results are displayed in Table 11. Again the grid is adapted in the same way as above and plotted in Figure 8. #Dofs 882 1922 3362 5202 γ = 0.6 ∞ ∞ ∞ ∞ γ = 0.7 ∞ 431 438 446 γ = 0.75 455 452 461 471 γ = 0.8 473 475 486 497 γ=1 564 571 586 599 Table 11: Number of iterations for the 2D Burgers equation with µ = 0.01 and ν = 0.1, with PETSc Figure 8: Adaptively refined grid for 2D Burgers equation with µ = 0.01 and ν = 0.1 6 Conclusions In this paper we studied the One Shot Method in a Hilbert space setting. For two example problems, the solid fuel ignition Model and the Burgers equation, we set up a convergence analysis. In detail, we showed that an augmented Lagrange function is an exact penalty function and that the One Shot iteration serves as a descent direction to the augmented Lagrangian. The numerical results show the convergence for a variety of preconditioners. Additionally it can be seen, that the method is mesh independent. Finally we extended the One Shot Method by an additional adaptive step. 28 References [1] J. M. Burgers, Application of a model system to illustrate some points of the statistical theory of free turbulence, Proc. Acad. Sci. Amsterdam, 43:2-12, 1940. [2] A. Borzi and K. Kunisch, The numerical solution of the steady state solid fuel ignition model and its optimal control, SIAM J. SCI. COMPUT., Vol.22, No. 1, pp. 263-284. [3] E. Casas, O. Kavian and J.-P. Puel, Optimal control of an ill-posed elliptic semilinear equation with an exponential nonlinearity, ESAIM Control, Optim. Calc. Var., Vol. 3, 1998, pp. 361-380. [4] J.C. De Los Reyes, Constrained Optimal control of Stationary Viscous Incompressible Fluids by Primal-Dual Active Set Methods, Dissertation Uni Graz, 2003 . [5] A. Hamdi and A. Griewank, Properties of an augmented Lagrangian for design optimization, Preprint-Number SPP1253-11-01. [6] A. Hamdi and A. Griewank, Reduced Quasi-Newton Method for simultaneous design and optimization, Preprint-Number SPP1253-11-02. [7] E. Hopf, The partial differential equation ut + uu! = µu!! , Communications an Pure and Applied Mathematics, pages 201-230, 1950. [8] K. Ito and K. Kunisch, Optimal control of the solid fuel ignition model with H 1 -cost, SIAM J. CONTROL OPTIM., Vol.40, No. 5, pp. 1455-1472. [9] A. Kauffmann, Optimal control of the solid fuel ignition model, Dissertation TU Berlin, 1998 . [10] A. Kauffmann and K. Kunisch, Optimal control of the solid fuel ignition model, ESAIM:PROCEEDINGS, Vol.8, pp.65-76, 2000. [11] E. Özkaya and N. Gauger, Single-step One-Shot aerodynamic Shape Optimization, Preprint-Number SPP1253-10-04. [12] E. Özkaya and N. Gauger, Automatic transition from simulation to oneshot shape optimization with Navier-Stokes equation , GAMM-Mitt. 33, No. 2, 133-147. [13] S. Ta’asan, One-shot methods for optimal control of distributed parameter systems I:Finite dimensional control, ICASE 91-2, 1991. [14] F. Tröltzsch, Vieweg, 2005. Optimale Steuerung partieller Differentialgleichungen, [15] S. Volkwein, Mesh independence of an augmented Lagrangian-SQP method in Hilbert spaces and control problems for the Burgers equation, Dissertation, 1997. 29
© Copyright 2026 Paperzz