Evolutionary and Deterministic Methods for Design, Optimization and Control with Applications to Industrial and Societal Problems EUROGEN 2005 R. Schilling, W.Haase, J. Periaux, H. Baier, G. Bugeda (Eds) c FLM, Munich, 2005 PROGRESS IN METAMODEL ASSISTED EVOLUTIONARY ALGORITHMS WITH APPLICATIONS IN EXTERNAL AERODYNAMICS AND TURBOMACHINERY Kyriakos C. Giannakoglou† † Associate Professor, National Technical University of Athens, Lab. of Thermal Turbomachines, P.O. Box 64069, Athens 157 10, GREECE, [email protected] Key words: Optimization, Evolutionary Algorithms, Surrogate Evaluation Models. Abstract. Research activities of the Design Optimization Group ( DOG) of the LTT/ NTUA concerning the development and assessment of design optimization methods and tools are summarized below. These are related to both evolutionary and deterministic (based on the adjoint method) optimization methods. Particular emphasis is laid on the use of computational intelligence to enhance the performance of evolutionary algorithms ( EAs), by reducing their CPU cost when used along with computationally expensive evaluation software. 1 Kyriakos C. Giannakoglou Cost Reduction in EAs The first section presents enhanced EA–based optimization methods which may handle complex engineering problems with affordable computing cost. These are associated with either single– (SOO) or multi–objective optimization (MOO) problems and become advantageous whenever the evaluation candidate solutions (through CFD codes, for instance) is costly. Reduction of the optimization CPU cost is equivalent to the reduction of the number of fitness evaluations required by the search method. Procedures which can reduce the CPU cost of EAs are mainly based on the use of (low–cost, approximate) surrogate evaluation models or metamodels. Through their use, the number of calls to the costly, accurate evaluation tool (such as a CFD code) can be kept low. Response surface models with low-order polynomials, artificial neural networks (ANNs) and kriging are known metamodels. Existing metamodel–assisted EAs (MAEAs) differ in the way metamodels are implemented. DOG/LTT/NTUA proposed the use of local, on–line trained metamodels, each of which corresponds to a single population member, [1, 2, 3]. The proposed algorithm relies on the concept of the Inexact Pre-Evaluation (IPE ) filter, [2, 3, 4, 5, 6, 7, 8]. In each generation, low-cost, local metamodels, trained on subsets of the previously evaluated patterns, are used to screen the population and pinpoint the top (most promising) individuals which are worth being exactly evaluated. A separate metamodel is associated with each and every new candidate solution. By means of the IPE, in SOO problems, the CPU cost per generation is kept even an order of magnitude lower than by standard EAs. Two further enhancements of MAEAs (abbreviated to EA-IPE-IF and D(EA-IPE-IF)) based on on–line trained metamodels are presented below. 1 1.1 EA-IPE-IF An enhanced EA-IPE scheme is the so–called EA-IPE-IF one, [4, 5]. This EA-IPE variant utilizes radial basis function networks (RBFNs) as metamodels. The EA-IPEIF concept is as follows: Modified neural networks (RBFNs) are trained to guess the response and its gradient with respect to the inlet parameters. During the training, these RBFNs take also into account the sensitivity of the response with respect to each one of the design variables (i.e. their input parameters). As sensitivity measures, the partial derivatives of the objective function with respect to the design variables are used. It is important to make clear that the required derivatives are computed by the RBFNs and, at the same time, used for their training. EA-IPE-IF outperforms EA-IPE which, in turns has a much better performance than conventional EAs. 1.2 D(EA-IPE-IF) The second promising method for lowering the cost of EAs is through their distributed variants (Distributed EAs, DEAs). DEAs rely on handling population subsets that evolve in a semi-isolated manner, by regularly exchanging their best individuals. Distributed schemes generally outperform single-population ones. The new idea is to employ 2 Kyriakos C. Giannakoglou metamodels in DEAs, [5], especially on a cluster of interconnected processors. The new algorithm is denoted by D(EA-IPE-IF) and incorporates all the aforementioned techniques. 1.3 A SOO Application An indicative comparison of the previously presented algorithmic variants, which is representative of a huge class of SOO problems, is presented. The relative positioning of a three–element airfoil components, that yields maximum lift, is sought. The six design variables determine the relative position of the slat and flap with respect to the main body, [9]. In fig. 1, the optimal airfoil configuration with computed iso–Mach contours, convergence plots for the four aforementioned algorithms as well as the computed IFs are illustrated. The gain achieved can be interpreted in two different ways: the cost of EA– IPE for reaching the solution quality of standard EAs is much lower or, by predetermining the allowable CPU cost (i.e. maximum number of evaluations), a better solution can be computed through EA–IPE. It is also clear that D(EA-IPE-IF) outperforms any other variant. 5.3 Conventional EA EA-IPE EA-IPE-IF D(EA-IPE-IF) Lift Coefficient 5.25 5.2 5.15 5.1 5.05 0 200 400 600 800 1000 Exact Evaluations (~ CPU Cost) Relative Importance 0.3 0.25 0.2 0.15 0.1 0.05 0 1 2 3 4 Design variables 5 6 Figure 1: Optimization of the three–element airfoil. EA-IPE in MOO Our experience has shown that, in MOO, the gain in CPU cost through the “standard“ (i.e. operating exactly as in SOO) IPE technique is not as noticeable as in SOO. Main reasons are: (a) in SOO problems, the population soon becomes clustered around the global optimum, in the vicinity of which adequate information exists to train the metamodels, which is not the case in MOO and (b) in MOO, the exact re-evaluation of just a few population members per generation has negative impact on the number of solutions which enter the Pareto front. 2 3 Kyriakos C. Giannakoglou In order to improve the performance of the EA–IPE method in MOO, one should proceed through carefully selecting the training patterns for the local metamodels (here RBFNs), identifying and differently treating outliers, i.e. new solutions lying far from the database (DB ) entries, selecting the RBF centers, etc. So: (a) Selection of Training Patterns - Outliers’ Identification: In “standard” IPE the selection of training patterns relies on distance–based criteria, with distances measured in the design variable space. A selection scheme which is more suitable for MOO is based on the formation of the minimum spanning tree. Through “walking” along the tree branches while moving away from the new individual, decisions are made about which points are to be used as training patterns, [10]. Through the same algorithm and the corresponding criteria, outlying individuals can be identified. For the outliers, the metamodel is expected to give rather inaccurate predictions and, thus, the exact evaluation tool should be used instead. (b) Selection of RBF centers and radii: It is known that RBFNs can be used as either interpolation or approximation tools and this can be determined through appropriate selection of their centers. So, EA-IPE for MOO should be enhanced through a new mechanism, which automatically decides about whether the approximate fitness value will be approximated or interpolated, [10]. Self-organizing maps (SOM ) is the heart of this mechanism, [10]. 2.1 A MOO Application An isolated airfoil shape optimization problem is presented. The objectives are to maximize CL and minimize CD at Re∞ = 6.2 · 106 , M∞ = 0.75 and a∞ = 2.734o. Starting shape is the RAE2822 airfoil. At these flow conditions, the starting airfoil yields equal to CL,ref = 0.749 and CD,ref = 0.0235. With the same CPU cost, the “standard” EA-IPE algorithm, which employs simple RBFN fails to achieve a better front than conventional EAs. The proposed, adaptive metamodel yields much better estimations and results to a considerably better front than other variants, fig. 2. 3,0x10 4 2,5x10 0 , 0 3 5 0 , 0 3 0 0 , 0 2 5 0 , 0 2 0 0 , 0 1 5 0 , 0 1 0 Conventional EA EA-IPE with simple RBFNs EA-IPE with the proposed RBFNs Reference airfoil 1500 Exact Evaluations 2,0x10 1,5x10 1,0x10 4 CD Non-Dominated Grid Points Conventional EA EA-IPE with simple RBFNs EA-IPE with the proposed RBFNs 4 4 4 1 , 0 500 1000 2 1 , 0 0 , 8 0 , 6 0 , 4 1500 -CL Evaluations Figure 2: Two–objective airfoil optimization. 3 Development of Adjoint Methods Continuous and discrete adjoint formulations for the inverse design of aerodynamic shapes and the minimization of viscous losses in turbomachinery cascades, have also been 4 Kyriakos C. Giannakoglou developed. With respect to the latter, viscous losses are expressed in terms of entropy generation caused by the boundary layer formation. The minimization of entropy s generation is expressed as the difference of incoming and outgoing mass averaged entropy, namely Z F = Si dṁsdS − Z So dṁsdS = Z S ρVn sdS where Si,o are the inlet and outlet boundaries. Using the Gauss’ divergence theorem and the continuity equation, this equation is transformed to a field integral as follows F = Z Ω ρui ∂s dΩ ∂xi and can also be written as F = Z Ω 1 ∂ui τij dΩ T ∂xj (ui are the velocity components, τij are the stress components). During the formulation of the adjoint equations, the field sensitivities are eliminated using the implicit function theory together with the Gauss’ divergence theorem and the final gradient formula depends solely on surface terms. For the augmented objective function Faug , formed after − → introducing the vector of costate variables Ψ , δFaug − →inv − →vis ∂fi − →T ∂ f i = δF + Ψ δ( − )dΩ ∂xi ∂xi Ω Z the appropriate development of terms yields the objective function gradient, as follows →T Z − ∂U → − →T − − → δFaug=− Ψ f i δ(ni dS)− Ai Tni Ψ δxk dS+ Sw Sw∂xk − →vis Z Z − →T ∂ f i δxk ni dS + Ψm qj δ(nj dS)− Ψ ∂xk Sw Sw " !# Z ∂ui ∂Ψj+1 ∂Ψm ∂Ψi+1 ∂Ψm +uj + +ui µ δxl nj dS− ∂xi ∂xi ∂xj ∂xj Sw∂xl " !# Z ∂ui ∂Ψk+1 ∂Ψm λδij +uk δxl nj dS− ∂xk ∂xk Sw ∂xl Z Z ∂uj 1 Rij ni δxk dS+ Rδxi ni dS ∂xk S S T Z ∂ui where Ai are the Jacobian matrices of the inviscid fluxes, R = τij ∂x and j Rij = 2(1 + δij ) ∂ui ∂uj 4 ∂uk + 2(1 − δij ) − δij ∂xj ∂xi 3 ∂xk 5 Kyriakos C. Giannakoglou y The proposed adjoint method is used to design a compressor cascade airfoil with minimum losses at laminar flow conditions. Geometrical constraints are taken into account in order to avoid creating thin airfoils. Results are shown in figs. 3, 4 and 5. 0.5 0.45 0.4 0.35 0.3 0.25 0.2 0.15 Initial Optimal 0.1 Initial 0.05 Optimal 0 -0.05 -0.1 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 x 0.0007 0.000698 0.000696 0.000694 0.000692 0.00069 0.000688 0.000686 0.000684 0.000682 0.00068 0.0174 0.0172 0.017 Pt Losses Entropy Generation Figure 3: Initial and optimal blade contour and control points. 0.0168 0.0166 0.0164 0.0162 0.016 0.0158 0.0156 0 20 40 60 80 100 120 140 160 0 20 40 Iteration 60 80 100 120 140 160 Iteration Figure 4: Optimization of a compressor blade: convergence plots. 0.0008 Continuous Adjoint Finite Differences 0.0006 Gradient 0.0004 0.0002 0 -0.0002 -0.0004 -0.0006 -0.0008 0 5 10 15 Variable 20 25 Figure 5: Optimization of a compressor blade: Objective function gradient. EAs with Gradient-Assisted Metamodels Recently, new metamodels trained by taking into consideration available information on both responses and their gradients have been proposed, [11, 12]. These will be referred to as gradient–assisted multi–layer perceptron (GAMLP) and gradient–assisted radial– basis function (GARBF) networks. These can be readily incorporated into MAEAs, in particular those based on off-line (separately) trained metamodels. For such a MAEA to be efficient, the additional cost for computing gradients (here, through the adjoint method) and training the metamodel using both responses and gradients should be overbalanced by the considerable reduction in the training set size. Note that the role of the adjoint equation solner is to compute objective function gradients to be used for training the metamodel, rather than to guide a descent optimization algorithm. The EA undergoes the search for the optimal solution(s). 4 6 Kyriakos C. Giannakoglou The inverse design of a turbine peripheral cascade is presented, fig 6. The target is to capture a known pressure distribution over its surfaces (parameterized using two Bezier surfaces with 33 control points each) at given flow conditions. The 3D Euler and the corresponding adjoint equations are solved for each candidate blade cascade. 0.0001 IPE_RBF(15,50,10) IPE_GARBF(15,50,10) Steepest Descent GARBF_OFFLINE_DB Cost Value 1e-05 1e-06 1e-07 1e-08 0 100 200 300 400 500 600 Equivalent Flow Solutions 700 800 Figure 6: Design of a 3D turbine cascade. Fig. 6 illustrates the convergence of the four optimization tools used. The first two methods employ the IPE technique, in which the metamodels are (a) conventional RBF and (b) GARBF networks. The cost for an exact evaluation using the IPE -GARBF method is multiplied by two, to account for the cost of the adjoint equations solver. Towards convergence, the additional information (gradient) used by the IPE –GARBF enhances the metamodel prediction abilities and faster convergence is achieved. The standard steepest descent method (third curve, figure 6), with gradient computed by the adjoint method, yields a solution with very small CPU cost. The fourth method is EAs supported by off–line trained GARBF. For the latter, an initial DB with 100 patterns is formed and objective function values and gradients are computed. Using the so–trained GARBF, the EA computes an ”optimal” solution with almost negligible cost. The ”optimal” solution is re–evaluated and the DB is updated. The metamodel is re– trained and a new cycle is carried out. Eight cycles are performed at the cost of 216 equivalent flow solutions. All EA or EA-IPE -based computations were performed using the optimization software EASY developed and brought to market by DOG/LTT/NTUA. Acknowledgement: This research program was co-funded by the European Social Funds (75%) and National Resources (25%) (Program PYTHAGORAS I). 7 Kyriakos C. Giannakoglou REFERENCES [1] Giannakoglou, K., ”Designing Turbomachinery Blades Using Evolutionary Methods”, ASME Paper 99-GT-181, 1999. [2] Giannakoglou, K., ”Acceleration of Genetic Algorithms Using Artificial Neural NetworksTheoretical Background, Von-Karman Institute Lecture Series (VKI-LS) 2000-07, 2000. [3] Giannakoglou, K., Giotis, A., ”Acceleration of Genetic Algorithms Using Artificial Neural Networks-Application of the Method”, VKI-LS 2000-07, 2000. [4] Giannakoglou, K., Giotis, A., Karakasis, M., ”Low–Cost Genetic Optimization Based on Inexact Pre–Evaluations and the Sensitivity Analysis of Design Parameters”, J. Inverse Problems in Eng., 9, 389-412, 2001. [5] Karakasis, M., Giotis, A., Giannakoglou, K., ”Inexact Information Aided, Low-Cost, Distributed Genetic Algorithms for Aerodynamic Shape Optimization”, Int. J. Num. Meth. Fluids, 43, 1149-1166, 2003. [6] Chiba, K., Obayashi, S., Nakahashi, K., Giotis, A., Giannakoglou, K., ”Design Optimization of the Wing Shape of the RLV Booster Stage using Evolutionary Algorithms and Navier-Stokes Computations on Unstructured Grids”, EUROGEN 2003, Barcelona, 2003. [7] Giannakoglou, K., ”Neural Network Assisted Evolutionary Algorithms, in Aeronautics and Turbomachinery”, VKI-LS 2004-07, 2004. [8] Giannakoglou, K., Papadimitriou, D., Kampolis, I., ”Coupling Evolutionary Algorithms, Surrogate Models and Adjoint Methods in Inverse Design and Optimization Problems”, VKI-LS 2004-07, 2004. [9] Giannakoglou, K., ”Design of Optimal Aerodynamic Shapes using Stochastic Optimization Methods and Computational Intelligence”, Progress in Aerospace Sciences, 38, 43-76, 2002. [10] Karakasis, M., Giannakoglou, K., ”On the Use of Surrogate Evaluation Models in MultiObjective Evolutionary Algorithms”, ECCOMAS 2004, Jyvaskyla, 2004. [11] Papadimitriou, D., Kampolis, I., Giannakoglou, K., ”Evolutionary Optimization using a New Radial Basis Function Network and the Adjoint Formulation”, Inverse Problems, Design and Optimization (IPDO) Symposium, Rio de Janeiro. 2004. [12] Papadimitriou, D., Kampolis, I., Giannakoglou, K., ”Stochastic and Deterministic Optimization for Turbomachinery Applications based on the Adjoint Formulation”, ERCOFTAC Design Optimization Int. Conf., Athens, Greece, 2004”. 8
© Copyright 2026 Paperzz