Multiobjective optimization with evolutionary algorithms and

Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Multiobjective optimization
with evolutionary algorithms and metaheuristics
in distributed computing environnement
Sébastien Verel
LISIC
Université du Littoral Côte d’Opale
Equipe OSMOSE
[email protected]
http://www.lisic.univ-littoral.fr/~verel
Séminaire SERMA,
February, 10, 2014
1/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Optimization
Inputs
Search space: Set of all feasible solutions,
X
Objective function: Quality criterium
f : X → IR
Goal
Find the best solution according to the criterium
x ? = argmax f
But, sometime, the set of all best solutions, good approximation of
the best solution, good ’robust’ solution...
2/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Contexte
Black box Scenario
We have only {(x0 , f (x0 )), (x1 , f (x1 )), ...} given by an ”oracle”
No information is either not available or needed on the definition
of objective function
Objective function given by a computation, or a simulation
Objective function can be irregular, non differentiable, non
continous, etc.
(Very) large search space for discrete case (combinatorial
optimization), i.e. NP-complete problems
3/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Contexte
Typical applications
Large combinatorial problems:
Scheduling problems, logistic problems, DOE,
problems from combinatorics (Schur numbers), etc.
Calibration of models:
Physic world ⇒ Model(params) ⇒ Simulator(params)
Model(Params) = argminM Error(Data, M)
Shape optimization:
Design (shape, parameters of design)
using a model and a numerical simulator
4/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Search algorithms
Principle
Enumeration of the search space
A lot of ways to enumerate the search space
Using random sampling: Monte Carlo technics
Local search technics:
5/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Search algorithms
Single solution-based: Hill-climbing technics,
Simulated-annealing, tabu search, Iterative Local Search, etc.
Population solution-based: Genetic algorithm, Genetic
programming, ant colony algorithm, etc.
Design components are well-known
Probability to decrease,
Memory of path, of sub-space
Diversity of population, etc.
6/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Position of the work
One Evolutionary Algorithm key point:
Exploitation / Exploration tradeoff
One main practicle difficulty:
Choose operators, design components, value of parameters,
representation of solutions
Parameters setting (Lobo et al. 2007):
Off-line before the run: parameter tuning,
On-line during the run: parameter control.
7/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Position of the work
One Evolutionary Algorithm key point:
Exploitation / Exploration tradeoff
One main practicle difficulty:
Choose operators, design components, value of parameters,
representation of solutions
Parameters setting (Lobo et al. 2007):
Off-line before the run: parameter tuning,
On-line during the run: parameter control.
Main question
How to combine correctly the design components
according to the problem ?
7/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Fitness landscapes
Definition
Fitness landscape (X , N , f )
X is the search space
N : X → 2X is a
neighborhood relation
f : X → IR is a
objective function
8/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Multimodal Fitness landscapes
Local optima x ∗
no neighbor solution with higher fitness value
∀x ∈ N (x ∗ ), f (x) ≤ f (x ∗ )
Fitness
Search space
9/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Local Optima Network (LON)
Join work
Gabriela Ochoa (univ. Stirling),
Marco Tomassini (univ. Lausanne),
Fabio Daolio (univ. Lausanne)
10/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Energy surface and inherent networks (Doye, 2002)
a Model of 2D energy surface
b Contour plot, partition of
the configuration space into
basins of attraction
surrounding minima
c landscape as a network
Inherent network:
Nodes: energy minima
Edges: two nodes are connected if the energy barrier
separating them is sufficiently low (transition state)
11/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Local optima network
[GECCO’08 - IEEE TEC’10]
0.76
0.65
0.185
fit=0.7657
fit=0.7133
0.29
0.4
Nodes :
local optima
Edges :
transition probabilities
0.27
0.05
0.055
fit=0.7046
0.33
12/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Structure of Local Optima Network
Quadratic Assignment Problem [CEC’10 - Phys A’11]
Random instance
Real-like instance
Structure of the Local Optima Network related to search difficulty
13/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Network Metrics
nv : ]vertices
lo: path to best
lv : avg path
fnn: fitness corr.
wii: self loops
cc: clust. coef.
zout: out degree
y 2: disparity
knn: degree corr.
14/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
ILS Performance vs LON Metrics
[GECCO’12]
1
Multiple linear regression on all possible predictors:
perf = β0 + β1 k + β2 log(nv ) + β2 lo + · · · + β10 knn + 2
Step-wise backward elimination of each predictor in turn.
Table : Summary statistics of the final linear regression model. Multiple
R-squared: 0.8494, Adjusted R-squared: 0.8471.
Predictor
(Intercept)
lo
zout
y2
knn
βbi
10.3838
0.0439
−0.0306
−7.2831
−0.7457
Std. Error
0.58512
0.00434
0.00831
1.63038
0.40501
p-value
9.24 · 10−47
1.67 · 10−20
2.81 · 10−04
1.18 · 10−05
6.67 · 10−02
15/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Tuning and Local Optima Network
Relevant features using ”complex system” technics
Search difficulty is related to the features of LON:
Classification of instances
Understanding the search difficulty
Guideline of design
Time complexity can be predicted:
Tuning with Portefolio technics
16/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Multiobjective optimization
Multiobjective optimization problem
X : set of feasible solutions in the decision space
M > 2 objective functions f = (f1 , f2 , . . . , fM ) (to maximize)
Z = f (X ) ⊆ IRM : set of feasible outcome vectors in the
objective space
x2
f2
Decision space
x1
Objective space f1
17/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Definition
Pareto dominance relation
A solution x ∈ X dominates a solution x 0 ∈ X (x 0 ≺ x) iff
∀i ∈ {1, 2, . . . , M}, fi (x 0 ) 6 fi (x)
∃j ∈ {1, 2, . . . , M} such that fj (x 0 ) < fj (x)
x2
f2
nondominated
vector
dominated
nondominated
solution
Decision space
vector
x1
Objective space f1
18/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Pareto set, Pareto front
x2
f2
Pareto front
Pareto
optimal
solution
Pareto optimal
set
Decision space
x1
Objective space f1
Goal
Find the Pareto Optimal Set,
or a good approximation of the Pareto Optimal Set
19/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Why multiobjective optimization ?
Position of multiobjective optimization
No a priori on the importance of the different objectives
a posterioro selection by a decision marker:
Selection of one Pareto optimal solution after a deep study of
the possible solutions.
20/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Motivations
Join work
Arnaud Liefooghe, Clarisse Dhaenens, Laetitia Jourdan, Jérémie Humeau,
DOLPHIN Team, INRIA Lille - Nord Europe
Goal
Understand the dynamics of search algorithms and
evaluate the time complexity
Guideline to design efficient local search algorithms
Portefolio: between of set of algorithms, select the best one
21/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Local search algorithms
Two main classes:
Scalar approaches: multiple scalarized aggregations of the
objective functions, only able to find a subset of Pareto
solutions (supported solutions)
Pareto-based approaches: directly or indirectly focus the
search on the Pareto dominance relation
f2
supported
solution
f2
Accept
No
accept
Objective space f1
Objective space f1
22/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Scalar approaches
• multiple scalarized aggregations of the objective functions
f2
supported
solution
Different aggregations
Weighted sum:
X
f (x) =
λi fi (x)
i=1..m
Weighted Tchebycheff:
f (x) = max {λi (ri∗ − fi (x))}
Objective space f1
i=1..m
...
23/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
A Pareto-based approach: Pareto Local Search
• Archive solutions using Dominance relation
• Iteratively improve this archive by exploring the neighborhood
f2
f2
Accept neighbor
No Accept
current
archive
f2
current
archive
Objective space f1
f2
Objective space f1
Accept
current
archive
Objective space f1
current
archive
Objective space f1
24/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
What makes a multiobjective optimization problem “difficult”?
The search space size?
The “difficulty” of underlying single-objective problems?
The number of objectives?
The objective correlation?
The number of non-dominated solutions?
The shape of the Pareto front?
...
. . . what are the relevant features?
25/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Summary of problem features
k
m
ρ
npo
hv
avgd
maxd
nconnec
lconnec
kconnec
nplo
ladapt
corhv
corlhv
low-level features
number of epistatic interactions
number of objective functions
objective correlation
high-level features
number of Pareto optimal solutions
hypervolume
average distance between Pareto sol.
maximum distance between Pareto sol.
number of connected components
size of the largest connected component
minimum distance to be connected
number of Pareto local optima
length of a Pareto-based adaptive walk
correlation length of solution hypervolume
correlation length of local hypervolume
unknown (black-box)
known
known (estimated?)
enumeration (bounds)
enumeration
enumeration
enumeration
enumeration
enumeration
enumeration
enumeration (estim.)
estimation
estimation
estimation
. . . correlation analysis?
26/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
0.8
0.8
0.7
0.7
0.7
0.6
0.6
0.6
0.5
Objective 2
0.8
Objective 2
Objective 2
MNK-landscapes with tunable objective correlation
0.5
0.5
0.4
0.4
0.4
0.3
0.3
0.3
0.2
0.2
0.3
0.4
0.5
0.6
Objective 1
0.7
0.8
Conflicting objectives
ρ = −0.9
• Large Pareto set
• Few Pareto sol.
0.2
0.2
0.3
0.4
0.5
0.6
Objective 1
0.7
0.8
Independent objectives
ρ = 0.0
0.2
0.2
0.3
0.4
0.5
0.6
Objective 1
0.7
0.8
Correlated objectives
ρ = 0.9
• Small Pareto set
• All supported sol.
27/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Summary of results and guidelines for local search design
Landscape features
Problem properties
N
K
M
ρ
Cardinality of the
Pareto optimal set
Proportion of
supported solutions
Connectedness of the
Pareto optimal set
Number of
Pareto local optima
Suggestion for the
design of local search
+
−
+
−
limited-size archive
unbounded archive
scalar efficient
scalar not efficient
ρ≥0
+ two-phase efficient
ρ<0
− two-phase not efficient
+ Pareto not efficient
− Pareto efficient
Following project...
28/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Join work
K. Tanaka, H. Aguirre,
Shinshu university,
Nagano, Japon
A. Liefooghe,
Inria Lille - Nord Europe
29/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Performance vs. problem features
30/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Performance vs. problem features
0 m, ρ, npo, hv
∼ connectedness (nconnec, lconnec, kconnec)
+ distance between Pareto solutions (avgd, maxd)
+ number of local optima (nplo, ladapt)
++ ruggedness (k, corhv, corlhv)
30/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Performance prediction
Predicting the performance of EMO algorithm?
Multiple linear regression model(s)
log(ert) = β0 + β1 · k + β2 · m + β3 · ρ + β4 · npo + . . . + β14 · corlhv + e
Towards a portfolio selection approach (in progress)
31/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Distributed Localized Bi-objective Search
Join work
Bilel Derbel, Arnaud Liefooghe, Jérémie Humeau,
Université Lille 1, INRIA Lille Nord Europe
32/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Motivations
Context
High computational cost
⇒ Subdivide the problem into subtasks which can be
processed in parallel
Most approaches: top-down, centralized need global
information to adapt its design components
⇒ Distributed computing, local information is often sufficient.
Crossroads approach
Parallel cooperative computation
Decomposition-based algorithms
Adaptive search
33/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Principles
Between fully independent and fully centralized approach
One node of computation = One solution
Each node optimize a localized fitness function
weighted sum, Tchebichev, etc. (based on search direction)
Communicate the state of the search
position to neighboring nodes
f2
f2
Objective space f1
Objective space f1
34/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Expected dynamic
35/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Observed dynamic
0.7
0.7
0.65
0.65
t=112
f2
0.8
0.75
f2
0.8
0.75
0.6
0.6
0.55
0.55
0.5
0.5
t=16
t=32
t=0
0.45
0.45 0.5 0.55 0.6 0.65 0.7 0.75 0.8
f1
0.45
0.45 0.5 0.55 0.6 0.65 0.7 0.75 0.8
f1
Tested on benchmark
Effective in terms of:
approximation quality,
computational time,
scalability
36/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Position of the work
One EA key point:
Exploitation / Exploration tradeoff
One main practicle difficulty:
Choose operators, design components, value of parameters,
representation of solutions
Parameters setting (Lobo et al. 2007):
Off-line before the run: parameter tuning,
On-line during the run: parameter control.
Increasing number of computation ressources (GPU, ...)
but, add parameters
Main question
Control the execution of different metaheuristics
in a distributed environment
37/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Distributed Adaptive Metaheuristic Selection (DAMS)
[GECCO, 2011]
Join work
Bilel Derbel,
Université Lille 1, INRIA Lille Nord Europe
38/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Execution of metaheuristics in a distributed environment
3 metaheuristics {M1 , M2 , M3 }
4 nodes of computation
Which algorithm can we design?
M1
M2
M3
39/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Execution of metaheuristics in a distributed environment
3 metaheuristics {M1 , M2 , M3 }
4 nodes of computation
Which algorithm can we design?
M1
M3
M3
M3
M3
M3
M3
M3
M3
M3
M3
M3
M3
M2
M3
39/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Execution of metaheuristics in a distributed environment
3 metaheuristics {M1 , M2 , M3 }
4 nodes of computation
Which algorithm can we design?
M1
M3
M3
M3
M3
M3
M3
M3
M2
M3
M2
M3
M2
M2
M3
39/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Execution of metaheuristics in a distributed environment
3 metaheuristics {M1 , M2 , M3 }
4 nodes of computation
M1
M2
M3
40/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Execution of metaheuristics in a distributed environment
3 metaheuristics {M1 , M2 , M3 }
4 nodes of computation
M1
M3
M3
M1
M2
M1
M1
M3
M3
M1
M2
M1
M1
M2
M3
Goal of control: Optimal strategy
The strategy that gives the best overall performances:
At first, execution of M3
then, M1 and M2 , and then, M1 .
40/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Control: Local strategy
mn attachements with n nodes and m metaheuristics
8
5
3
4
4
2
Local distributed strategy:
For each computational node:
Measure the efficiency of neighboring metaheuristics
Select a metaheuristic according to the local information
Trade-off between:
Exploitation of the most promizing metaheuristics
Exploration of news ones
41/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Select Best and Mutate strategy (SBM)
For each nodes:
k ← initMeta()
P ← initPop()
reward ← 0
while Not stop do
Migration of populations
Send (k, reward)
Receive ∀i (ki , rewardi )
k ← best meta from ki ’s
if rnd(0, 1) < pmut then
k ← random meta
end if
Pnew ← metak (P)
reward ← Reward(P, Pnew )
P ← Pnew
end while
42/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Select Best and Mutate strategy (SBM)
For each nodes:
8
5
3
4
2
4
Select ’best’ one (rate 1 − pmut )
8
5
3
4
2
4
Select random one (rate pmut )
8
5
3
k ← initMeta()
P ← initPop()
reward ← 0
while Not stop do
Migration of populations
Send (k, reward)
Receive ∀i (ki , rewardi )
4
4
2
k ← best meta from ki ’s
if rnd(0, 1) < pmut then
k ← random meta
end if
Pnew ← metak (P)
reward ← Reward(P, Pnew )
P ← Pnew
end while
42/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Dynamics of SBM-DAMS
Average frequency and fitness
1
fitness
5 bits
3 bits
1 bit
bit-flip
Frequency
0.8
0.6
0.4
0.2
0
0
500
1000
1500
Rounds
2000
2500
3000
First, 1/l bit-flip mutation (opposite of sequential oracle)
Then 5-bits, 3-bits, 1-bit, 1/l bit-flip
See sbmDams.mov
43/44
Contexte
Fitness landscapes
Multiobjective Optimization
Distributed MO
DAMS
Conclusion
One EA key point: Exploitation / Exploration tradeoff
One main practicle difficulty:
Choose operators, design components, value of parameters
Fitness Landscapes:
Features, description of the search space
Off-line tuning:
Model to predict the performances,
Design guideline, portefolio
On-line control:
Machine learning technics
⇒ Toward (more) automatic design
44/44