Hyper-volume evolutionary algorithm
VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
Hyper-volume Evolutionary Algorithm
Khoi Nguyen Le1,∗, Dario Landa-Silva2
1VNU University of Engineering and Technology, Hanoi, Vietnam
2The University of Nottingham, Nottingham, United Kingdom
Abstract
We propose a multi-objective evolutionary algorithm (MOEA), named the Hyper-volume Evolutionary
Algorithm (HVEA). The algorithm is characterised by three components. First, individual fitness evaluation
depends on the current Pareto front, specifically on the ratio of its dominated hyper-volume to the current Pareto
front hyper-volume, hence giving an indication of how close the individual is to the current Pareto front. Second,
a ranking strategy classifies individuals based on their fitness instead of Pareto dominance, individuals within
the same rank are non guaranteed to be mutually non-dominated. Third, a crowding assignment mechanism that
adapts according to the individual’s neighbouring area, controlled by the neighbouring area radius parameter,
and the archive of non-dominated solutions. We perform extensive experiments on the multiple 0/1 knapsack
problem using different greedy repair methods to compare the performance of HVEA to other MOEAs including
NSGA2, SEAMO2, SPEA2, IBEA and MOEA/D. This paper shows that by tuning the neighbouring area
radius parameter, the performance of the proposed HVEA can be pushed towards better convergence, diversity or
coverage and this could be beneficial to different types of problems.
Received 05 December 2015, revised 20 December 2015, accepted 31 December 2015
Keywords: Multi-objective Evolutionary Algorithm, Pareto Optimisation, Hyper-volume, Knapsack Problem.
1. Introduction
An important issue in a MOEA is how to
establish superiority between solutions within
the population, i.e. how to assess solution
fitness in a multi-objective sense. In this paper,
we proposed the Hyper-Volume Evolutionary
Algorithm (HVEA), a MOEA that employs
the concepts of volume dominance proposed
by Le et at [1] to assess solution fitness in
multi-objective optimisation.
The development of heuristic and evolutionary
techniques to solve real-world multi-objective
optimisation problems is an an active
research area. In Pareto-based multi-objective
optimisation, a set of non-dominated solutions,
also known as Pareto front, is sought so
that the decision-maker can select the most
appropriate one.
Evolutionary xalgorithms
The paper is organised as follows. Section 2
and other population based heuristics seem
well suited to deal with Pareto based multi-
objective optimisation problems because they
can evolve a population of solutions towards the
Pareto-optimal front in a single run. A good
multi-objective evolutionary algorithm (MOEA)
should be able to obtain Pareto fronts that
are both well-distributed and well-converged.
describes the HVEA algorithm. In Section 3,
most well-known and recent MOEAs in the
literature together with the multiple 0/1 knapsack
problem are discussed.
In Section 4, the
experimental results and discussion are presented.
Then some conclusions and perspectives are
discussed in Section 5.
∗
Corresponding author. Email: khoi.n.le@vnu.edu.vn
10
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
11
2. Hyper-volume Evolutionary Algorithm
Step 1: Initialisation: Start with an empty
population P and create an initial archive P
(each of size N).
We propose a new approach to multi-objective
optimisation, the Hyper-volume Evolutionary
Algorithm (HVEA). HVEA deploys techniques
that are well established in the literature as well
as presents new ones in order to find the Pareto
front of the problem. As other population-based
MOEAs, the proposed HVEA
Step 2:
Fitness assignment:
Calculate
S
fitness values of individuals in P
(cf. Section 2.1).
P
Step 3: Ranking assignment: Assign ranking to
S
individuals in P P (cf. Section 2.2).
• Uses a population of current solutions and an
archive for storing the best solutions found
so far,
Step 4: Environmental selection: Repeatedly
copy all individuals having the best ranking
from P to P and assign their crowding values
until the size of P exceeds N then reduce the
size of P by means of the truncation operator
(cf. Section 2.3 and Section 2.4).
• Assigns scalar fitness values to individuals
and uses the Pareto dominance relationship,
• Employs a crowding strategy to maintain the
diversity of the population and, if necessary,
control the size of the archive during
the search.
Step 5: Termination: Stopping criteria are
satisfied then present all non-dominated
individuals in P as solutions to the problem,
and then the algorithm stops here.
However, HVEA is distinguished by
four features:
Step 6: Generating offspring: Apply crossover
and mutation operators on parents, which
are selected with binary tournaments, to
produce offspring (cf. Section 2.5). Go to
Step 2.
• The fitness of an individual is calculated
using the hyper-volume of that individual
and the hyper-volume of the current
representative Pareto front.
2.1. Fitness Assignment
• Individuals are ranked based on their
fitness values, individuals having fractional
difference in fitness values between them are
classified into the same rank.
The fitness assignment procedure of HVEA
deploys the concept of volume dominance
proposed by Le et at [1]. Le et at. proposed a
new relaxed Pareto dominance for multi-objective
optimisation, named volume dominance. They
assign strength values to individuals in the
population using the ratio between the hyper-
volume of an individual and the hyper-volume
of the current representative Pareto front. The
dominance relationship between two individuals
is established by comparing strengths of the pair
with respect to the threshold strength rS tr. The
fitness assignment of HVEA is as follows. Let
V(x) be the hyper-volume of individual x w.r.t.
• A niching technique to preserve the diversity
of the population based on distance between
the individual and other solutions in its
neighbouring area.
• Offspring that improve the current Pareto
front (in other words offspring that dominate
individuals in the current Pareto set) are
guaranteed to be selected not only for the
archive but also for the tournament selection
to fill the mating pool.
~
x
the reference point r ; ParetoFront is the current
Pareto front of the archived population P; the
current representative Pareto front, RP(x), is the
Overall, HVEA works as follows:
12
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
set of all individuals in ParetoFront that Pareto-
dominate x.
representative
Pareto front
V ref (x)
f2
r sup
x ref
ꢀ
ꢁ
^
*
*
*
x
RP(x) = x | x ꢀ x
x ∈ ParetoFront (1)
current
Pareto front
ꢀ
ꢁ
xiref = sup fi(x ) | x ∈ RP(x)
(2)
*
*
rinf
r x
ref
ref (x) is the hyper-volume of x , given
V(x)
V
~
x
O
f1
by (2), w.r.t. the reference point r , given by (4),
then the fitness value of individual x is as follows:
Fig. 1: The fitness assignment in HVEA.
V(x)
Vref (x)
fitness(x) = 1 −
(3)
Step 1: Use all individuals in P and P to update:
ꢂ
ꢃ
As RP(x) is an empty set if x is a Pareto
rinf = r1inf, r2inf, . . . , rminf
(9)
~
solution, (3) is relaxed by setting fitness(x) = 0.
~
x
The estimation of the reference point r and
ꢂ
ꢃ
sup
the hyper-volume V(x), Vref (x) is identical to the
~
r
= r1sup, r2sup, . . . , rmsup
(10)
one proposed by Le et al. [1]. The reference point
ꢂ
ꢃ
x x
x
where riinf and risup are given by (5) and (6)
respectively.
~
x
r = r1 , r2 , . . . , rm is as follows:
sup
ri = fi(x) − (r − riinf)
(4)
x
i
Step 2: Update ParetoFront using all newly
generated offspring in P. Assign fitness
value of −1 to offspring that Pareto-
dominate at least one individual in
ParetoFront of the previous generation.
Assign fitness value of 0 to the remaining
individuals in ParetoFront of the current
generation. Apply (3) to assign fitness value
to all remaining individuals in P and P.
for ∀i = 1, 2, . . . , m
ꢀ
ꢁ
[
riinf = inf fi(x ) | x ∈ P
P
(5)
(6)
*
*
ꢀ
ꢁ
[
risup = sup fi(x ) | x ∈ P
P
*
*
and the hyper-volume V(x), Vref (x):
m
Y
Assigning fitness value of −1 to offspring
that Pareto-dominate at least one individual
in ParetoFront enables HVEA to distinguish
individuals that make a direct contribution in
moving the Pareto front forward during the
environmental selection and mating selection.
This fitness assignment strategy emphasises the
idea of preferring individuals near the Pareto front
especially individuals that improve the Pareto
front of the previous generation.
x
V(x) =
(fi(x) − ri )
(7)
(8)
i=1
m
Y
(xiref − ri )
x
V
ref (x) =
i=1
Fig. 1 illustrates the estimation of the above
terms that are required for the calculation of
individual fitness in HVEA. Lower values of
fitness(x) are preferred.
The procedure to
2.2. Ranking Assignment
determine the fitness values of individuals in
the offspring population P and the archive P is
as follows:
As commonly found in the literature, the
ranking assignment in evolutionary algorithms
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
13
l−1
X
classifies individuals with similar (or identical)
characteristic(s) into the same category(ies).
The characteristic usually used is the Pareto
|Fi| + |Fl| > N
(14)
i=−1
Individuals in Fl are sorted by their crowding
values. The individual with the highest crowding
value in Fl is removed from P and Fl. The
crowding value of remaining individuals in P, not
just Fl, are updated accordingly. These processes
are repeated until the size of P equals to N.
non-dominance. In other words, Pareto non-
dominated individuals are classified into one
category (rank). Therefore it is guaranteed that
individuals with the same rank are all mutually
non-dominated. An example is NSGA2 [2] with
its fast non-dominated sorting approach. HVEA
takes a paradigm shift in ranking individuals. In
HVEA, it is the case that an individual is Pareto
dominated by another individual having the same
rank. The main intention of this mechanism
is to allow slightly worse quality individuals to
be able to compete for survival and/or selection.
HVEA computes ranking from the individual
fitness value as follows:
2.4. Crowding Assignment
As NSGA2 and SPEA2, HVEA employs
the distance-based approach to estimate the
density of individuals around a particular
individual.
However HVEA’s crowding
assignment mechanism is different from NSGA2
and SPEA2. Both NSGA2 and SPEA2 only
consider “one neighbour” in the neighbouring
area in order to determine the density of a
particular individual. NSGA2 combines the
distance to the adjacent neighbour in each
objective to estimate the crowding of a particular
individual whereas SPEA2 considers the distance
to the k-th nearest neighbour. HVEA takes
into account all individuals in the neighbouring
area to determine the crowding of a particular
individual. The neighbouring area is defined
by the current state of the combined population
$
%
1
µ
rank(x) = fitness(x) ×
(11)
where parameter µ indicates the ‘size’ of each
1
ranking level whereas
could be referred as
the desired number of µranks (desired number
of ‘fronts’). The advantage of this ranking
assignment mechanism is that only Pareto non-
dominated individuals are assigned rank 0 with
exception of individuals, which improve the
Pareto front, are assigned rank −1. Other Pareto
dominated individuals are assigned rank with
some noise induction by applying (11).
S
*
P
P and a radius ω. An individual x is in the
neighbouring area of an individual x (in other
word x and x are neighbours) if the following
*
2.3. Environmental Selection
condition is satisfied:
S
Let Fi is the set of individuals in P P with
ꢂ
ꢃ
ꢄ
ꢄ
sup
i
ꢄ
∗ꢄ
xi − xi ≤ r − riinf × ω
(15)
ꢄ
ꢄ
rank(x) = i:
[
n
o
for ∀i = 1, 2, . . . , m and riinf and risup are given
Fi = x | rank(x) = i ∧ x ∈ P
P
(12)
in (5) and (6). Let NB(x) is the set of all
One set at a time, all individuals from
each set Fi, with i = −1, 0, 1, 2, . . ., are copied
to the archive P until the size of P equals to
or exceeds N. If the size of P equals to N,
the environmental selection is completed.
Otherwise, individuals in the last front Fl copied
to P, are removed from P until reducing the size
of P to N. The last copied front Fl satisfies the
following conditions:
*
neighbouring individuals of x that means x
NB(x) if and only if x and x are satisfied (15).
Then, the crowding value of x is as follows:
∈
*
X
*
crowding(x) =
(d(x, x ) + 1)−1(16)
*
x ∈NB(x)
*
where d(x, x ) is the Euclidean distance between
*
x and x :
v
t
m
l−1
X
X
*
d(x, x ) =
(xi − xi∗)2
(17)
|Fi| < N
(13)
i=1
i=−1
14
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
and lower value of crowding(x) is preferred.
Apart from considering all individuals in the
neighbouring area to estimate the crowding of a
particular individual, another main difference in
crowding assignment mechanism between HVEA
and NSGA2, SPEA2 is that NSGA2 only uses
each front separately to estimate crowding value
of individuals in that front, SPEA2 uses all
individuals in both the offspring population P
and the archive P to estimate crowding value of
individuals, but HVEA only uses individuals in
fronts, which are already added to the archive P,
to estimate the crowding value. Additionally, the
crowding value of individual x ∈ P is repeatedly
adjusted when a new front Fi is added to P which
is not the case in SPEA2. The pseudocode for the
crowding assignment of front Fi (−1 ≤ i ≤ l) is
as follows:
3. Comparative Case Study
3.1. Multi-objective Evolutionary Algorithms
The performance of HVEA is compared to
NSGA2 [2], SEAMO2 [3], SPEA2 [4], IBEA [5]
and MOEA/D [6].
3.1.1. Non-dominated Sorting Genetic Algorithm 2
(NSGA2)
Deb et at. [2] used a fixed population and
a fixed archive of size N for their NSGA2.
They proposed a fast non-dominated sorting
approach in NGSA2 to classify individuals in a
population into different non-domination levels
or different fronts. NSGA2 employed a density
estimation metric to preserve the population
diversity. It required to sort each front according
to each objective function value in ascending
order of magnitude. The boundary individuals are
assigned an infinite distance value. The density
of individuals surrounding other intermediate
individual in the front is the sum of the average
normalised distance of two individuals on either
side of this individuals along each of the
objectives. The mating pool is filled by the
binary tournament selection with the following
priority, non-domination level then crowding
distance. The offspring population is formed by
applying crossover and mutation on the mating
pool. The best fronts of the offspring population
and the archive combined are selected for the
next archive. If the size of the archive is greater
than N, individuals in the last front is removed
to reduce the size of the archive to N based on
their crowding value. There is no treatment to
prevent duplication of individuals whilst forming
the archive. The duplication individuals could
be spotted by the crowding value and removed
during the truncation of the archive but it is
not exhaustive.
Procedure CrowdingAssignment(Fi)
begin
for each x in Fi
*
*
for each x in Fi (x , x)
*
if x and x are neighbours
then add (d(x, x ) + 1)−1 to crowding(x) and crowding(x )
*
*
endif
endfor
for each front Fj (−1 ≤ j < i)
*
for each x in Fj
*
if x and x are neighbours
then add (d(x, x ) + 1)−1 to crowding(x) and crowding(x )
*
*
endif
endfor
endfor
endfor
end
2.5. Generating Offspring
HVEA chooses parents to fill the mating pool
by binary tournament selection. Individuals
competed for the second parent are different
from the first parent. The binary tournament
selection prioritises the following order: ranking,
crowding, fitness values of individuals. Crossover
and mutation are applied on the mating pool
to form the offspring population P.
Any
3.1.2. Simple Evolutionary Algorithm for Multi-
objective Optimization 2 (SEAMO2)
SEAMO2 [3] uses a steady-state population
and a simple elitist replacement strategy. Each
individual of the population act as the first parent
once and a second parent is chosen at random.
duplicated offspring dies out of the combined
S
population P P. This guarantees that there
is no duplication in the combined population
S
P
P
before the environmental selection
process. Consequently, there is no duplication in
the archive P.
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
15
S
Crossover and mutation is applied on the pair
of parents to produce offspring. The offspring
replaces one of the parents if either its objective
vector improves on any best-so-far objective
function or it dominates that parent. If neither the
offspring dominates the parents nor the parents
dominate the offspring, the offspring replaces
a random individual in the population that the
offspring dominates. SEAMO2 does not allow
non-dominated individuals in P P, are copied
to the next archive. Zitzler et at. pointed out
that there are two situation: either the archive
is too small or too large. If the archive is
too small, the best dominated individuals, based
on the fitness value, are copied to fill P. In
the later situation, non-dominated individuals
in the archive are iteratively removed until the
archive’s size is not exceeded. The removal of
non-dominated individuals from the archive is
carefully managed by using an archive truncation
method that guarantees the preservation of
boundary solutions. As in NSGA2, SPEA2 relies
on the archive truncation to remove duplicated
individuals but it does not guarantee that the
archive contains no duplication.
duplication in its population.
Therefore,
any duplicated offspring dies before the
replacement process.
3.1.3. Strength Pareto Evolutionary Algorithm 2
(SPEA2)
Zitzler et at. [4] employed a fixed size archive
to store non-dominated individuals in addition
to a population for SPEA2. SPEA2 uses a
fine-grained fitness assignment strategy which
takes for each individual into account how many
individuals it dominates and it is dominated
by. Each individual in both the archive and
the population is assigned a strength value
S (i), indicating the number of individuals it
dominated. The raw fitness R(i) of an individual
is defined as the sum of the strength value of all
individuals by which it is dominated.
3.1.4. Indicator Based Evolutionary Algorithm
(IBEA)
Zitzler and Ku¨nzli [5] proposed a general
framework
indicator-based
evolutionary
algorithm (IBEA). IBEA could be referred
as classic MOEAs but guided by a general
preference information of the dominance
relationship, a binary quality indicator. IBEA
needs neither specifications of weights or targets
(as in aggregation methods) nor the dominance
relationship and distribution techniques (as
in classic MOEAs). IBEA uses this binary
indicator guiding the search to optimise the
approximation set. The binary quality indicator
in IBEA maps an ordered pair of individuals to
a real number which therefore could be used
for fitness calculation. Zitzler and Ku¨nzli [5]
proposed two indicators, the additive ꢀ-indicator
ꢄ
ꢄ
[
n
o
ꢄ
ꢄ
ꢄ
ꢄ
ꢄ
S (i) = j | j ∈ P
P ∧ i ꢁ j
(18)
(19)
ꢄ
X
R(i) =
S (j)
j∈PS P∧ jꢁi
SPEA2 uses an adaptation of the k-th nearest
neighbour method to estimate the density of an
individuals. The density of an individual is
estimated as the inverse of the distance to the
k-th nearest neighbour. Then, the density is
added to the raw fitness Ri to yields the fitness
of an individual i. Similarly to NSGA2, SPEA2
fills the mating pool using binary tournament
selection on the archive based on the fitness
of individuals participated in the tournament.
Crossover and mutation is applied on the mating
pool to create the offspring population. From
the offspring and the archive population, all
individuals having fitness less than one, i.e. all
+
Iꢀ and the hyper-volume-indicator IHD. For an
1
2
1
2
+
order pair of individuals (x , x ), Iꢀ (x , x ) gives
minimum distance for which x1 is translated
in each in objective space to weakly dominate
x2. IHD(x1, x2) measures the space volume that
is dominated by x2 but not by x1 with respect
to a predefined reference point. The fitness
of an individual is estimated as the sum of an
exponential function of the indicator values.
A positive constant κ = 0.05 is applied to
the indicator value to amplify the influence of
dominating individuals over dominated one. It
16
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
is suggested to use adaptive scaling for both
indicator values and objective values to avoid
the issue of estimating a good reference point.
As aforementioned population based MOEAs
(NSGA2 and SPEA2), IBEA uses a fixed size
archive and an offspring population. Offspring
is produced by recombination and mutation on
a pair of parents selected from the archive. The
archive and the offspring population is combined
and truncated to generate the new archive for the
next iteration.
annealing) and adapting the weight vectors, so
called EMOSA. EMOSA deploys the cooling
technique in simulated annealing to adaptively
modified weight vectors at the lowest temperature
in order to diversify the population towards the
unexplored part of the search space. EMOSA also
uses simulated annealing technique to further
improve each individual in the population.
Finally, EMOSA use ꢀ-dominance to update
the external population in order to maintain
the diversity of the external population. The
performance of EMOSA is then compared
against a set of multi-objective simulated
annealing algorithms and a set of multi-objective
memetic algorithms using multi-objective 0/1
knapsack problem and multi-objective travelling
salesman problem. EMOSA performs better than
all these algorithms and is capable of finding very
good distribution of non-dominated solutions.
3.1.5. Multi-objective Evolutionary Algorithm
Based on Decomposition (MOEA/D)
MOEA/D, proposed by Zhang and Li [6],
decomposes
a
multi-objective optimisation
problem into a number of scalar optimisation sub-
problems and optimises them simultaneously.
Each sub-problem is optimised by utilising
information from it neighbouring sub-problems.
Non-dominated individuals found during
the search are stored to an external archive.
MOEA/D predefines a set of even spread weight
3.1.7. S-Metric Selection EMOA (SMS-EMOA)
SMS-EMOA is a steady state population and
selection scheme, proposed by Beume et al. [8].
SMS-EMOA’s goal is to maximise the hyper-
volume, S-metric value, of the population by
removing an individual which exclusively
n
o
vector λ1, . . . , λN , where N is the number
of sub-problems, or the population size of
MOEA/D. Each sub-problem ith corresponds to
a weight vector λi. The neighbourhood of the ith
sub-problem consists of all sub-problems with
the weight vectors closest to λi which include the
ith sub-problem itself. Then, the population of
MOEA/D consists the best solution xi found so
far for each sub-problem ith. At each iteration,
an offspring to ith sub-problem is produced by
recombination and mutation on parents which
are the current solutions to the neighbohood
of the ith sub-problem. The offspring replaces
current solutions of the ith sub-problem and its
neighbouring sub-problems if the fitness value of
the offspring is better than that of these current
solutions. See [6] for more details of MOEA/D
and the fitness calculation.
contributes the least hyper-volume.
each iteration, an offspring is produced by
recombination and mutation. The offspring
At
replace an individual in the population Pt if it
leads to higher quality of the population with
respect to the S-metric. The resulting population
P
t+1, formed by combining the offspring and
the previous population Pt, is partition into
different non-dominated sets, fronts, using the
fast non-dominated sorting algorithm as describe
in NSGA2 [2]. The first front R1 contains all
non-dominated solutions of Pt+1, the second front
contains all individual that are non-dominated
in the set (Pt+1\R1), etc.. An individual r is
discarded from the last front Rv, the worst ranked
front, for which that individual exclusively
contributes the least hyper-volume to the last
front Rv.
3.1.6. Adaptive Evolutionary Multi-objective
Simulated Annealing (EMOSA) [7]
Li and Landa-Silva [7] proposed an improved
version of MOEA/D by incorporating an
advanced local search technique (simulated
r = arg min [∆ (s, Rv)]
(20)
S
s∈Rv
∆ (s, Rv) = S (Rv) − S (Rv\ {s})
S
(21)
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
17
Beume et al. also investigated several selection
variants of SMS-EMOA and pointed out that
if there is more than one front in Pt+1 (v >
1), the individual with the most number of
dominating points should be discarded. The
equation (20) should be replaced by the following
equation (22):
by a niching criterion applied in the decision
space (i.e. using spread direction), whereas the
other half is filled by non-dominated dominated
solutions from the combined population using
convergence direction. The archive is updated
by scanning a bundle of rays from the estimated
ideal point into the part of objective space
containing the current pareto front. For each
ray scanned, the non-dominated solution in the
combined population, which are closest to the ray,
is selected for the next archive. The experimental
result on continuous problems including ZDT and
DTLZ obtained by DMEA is slightly better than
those by NSGA2 and SPEA2.
r = arg max [d (s, Pt+1)]
(22)
(23)
s∈Rv
d (s, Pt) = |{y ∈ Pt|y ≺ s}|
The main motivation of this selection variant
is to reduce the runtime complexity and to
emphasise on sparsely filled regions of the
solution space. It is expected that dominating
individual will rise in rank to better fronts and fill
those vacancies [8]. The experimental results on
continuous problems including ZDT and DTLZ
indicated that SMS-EMOA outperforms NSGA2
and SPEA2.
It is emphasised that HVEA is clearly
different from SMS-EMOA. SMS-EMOA uses
the exclusive hyper-volume contribution of
individuals in the worst front to eliminate
an individual during the steady-state selection
3.1.9. Direction-based Multi-objective Evolutionary
Algorithm II (DMEA-II)
The DMEA is further improved and so
called DMEA-II [10]. The main differences
between these two algorithms are: (1) an
adaptive approach for choosing directional types
in forming the parental pool, (2) a new concept
for ray-based density for niching and (3) a
new selection scheme based on this ray-based
density. In DMEA, a half of the parental pool
is formed by using the convergence direction
and the other half is formed by using the
spread direction. DMEA-II deployed an adaptive
approach, in which this ratio is depended on the
number of non-dominated solutions in the current
population. The higher the number of non-
dominated solutions in the current population is,
the more use of the spread direction. In other
words, if the current population contains all non-
dominated solutions then the spread direction is
used to generate the whole parental population.
The second improvement is a new ray-based
density approach by counting the number of
ray that a solution is the closest one. Then a
half of the archive is updated by non-dominated
solutions with highest ray-based density, whereas
the other half is formed by solutions closest to
each ray. Continuous problems including ZDT,
DTLZ and UF are used as benchmark problems
scheme.
an individual to determine its fitness and
its ranking. The next archive is selected
HVEA use the hyper-volume of
from the archive population combining with
the offspring population using only individual
ranking and crowding.
3.1.8. Direction-based Multi-objective Evolutionary
Algorithm (DMEA)
Bui et al. [9] proposed a population-based
evolutionary algorithm that evolves a population
along directions of improvement. They are two
types of directions employed in DMEA i.e. (1)
the convergence direction between an archived
non-dominated solution and a dominated solution
from the parental pool and (2) spread direction
between two non-dominated solutions in the
archive. These directions are used to perturb
the parental population at each generation. A
half of the next-generation parental pool is
formed by non-dominated whose spread is aided
for comparison.
Experimental results show
that DMEA-II outperforms DMEA, NSGA2 and
SPEA2 but remains competitive to MOEA/D.
18
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
Table 1: Parameter Setting for The Multiple
3.2. Experimental Design
0/1 Knapsack Problem
3.2.1. Multi-objective Optimisation Problem
The performance of HVEA is assessed by
comparing to NSGA2, SEAMO2, SPEA2, IBEA
and MOEA/D on the multiple 0/1 knapsack
problem proposed by Zitzler and Thiele [11].
The multiple 0/1 knapsack is defined as a set of n
items and a set of m knapsacks, weight and profit
associated with each item in a knapsack, and an
upper bound for the knapsack capacity. The goal
is to find a subset of items that maximise the
profit in each knapsack and the knapsack weight
does not exceed its capacity.
Instance Population Size (S) N in MOEA/D
ks2 250
ks2 500
ks2 750
ks3 250
ks3 500
ks3 750
ks4 250
ks4 500
ks4 750
150
200
250
200
250
300
250
300
350
150
200
250
351
351
351
455
455
455
There are several representations, repair
methods and offspring productions for the
multiple 0/1 knapsack problem. Our preliminary
experiment show that the performance of MOEAs
on the multiple 0/1 knapsack problem is affected
by those. Therefore we outline 4 different repair
methods and corresponding representations.
Zitzler and Thiele [11] represented solutions
to the multiple 0/1 knapsack problem as binary
strings. Offspring is produced by applying one
point crossover followed by bit-flip mutation
on a pair of parents. The knapsack capacity
constraints are confronted by a greedy repair
method. This repair method repeatedly removed
items from the encoded solutions until all
capacity constraints are satisfied. Items are
deleted based on the order determined by the
maximum profit/weight ration per item (qj); for
item j the maximum profit/weight ratio (qj) is
given by the equation (27):
pi, j = profit of item j in knapsack i
wi, j = weight of item j in knapsack i
ci = capacity of knapsack i
n
find a vector x = (x1, x2, . . . , xn) ∈ {0, 1} such
that:
n
X
∀i ∈ {1, 2, . . . , m} :
wi, j × xj ≤ ci
(24)
j=1
and maximise f(x) = (f1(x), f1(x), . . . , fm(x)),
where
n
X
fi(x) =
pi, j × xj
(25)
j=1
and xj = 1 if and only if item j is selected.
There are 9 instances for the multiple 0/1
knapsack problem [11]. The population size for
each instance is as follows:
(
)
pi, j
wi, j
m
qj = max
i=1
(27)
We perform 50 independent runs for statistical
analysis, each run is of 2000 generations. It
is noticed that MOEA/D predefines the set of
Items with lowest qj are removed first until
the capacity constraints are fulfilled. This
mechanism diminishes the overall profit as little
as possible [11].
Mumford [12] used a different set of
representation and repair method for the
multiple 0/1 knapsack problem. The solution to
the knapsack problem in SEAMO2 is represented
as a simple permutation of items to be packed.
Offspring is produced from a pair of parents by
n
o
even spread weight vector λ1, . . . , λN where
m−1
N = C
(H: controlling parameter). It is
H+m−1
difficult to change N to the required population
size. Therefore it is suggested to respect the
population size set by MOEA/D but alter the
number of generations for MOEA/D as follows
ꢅ
ꢆ
S
N
g =
× 2000
(26)
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
19
ꢇ
ꢈ
ꢄ
ꢄ
ꢄ
fte x λ,~z = max {λi × |zi − fi(x)|}
(31)
applying cycle crossover followed by random
mutation swapping two arbitrarily selected
items within a single permutation list. The
repair method starts from the beginning of the
permutation list, packing item once at a time
until the weight for any knapsack exceeds its
capacity. Packing is discontinued and the final
item is removed from all knapsacks.
Jaszkiewicz [13] used binary strings to
represent solutions of the multiple knapsack
problem. As Zitzler and Thiele [11], Jaszkiewicz
used one point crossover followed by bit-flip
mutation on a pair of parents to produce offspring.
However Jaszkiewicz used the weighted linear
scalarising approach to repair infeasible solutions
rather than the maximum profit/weight ratio
(qj) [11]. Items are sorted according to the
following ratio:
~
ꢄ
1≤i≤m
where ~z = (z1, z2, . . . , zm) is the reference point
with respect to the current population, zi
max { fi (x) |x ∈ P}.
=
As aforementioned, our initial investigation
shows that there is difference in performance
regarding the greedy repair methods and their
corresponding representations and recombination
methods for the multiple 0/1 knapsack problem.
Therefore we assess the performance of HVEA,
NSGA2, SEAMO2, SPEA2, IBEA and MOEA/D
using different methods separately.
3.2.2. Performance Metrics
We use the hyper-volume measurement
proposed by Zitzler and Thiele [11], the
generational distance and the inverted
generational
evaluation.
Regarding the hyper-volume metric, Zitzler
and Thiele suggested that the reference point
to estimate the hyper-volume at the origin in
the objective space. However, this suggestion
gives extreme points in the objective space much
higher exclusive contribution to the hyper-volume
metric especially when the objective values of the
solutions are high. Therefore, we propose that
the reference point ~r = (r1, r2, . . . , rm) should
be close to the solution set S obtained by all
algorithm. The estimation of the reference point
is as follows:
P
n
distance
measurements
for
i=1 λi × pi, j
lj =
P
(28)
n
i=1 wi, j
~
where λ
=
(λ1, λ2, . . . , λn) is the weight
vector used in the current iteration. Later,
Jaszkiewicz improved this greedy repair method
which is subsequently deployed in MOEA/D.
Jaszkiewicz’s improved greedy repair method is
as follows.
n
o
Let set J
a subset of selected items and set I
i|1 ≤ i ≤ m ∧ j=1 wi, j × xj > ci is a set of
=
j|1 ≤ j ≤ n ∧ xj = 1 is
=
n
o
P
n
overfilled knapsacks. Repeatedly select item k ∈
J to remove until none knapsack is overfilled,
such that:
ui = max {fi(x)}
x∈S
(32)
(33)
(34)
f(xj−) − f(x)
k = arg min
P
(29)
j∈J
i∈I wi, j
li = min { fi(x)}
x∈S
where xj− is different from x only by item j,
ri = li − (ui − li) × 0.1
xij− = xi for all i , j and xj− = 0, and
j
f(x) is the fitness value of x. We investigated
two approaches, weighted sum approach and
Tchebycheff approach, determining the fitness
value of x addressed in MOEA/D [6].
The generational distance measures the
distance from a non-dominated solutions set Pf
to the true Pareto front Pt:
q
P
(d(x, Pt))2
m
ꢇ
ꢈ
ꢄ
X
x∈Pf
ꢄ
ꢄ
fws x λ,~z =
λi × (zi − fi (x))
(30)
ꢄ ꢄ
~
gd(Pf , Pt) =
(35)
ꢄ
ꢄ ꢄ
ꢄ ꢄ
Pf
i=1
20
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
where d(x, Pt) is the minimum Euclidean distance
between x and points in Pt:
4.1. Comparison to MOEAs
The performance of HVEA is compared
against NSGA2, SEAMO2, SPEA2, IBEAꢀ
and IBEAHV abbreviated as ns2, sea2, sp2,
v
t
m
+
X
ꢂ
ꢃ
2
*
d(x, Pt) = min
fi(x) − fi(x ) (36)
*
x ∈Pt
i=1
+
ibꢀ (ibe) and ibHV (ibhv) respectively. The
performance of these MOEAs is assessed using
four greedy repair methods for the multiple 0/1
knapsack problem, included binary encoding
approach proposed by Zitzler and Thiele [11],
permutation encoding approach proposed by
Mumford [12], te binary encoding (Tchebycheff)
and ws binary encoding (weighted sum) different
weighted linear scalarising approaches proposed
by Jaszkiewicz [13].
The generational distance measurement
indicates how close a non-dominated solutions
set to the true Pareto front is. In other words,
this metric indicates the convergence of a
non-dominated solutions set to a particular part
of the true Pareto front. The lower the value
of the generational distance, the closer of a
non-dominated solutions set to a particular part
of the true Pareto front, the better performance of
the algorithm.
Table 2: Generational distance
(permutation encoding)
The inverted generational distance works on
the opposite manner, measuring the diversity of a
non-dominated solutions set along the whole true
Pareto front or how close the true Pareto front to
a non-dominated solutions set is. The lower the
value of the inverted generational distance, the
more diversity of a non-dominated solutions set,
the better performance of the algorithm.
Instance hv1 hv100 ns2 sea2 sp2 ibꢀ+ ibHV
ks2 250 3.20 3.24 1.99 4.57 1.74 3.31 3.45
ks2 500 10.10 10.37 6.00 14.47 5.46 10.65 12.02
ks2 750 17.86 17.48 12.73 26.54 11.36 21.86 23.71
ks3 250 6.96 14.59 24.31 8.04 10.83 6.15 5.73
ks3 500 16.68 34.11 54.61 18.24 18.04 13.91 12.75
ks3 750 26.57 54.02 73.69 29.58 23.35 22.61 20.36
ks4 250 11.43 38.73 44.97 12.60 19.93 12.27 11.39
ks4 500 22.37 66.97 95.02 27.28 27.74 22.97 20.84
ks4 750 34.44 99.30 141.16 42.77 34.04 35.25 30.96
q
ꢂ
ꢃ
2
P
d(x, Pf )
x∈Pt
igd(Pf , Pt) =
(37)
|Pt|
The true Pareto front Pt is not available for
every instance of the multiple 0/1 knapsack
problem. Therefore we use the approximation
of Pt estimated by Jaszkiewicz [13] as suggested
in MOEA/D [6] to determine the generational
distance and inverted generational distance
measurements.
Table 3: Inverted generational distance
(permutation encoding)
Instance hv1 hv100 ns2 sea2 sp2 ibꢀ+ ibHV
ks2 250 10.13 9.06 8.45 11.49 8.67 8.48 8.48
ks2 500 10.96 9.98 10.07 13.93 9.86 9.50 8.62
ks2 750 48.07 46.02 42.92 64.14 44.40 38.07 38.01
ks3 250 21.11 5.59 10.37 18.35 12.41 13.25 15.74
ks3 500 48.73 15.58 25.60 43.54 39.14 33.71 37.69
ks3 750 69.65 27.06 39.73 61.16 61.03 48.20 54.06
ks4 250 19.33 9.39 13.57 14.98 14.53 12.30 15.03
ks4 500 43.24 19.71 30.15 33.84 39.64 29.83 35.81
ks4 750 68.36 33.56 49.35 54.58 65.93 49.84 57.93
4. Experimental Results and Discussion
We set µ = 0.01 to determine the rank of an
individual based on its fitness as in equation (11).
We experiment different neighbouring area
radius values ω which is given in (15) for HVEA.
The value of ω should lie in the range of 0 and
1 (0 ≤ ω ≤ 1). We report results for ω = 0.01
and ω = 1.0 which abbreviate as hv1 and hv100
respectively in both sets of figures and tables
and as HVEA0.01 and HVEA1.0 respectively in
the text.
Regarding
the
hyper-volume
metric,
Fig. 2,3,4,5 show that HVEA1.0 outperforms
or at least remains competitive to NSGA2,
SEAMO2, SPEA2 on all 9 instances of the
multiple 0/1 knapsack problem using all 4
different greedy repair methods. HVEA1.0 is
worse in only few cases out of 36 instance-
repair method combinations which are mainly
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
21
(a) ks2 250
(d) ks3 250
(g) ks4 250
(b) ks2 500
(c) ks2 750
(f) ks3 750
(i) ks4 750
(e) ks3 500
(h) ks4 500
Fig. 2: The percentage of hypervolume (permutation encoding).
Table 4: Running time in seconds (permutation)
problems. It looses its competitiveness to IBEA
+
for 2,4-knapsack problems. IBEAꢀ tends to be
Instance hv1 hv100 ns2 sea2 sp2 ibꢀ+ ibHV
the most effective MOEA to provide the coverage
(hyper-volume) for the Pareto sets. It is also
noticed that HVEA1.0 performs much better than
HVEA0.01. The reason is that HVEA0.01 only
looks at a tiny neighbouring area to determine
the crowding of an individual whereas HVEA1.0
examines the objective space as the whole picture,
i.e. every individuals contributing towards the
ks2 250
7
6
6
4
48
97
50
96
50
91
ks2 500 14
ks2 750 27
ks3 250 14
ks3 500 28
ks3 750 46
ks4 250 24
ks4 500 44
ks4 750 68
14
14
26
11
23
39
18
36
56
9
21
7
15
32
11
22
48
28
18
38
66
52
114
204
165 160 144
103 103 158
179 175 242
272 264 368
184 183 354
294 286 545
415 407 771
crowding of an individual.
Results on the
generational and inverted generational distance
show more prominent evidence regarding
this issue.
in the lowest dimension space, the 2-knapsack
problems. HVEA1.0 only outperforms IBEA
+
(both IBEAꢀ and IBEAHV) for the 3-knapsack
22
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
(a) ks2 250
(d) ks3 250
(g) ks4 250
(b) ks2 500
(c) ks2 750
(f) ks3 750
(i) ks4 750
(e) ks3 500
(h) ks4 500
Fig. 3: The percentage of hypervolume (binary encoding).
Table 5: Generational distance
(binary encoding)
distance from the true Pareto front to non-
dominated sets. Bold figures indicate the best
results for each instance. Both sets of tables
clearly indicate that HVEA1.0 provides better
diversity than HVEA0.01 whereas HVEA0.01
provides better convergence than HVEA1.0 due to
the use of neighbouring area radius ω. On the
performance regarding the generational distance
metric, HVEA0.01 outperforms NSGA2 and
SPEA2 but is competitive to SEAMO2 and IBEA.
SEAMO2 is one of the best algorithms to provide
convergence, although SEAMO2 only exploits
Pareto dominance without fitness, ranking and
crowding estimation. The reason behinds it is
Instance hv1 hv100
ks2 250 3.91 3.78
ns2
sea2
sp2
ibꢀ+ ibHV
4.17 6.16 4.09 6.95 6.99
ks2 500 11.42 11.03 13.39 15.10 13.78 17.66 18.07
ks2 750 26.86 27.10 32.87 33.91 31.67 29.48 30.05
ks3 250 8.45 13.95 22.16 8.62 14.18 10.06 9.77
ks3 500 20.83 34.04 50.69 21.39 32.63 24.47 22.82
ks3 750 35.71 55.08 74.63 40.78 52.56 38.58 35.84
ks4 250 14.18 33.23 38.58 13.91 27.10 15.37 14.99
ks4 500 31.13 75.38 96.31 30.97 66.35 32.20 28.85
ks4 750 52.80 119.03 150.58 52.89 106.08 53.40 44.81
Table 2,5,8,11 show the generational distance
from non-dominated sets to the true Pareto front.
Table 3,6,9,12 show the inverted generational
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
23
Table 8: Generational distance
that SEAMO2 is a steady-state algorithm which
allows offspring to compete for recombination
(te binary encoding)
immediately within the current generation. It
is not the case for the other population-based
MOEAs. However it leads SEAMO2 to be one
of the worst MOEAs to provide diversity for
non-dominated sets. The inverted generational
distance metric show that HVEA1.0 outperforms
other MOEAs in most cases of the 36 instance-
repair method combination. HVEA1.0 is clearly
the best algorithm to provide diversity in high
dimensional knapsack, 3,4-knapsack problems.
There is no strong evidence for which algorithm
is the best in the 2-knapsack problem.
Instance hv1 hv100 ns2 sea2 sp2 ibꢀ+ ibHV
ks2 250 4.42 4.46 5.18 6.94 5.41 6.84 6.93
ks2 500 16.33 16.60 17.68 23.92 17.62 18.35 19.01
ks2 750 44.67 43.98 46.88 64.90 44.96 41.62 43.42
ks3 250 13.99 18.92 28.85 11.79 21.06 13.64 13.50
ks3 500 36.35 44.72 70.27 33.93 50.40 31.61 29.84
ks3 750 64.61 78.85 113.39 67.64 85.45 59.47 56.29
ks4 250 22.82 40.94 47.44 16.53 37.85 17.53 17.29
ks4 500 51.52 89.40 114.67 40.71 89.75 43.05 37.09
ks4 750 90.24 144.94 183.81 79.45 148.89 78.48 67.98
Table 9: Inverted generational distance
(te binary encoding)
Instance hv1 hv100 ns2 sea2 sp2 ibꢀ+ ibHV
ks2 250 3.06 2.93 2.62 4.52 2.87 2.84 2.89
ks2 500 6.88 6.80 5.79 8.72 6.03 5.52 5.72
ks2 750 44.16 42.82 41.38 56.88 39.69 38.89 40.27
ks3 250 14.83 7.03 12.19 17.76 9.19 9.10 11.10
ks3 500 37.33 19.40 30.71 40.87 26.50 23.96 27.44
ks3 750 52.74 36.67 52.38 59.90 42.18 37.78 41.14
ks4 250 15.02 10.27 14.10 14.49 11.77 8.90 11.58
ks4 500 35.20 25.91 35.61 32.69 30.65 22.04 26.31
ks4 750 56.33 45.72 60.56 53.05 52.58 37.70 43.18
Table 6: Inverted generational distance
(binary encoding)
Instance hv1 hv100 ns2 sea2 sp2 ibꢀ+ ibHV
ks2 250 14.55 13.88 9.75 16.81 10.51 10.67 9.89
ks2 500 17.97 17.91 11.75 20.95 12.85 12.19 11.79
ks2 750 81.69 81.09 54.68 98.40 54.63 56.50 55.30
ks3 250 20.33 6.76 10.20 20.46 11.16 14.33 16.36
ks3 500 46.55 22.41 27.22 49.62 31.86 36.09 39.43
ks3 750 64.04 40.16 43.82 72.18 48.56 53.63 56.87
ks4 250 18.37 8.47 12.53 16.64 12.67 13.45 15.90
ks4 500 39.55 22.05 30.54 37.02 30.33 29.45 34.11
ks4 750 62.33 38.87 51.83 60.39 51.45 49.51 55.66
Table 10: Running time in seconds (te binary)
Instance hv1 hv100 ns2 sea2 sp2 ibꢀ+ ibHV
ks2 250 13
ks2 500 31
ks2 750 58
ks3 250 22
ks3 500 51
ks3 750 93
ks4 250 36
ks4 500 75
ks4 750 134
13
31
58
28
61
109
66
143
12
30
56
18
46
86
30
66
10
54 55
60
26 112 109 117
51 192 182 198
16 109 110 150
40 195 191 256
76 310 301 392
24 195 194 367
57 307 308 565
Table 7: Running time in seconds (binary)
Instance hv1 hv100 ns2 sea2 sp2 ibꢀ+ ibHV
ks2 250
8
8
18
32
20
39
61
50
107
168
7
6
50 51
99 96
51
95
ks2 500 18
ks2 750 31
ks3 250 16
ks3 500 31
ks3 750 50
ks4 250 27
ks4 500 47
ks4 750 73
17
30
12
25
43
19
37
60
14
26 165 157 149
9
21 175 170 236
36 266 256 350
14 185 183 351
28 281 278 537
48 396 392 746
256 129 108 462 461 810
104 103 145
to IBEA with much faster computational time
than IBEA.
4.2. Comparison to MOEA/D
The running time (in seconds) is summarised in
table 4,7,10,??. SPEA2 and IBEA are the slowest
algorithms due to the complexity in computation
of fitness values (SPEA2) and indicator values
(IBEA). SEAMO2, NSGA2, HVEA are much
faster algorithms.
Taken all the above criteria into account, it
is concluded that HVEA outperforms NSGA2,
SEAMO2, and SPEA2 but remains competitive
We compare HVEA against MOEA/D
separately from other MOEAs due to the setting
and approach employed by MOEA/D. As
discussed in Section 3.2.1, it is problematic
to alter the population size N in MOEA/D
to that population size S suggested for the
multiple 0/1 knapsack problem. Furthermore,
MOEA/D stores all non-dominated solutions
found so far in an external archive which is
24
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
(a) ks2 250
(d) ks3 250
(g) ks4 250
(b) ks2 500
(c) ks2 750
(e) ks3 500
(f) ks3 750
(h) ks4 500
(i) ks4 750
Fig. 4: The percentage of hypervolume (te binary encoding).
then reported as the final result. Therefore, it is
suggested to adapt HVEA to the setting similar
to MOEA/D included the population size and
an external archive for non-dominated solutions
found so far. Experimental results, including
figures and tables, are presented in a similar
format as in Section 4.1. MOEA/D employed
Tchebycheff and weight sum approach for the
fitness calculation are abbreviated as “mo/t” and
“mo/w”. The extracted population, the truncated
external population, is also reported so that the
size of the extracted population does not exceed
the suggested population size S . NSGA2’s
truncation approach is used here. Fig. 6,7,8,9
present extracted populations using prefix “e”.
Table 11: Generational distance
(ws binary encoding)
Instance hv1 hv100 ns2 sea2 sp2 ibꢀ+ ibHV
ks2 250 2.02 1.96
ks2 500 6.13 6.07
2.24 3.89 2.07 2.91 2.92
6.70 10.80 6.74 8.07 8.59
ks2 750 16.53 16.38 18.42 27.96 17.26 17.40 17.62
ks3 250 9.52 12.17 20.82 7.26 14.23 7.87 7.46
ks3 500 20.87 27.25 47.72 16.50 31.60 16.88 15.90
ks3 750 30.72 42.91 67.05 25.88 45.73 27.35 24.99
ks4 250 17.06 29.66 35.52 12.52 27.13 13.59 13.35
ks4 500 33.58 60.66 82.87 25.34 59.50 27.40 23.68
ks4 750 54.58 97.91 131.27 42.68 96.22 44.91 38.29
MOEA/D outperforms HVEA in most cases
out of 36 instance-repair method combinations
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
25
(a) ks2 250
(d) ks3 250
(g) ks4 250
(b) ks2 500
(c) ks2 750
(f) ks3 750
(i) ks4 750
(e) ks3 500
(h) ks4 500
Fig. 5: The percentage of hypervolume (ws binary encoding).
Table 12: Inverted generational distance
(ws binary encoding)
on the hyper-volume metric and the inverted
generational distance metric. The convergence
of MOEA/D is only better than that of HVEA
in the linear scalarising greedy repair method.
However the running time of MOEA/D is
significantly higher than that of HVEA. In the
other 2 repair methods, HVEA0.01 outperforms
MOEA/D regarding the generational distance
metric. Under a restriction of the size of the
population, where the truncation is applied to the
external population, HVEA is able to compete
to MOEA/D. However HVEA is much faster
than MOEA/D. The running time (in seconds) is
reported in Table 18, 15, 21, 24.
Instance hv1 hv100 ns2 sea2 sp2 ibꢀ+ ibHV
ks2 250 3.80 3.40 2.70 4.90 3.05 3.14 3.34
ks2 500 7.02 6.62 4.97 9.05 5.27 4.53 4.78
ks2 750 41.14 40.31 28.86 52.55 27.86 26.11 27.82
ks3 250 16.83 4.78 9.71 17.31 8.93 9.06 11.05
ks3 500 40.25 13.21 22.43 40.48 24.74 23.92 27.69
ks3 750 56.43 23.72 35.38 59.39 37.02 36.85 41.70
ks4 250 16.08 7.75 11.47 14.33 11.35 9.17 11.85
ks4 500 37.35 18.10 27.27 32.41 28.35 22.44 27.55
ks4 750 58.56 31.92 46.15 52.21 47.00 38.13 45.26
4.3. Further Discussion
HVEA employs a parameter µ in equation (11)
to define the size of each front. In this paper,
26
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
(a) ks2 250
(d) ks3 250
(g) ks4 250
(b) ks2 500
(c) ks2 750
(f) ks3 750
(i) ks4 750
(e) ks3 500
(h) ks4 500
Fig. 6: The percentage of hypervolume (permutation encoding).
µ is set to 0.01, which is not only bounded to
the multiple 0/1 knapsack problem, could be used
for other multi-objective optimisation problems.
It is suggested that this parameter µ is fixed to
0.01 rather than tunable. The more important
parameter in HVEA is ω, the neighbouring area
radius, which is in the range [0, 1]. Section 4.1
shows a significant difference in performance
tune ω to obtain desirable requirement. For
example, ω could be set to 1.0 for theoretical
problems such as the multiple 0/1 knapsack
problem to obtain non-dominated set with better
diversity (generational distance metric) and better
coverage (hyper-volume metric). However, for
real-world applications, where running time is
expensive, ω
=
0.01 could be deployed.
between HVEA0.01 and HVEA1.0
.
HVEA1.0
Furthermore, in real-world applications, extreme
solutions are likely less of interest whereas better
convergence, which could strike a good balance
amongst objectives, is more important. We
experiment different ω values in the range [0, 1]
gives a much better performance regarding the
hyper-volume metric and the diversity of the non-
dominated set. However HVEA0.01 is better
in convergence than HVEA1.0
.
One could
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
27
Table 15: Running time in seconds (permutation)
but the performance of HVEA degrades quite
significantly. Therefore, it is suggested to use
Instance hv1 hv100 mo/t mo/w
ω = 1.0 for better diversity and better coverage
and 0.01 ≤ ω ≤ 0.05 for better convergence and
faster computational time.
ks2 250
ks2 500
ks2 750
ks3 250
ks3 500
ks3 750
ks4 250
4
8
15
16
29
46
47
5
9
16
26
3
4
9
19
38
133
181
315
972
8
17
40
118
172
357
1025
84
Table 13: Generational distance
(permutation encoding)
177
156
602
870
ks4 500 112
ks4 750 163
External Population
Extracted Population
1784 1807
Instance hv1 hv100 mo/t mo/w hv1 hv100 mo/t mo/w
ks2 250 3.16 3.21 5.98 7.00 3.19 3.23 6.18 7.18
ks2 500 10.00 10.17 17.57 16.05 10.08 10.33 18.64 16.84
ks2 750 17.49 17.06 29.28 24.43 17.78 17.40 33.04 26.86
ks3 250 2.08 2.84 4.64 3.86 7.99 9.90 14.77 13.18
ks3 500 4.97 6.81 10.31 6.84 16.72 27.44 37.67 26.73
ks3 750 9.29 12.21 17.22 9.99 26.94 48.53 59.89 38.67
ks4 250 2.24 3.75 3.80 2.92 14.11 30.34 20.63 17.17
ks4 500 4.23 6.79 8.21 5.11 24.22 57.09 39.98 34.94
ks4 750 6.79 10.63 13.32 7.47 34.17 86.46 61.01 51.07
Table 16: Generational distance
(binary encoding)
External Population
Extracted Population
Instance hv1 hv100 mo/t mo/w hv1 hv100 mo/t mo/w
ks2 250 3.91 3.78 4.47 7.28 3.91 3.78 4.54 7.29
ks2 500 11.42 11.03 15.53 17.39 11.42 11.03 15.53 17.39
ks2 750 26.86 27.10 34.07 34.88 26.86 27.10 34.07 34.88
ks3 250 3.26 4.08 4.77 4.57 10.18 12.28 14.33 14.01
ks3 500 10.67 10.69 12.18 10.15 24.41 31.16 37.44 32.18
ks3 750 22.54 22.11 23.62 19.32 38.93 53.36 64.60 52.75
ks4 250 3.26 4.46 4.08 3.63 16.27 26.40 21.17 19.24
ks4 500 9.37 11.80 10.06 8.09 35.47 69.12 49.77 44.30
ks4 750 18.56 20.13 16.93 12.85 57.80 114.62 80.97 71.23
Table 14: Inverted generational distance
(permutation encoding)
External Population
Extracted Population
Instance hv1 hv100 mo/t mo/w hv1 hv100 mo/t mo/w
ks2 250 10.13 9.06 3.07 4.26 10.13 9.06 3.07 4.26
ks2 500 10.96 9.98 6.60 6.55 10.96 9.98 6.60 6.55
ks2 750 48.07 46.02 37.36 31.90 48.07 46.02 37.37 31.90
ks3 250 15.86 5.22 5.20 6.26 15.95 6.52 7.83 8.37
ks3 500 42.80 15.02 16.43 15.10 42.85 16.99 19.83 18.18
ks3 750 63.84 25.89 30.90 23.10 63.89 27.77 34.74 27.54
ks4 250 14.97 5.32 5.76 5.91 15.08 11.06 12.70 9.77
ks4 500 38.54 14.94 16.82 13.49 38.60 22.45 31.17 20.77
ks4 750 63.77 29.00 31.73 22.26 63.87 36.84 49.64 31.97
Table 17: Inverted generational distance
(binary encoding)
External Population
Extracted Population
Instance hv1 hv100 mo/t mo/w hv1 hv100 mo/t mo/w
ks2 250 14.55 13.88 2.78 5.35 14.55 13.88 2.79 5.35
ks2 500 17.97 17.91 5.36 7.37 17.97 17.91 5.36 7.37
ks2 750 81.69 81.09 33.08 35.57 81.69 81.09 33.08 35.57
ks3 250 16.31 8.36 5.22 8.37 16.36 8.93 7.41 9.67
ks3 500 41.47 25.90 17.05 24.17 41.50 26.38 20.07 25.59
ks3 750 61.05 42.30 30.87 34.95 61.08 42.67 33.83 36.71
ks4 250 14.87 6.95 6.10 8.84 14.97 10.12 11.59 10.91
ks4 500 34.46 19.55 17.22 19.04 34.53 24.15 24.86 22.39
ks4 750 56.85 36.89 32.02 32.62 56.92 41.66 43.02 36.28
We further examine the performance of HVEA
against other MOEAs using a fixed running time,
although this approach is not widely adopted
due to its low reliability. The running time (in
seconds), reported in Table 25, is deduced from
the minimum running time of all MOEAs on
each knapsack instances over 2000 generations.
The running time for the linear scalarising greedy
repair method proposed by Jaszkiewicz [13]
is twice as much as the ones proposed by
Mumford [12] and Zitzler and Thiele [11].
Table 18: Running time in seconds (binary)
Instance
ks2 250
ks2 500
ks2 750
ks3 250
ks3 500
ks3 750
ks4 250
ks4 500
ks4 750
hv1
8
hv100
8
19
33
35
56
74
139
285
427
mo/t
5
mo/w
5
18
32
25
39
56
55
82
110
13
13
25
25
18
18
40
57
87
39
59
113
359
536
Under the time restriction, HVEA outperforms
342
538
+
NSGA2, SPEA2, IBEAꢀ and IBEAHV. HVEA
only outperforms SEAMO2 on 2,3-knapsack
problems. SEAMO2 is slightly better than HVEA
on 4-knapsack problems. It is suggested that due
to the steady-state approach and the simple Pareto
dominance, SEAMO2 is able to perform more
evaluations than any other MOEAs which could
lead to a better performance. HVEA is slightly
worse than but competitive to MOEA/D.
28
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
(a) ks2 250
(d) ks3 250
(g) ks4 250
(b) ks2 500
(c) ks2 750
(f) ks3 750
(i) ks4 750
(e) ks3 500
(h) ks4 500
Fig. 7: The percentage of hypervolume (binary encoding).
Table 19: Generational distance
(te binary encoding)
Table 20: Inverted generational distance
(te binary encoding)
External Population Extracted Population
External Population
Extracted Population
Instance hv1 hv100 mo/t hv1 hv100 mo/t
Instance hv1
ks2 250
ks2 500
hv100
2.93
6.80
mo/t
hv1
3.06
6.88
hv100
2.93
6.80
42.82
7.68
mo/t
1.53
ks2 250 4.40
ks2 500 16.33 16.60
ks2 750 44.67 43.98 23.54 44.67 43.98 23.63
ks3 250 5.39 6.09 4.54 14.62 15.61 14.80
4.43
2.83 4.41
4.45
3.05
3.06
6.88
1.52
3.69
9.88 16.33 16.60 10.03
3.69
ks2 750 44.16
ks3 250 10.22
ks3 500 30.69
ks3 750 47.42
ks4 250 10.89
ks4 500 29.43
ks4 750 49.19
42.82
6.18
19.09
35.89
7.15
22.33
42.32
26.22 44.16
5.03 10.66
15.73 30.86
31.87 47.63
5.52
16.88 29.56
32.32 49.37
26.22
7.45
ks3 500 16.66 15.73 10.49 37.71 42.51 37.10
ks3 750 35.66 30.38 19.88 66.81 77.08 65.21
20.57
37.57
12.61
29.20
50.01
18.94
35.11
12.42
24.87
42.92
ks4 250 4.91
ks4 500 14.45 14.45
ks4 750 30.20 24.54 13.47 93.91 141.43 65.90
5.75
3.61 22.65 34.78 20.41
8.09 52.51 83.35 46.14
11.17
K.N. Le, D. Landa-Silva / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 32, No. 1 (2016) 10–32
29
(a) ks2 250
(d) ks3 250
(g) ks4 250
(b) ks2 500
(c) ks2 750
(f) ks3 750
(i) ks4 750
(e) ks3 500
(h) ks4 500
Fig. 8: The percentage of hypervolume (te binary encoding).
Table 21: Running time in seconds (te binary)
Table 22: Generational distance
(ws binary encoding)
Instance
ks2 250
ks2 500
ks2 750
ks3 250
ks3 500
ks3 750
ks4 250
ks4 500
ks4 750
hv1
14
33
62
26
58
102
49
hv100
13
32
59
42
mo/t
9
External Population
Instance hv1
hv100
1.85
mo/w
25
ks2 250 1.91
ks2 500 6.13
ks2 750 16.53
ks3 250 2.61
ks3 500 7.64
ks3 750 14.23
ks4 250 3.02
ks4 500 7.42
ks4 750 14.49
1.81
3.74
8.10
1.84
3.15
5.15
1.56
2.72
4.34
51
6.07
16.38
3.24
24
81
60
126
164
331
527
105
98
424
769
8.11
14.80
3.66
92
157
9.11
16.17
Tải về để xem bản đầy đủ
Bạn đang xem 20 trang mẫu của tài liệu "Hyper-volume evolutionary algorithm", để tải tài liệu gốc về máy hãy click vào nút Download ở trên
File đính kèm:
- hyper_volume_evolutionary_algorithm.pdf