Evolutionary and Estimation of Distribution
Algorithms for Unconstrained, Constrained,
and Multiobjective Noisy Combinatorial
Optimisation Problems
Aishwaryaprajna
School of Computer Science, University of Birmingham, Birmingham,
Reino Unido
aishwaryaprajna@gmail.com
Jonathan E. Rowe
School of Computer Science, University of Birmingham, Birmingham,
United Kingdom and The Alan Turing Institute, Londres, Reino Unido
J.E.Rowe@cs.bham.ac.uk
yo
D
oh
w
norte
oh
a
d
mi
d
F
r
oh
metro
h
t
t
pag
:
/
/
d
i
r
mi
C
t
.
metro
i
t
.
/
/
mi
d
tu
mi
v
C
oh
a
r
t
i
C
mi
–
pag
d
yo
F
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
mi
v
C
oh
_
a
_
0
0
3
2
0
pag
d
.
/
F
b
y
gramo
tu
mi
s
t
t
oh
norte
0
7
S
mi
pag
mi
metro
b
mi
r
2
0
2
3
https://doi.org/10.1162/evco_a_00320
Abstracto
We present an empirical study of a range of evolutionary algorithms applied to var-
ious noisy combinatorial optimisation problems. There are three sets of experiments.
The first looks at several toy problems, such as OneMax and other linear problems.
We find that UMDA and the Paired-Crossover Evolutionary Algorithm (PCEA) son los
only ones able to cope robustly with noise, within a reasonable fixed time budget. En
the second stage, UMDA and PCEA are then tested on more complex noisy problems:
SubsetSum, Knapsack, and SetCover. Both perform well under increasing levels of
ruido, with UMDA being the better of the two. In the third stage, we consider two
noisy multiobjective problems (CountingOnesCountingZeros and a multiobjective
formulation of SetCover). We compare several adaptations of UMDA for multiob-
jective problems with the Simple Evolutionary Multiobjective Optimiser (SEMO) y
NSGA-II. We conclude that UMDA, and its variants, can be highly effective on a va-
riety of noisy combinatorial optimisation, outperforming many other evolutionary
algoritmos.
Palabras clave
Noisy combinatorial optimisation, noisy multiobjective optimisation, expected run-
tiempo, crossover, estimation of distribution algorithms, NSGA-II.
1
Introducción
Realistic optimisation problems are often affected with noisy fitness measurements. El
recent theoretical analyses of evolutionary algorithms (EAs) on noisy problems defined
in discrete spaces are mostly focused on simple and classical benchmark problems, semejante
as OneMax. Harder and more realistic combinatorial problems such as Knapsack or
SetCover, in presence of noisy fitness evaluations, are not easily amenable to theoreti-
análisis calórico. This article discusses a thorough empirical analysis of an array of EAs on
several simple and harder noisy combinatorial problems. This study attempts to iden-
tify which EAs to choose when solving realistic noisy combinatorial problems.
Noise may affect fitness evaluation in a number of ways. It may be considered as
previo, in which the search point is randomly tampered with, and fitness evaluation is
Manuscrito recibido: 4 Agosto 2021; revisado: 5 Julio 2022 y 2 Febrero 2023; aceptado: 19 Febrero 2023.
© 2023 Instituto de Tecnología de Massachusetts
Computación evolutiva 31(3): 259–285
Aishwaryaprajna and J.E. Rowe
performed on the noisy search point. Alternativamente, posterior noise (which will be the
focus of our study) is where a random value gets added to the fitness of a search point
during the optimisation process. The same search string would have a different fitness
value each time it is being evaluated. With more complex combinatorial problems (p.ej.,
ones with constraints) the noise may enter in different ways (p.ej., in the evaluation of
those constraints).
An early theoretical result by Droste (2004) examined the performance of the hill-
climber (1 + 1)-EA on OneMax with prior noise. This was generalised to the (μ + λ)-EA
by Gießen and Kötzing (2016), showing that populations can help in both prior and pos-
terior noise. They show the (1 + 1)-EA, sin embargo, can tolerate posterior Gaussian noise
only when the variance is very small (less than 1/(4 iniciar sesión sustantivo, masculino—)). It has been recognised for a
long time that the population size can affect the ability of an EA to handle noise (Gold-
berg et al., 1991; Rattray and Shapiro, 1997). A more recent theoretical study by Dang and
Lehre (2015) shows that a low mutation rate enables a particular mutation-population
algorithm to handle arbitrary posterior noise for the OneMax problem in polynomial
tiempo, although the bounds given are large. Similarmente, the compact genetic algorithm
(cGA) is shown to handle noise with (grande) polynomial runtime (Friedrich et al., 2015).
A better asymptotic runtime for OneMax with posterior Gaussian noise is proved for
the Paired-Crossover Evolutionary Algorithm (PCEA) which just uses crossover, y
no mutation (Prügel-Bennett et al., 2015).
Por supuesto, it is possible to handle noise simply by resampling the fitness of a poten-
tial solution many times, and taking the average as an estimate of the true fitness. Sup-
pose the noisy problem is defined by taking a deterministic fitness function and adding
Gaussian noise with mean 0 and variance σ 2. There is a general result (Akimoto et al.,
2015) that states if the runtime of a black-box algorithm on a problem with no noise is T ,
then σ 2 log T samples are required at each step leading to a runtime of σ 2T log T . En el
case of OneMax, the most efficient algorithm (Anil and Wiegand, 2009) has a runtime of
(cid:2)(n/ log n). Using this algorithm with resampling, gives a runtime for noisy OneMax
de (cid:2)(σ 2n). Por el contrario, the PCEA algorithm (Prügel-Bennett et al., 2015), when σ 2 = norte,
has a runtime of O(norte(iniciar sesión sustantivo, masculino—)2) which is already faster than the resampling algorithm.
It has been suggested by Doerr and Sutton (2019) that using the median rather than
mean provides a better estimate when resampling, but this is significant only when the
variance is small (less than a constant).
A recent study by Rowe and Aishwaryaprajna (2019) of a new voting algorithm on
OneMax shows a runtime of O(n log n), when the variance of the noise distribution is
pag 2 =O(norte) and in O(pag 2 iniciar sesión sustantivo, masculino—) when the noise variance is greater than this. This upper
bound is the best proven runtime that we are aware of to date. Some empirical results
show that the use of voting in population-based algorithms (UMDA, PCEA, and cGA)
are effective for large population sizes. The voting algorithm is shown to solve the Jump
función, which has a fitness gap next to the global optimum, en O(n log n) fitness eval-
uations with high probability any gap size m < (1 − α)n where, the constant α ∈ [ 1
2 , 1].
The performance of voting algorithm and the cGA are also analysed for other variants
of the Jump function in the recent paper by Witt (2021). However, in our work, problems
with fitness gaps have not been considered.
In this article, we are interested in whether any of the algorithms with polyno-
mial theoretical runtimes for noisy OneMax would be capable of solving combinatorial
problems with added noise in practice, when given a reasonable but fixed time budget.1
1This article is an extended version of Aishwaryaprajna and Rowe (2019).
260
Evolutionary Computation Volume 31, Number 3
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
.
/
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
EAs and EDAs for Solving Noisy Combinatorial Problems
We proceed in three stages. First, we will experimentally compare a collection of algo-
rithms on noisy OneMax and noisy Linear problems, to see which can find solutions
within a reasonable amount of time (to be defined below), bearing in mind that the
asymptotic bounds for some of these algorithms, while polynomial, are actually very
large. Second, we will take those algorithms which pass this first test, and see how
well they handle noise in three combinatorial problems: SubsetSum, Knapsack, and
SetCover. We choose these, as they have a “packing” structure which might make them
amenable to algorithms which can solve noisy OneMax efficiently. We generate random
problem instances within the “easy” regime (so that the algorithms can be expected to
solve them when there is no noise) and then empirically study how they degrade with
added Gaussian noise.
There has been considerable research in the last decades on stochastic combinato-
rial optimisation problems, like the stochastic knapsack problem (refer to Steinberg and
Parks, 1979; Sniedovich, 1980; Ross and Tsang, 1989; Henig, 1990; Carraway et al., 1993;
Fortz et al., 2013; and Bianchi et al., 2009 for a survey). In stochastic knapsack formula-
tions, in some instances the profits are considered as random variables (Steinberg and
Parks, 1979; Sniedovich, 1980; Carraway et al., 1993; Henig, 1990), as well as, the distri-
bution of weights are known in some formulations (Fortz et al., 2013). This means, that
the approaches in handling stochastic problems involve having access to more details
of the problem, for example, the distribution of the decision variables is known. How-
ever, in this article, we assume that the noise contribution to the fitness value is part of
the black-box, and so we do not have access to details such as the noise distribution, or
the exact distribution of the decision variables.
In the last stage of our analysis, we look at noisy multiobjective problems. Initially,
we analyse the performance of a collection of multiobjective algorithms on a toy mul-
tiobjective problem COCZ without and with high levels of noise and we attempted to
identify which algorithms perform better. We study the simple hill-climber algorithm
SEMO, the popular NSGA-II, and some other algorithms designed on the basis of our
previous experimental results. We compare our algorithms on the basis of the perfor-
mance indicator, hypervolume, which provides an analysis of the spread of the non-
dominated solutions found, in a reasonable time budget. We then formulate the noisy
constrained SetCover problem as a multiobjective problem and we empirically analyse
the performance of the better algorithms on this.
It should be noted that in our empirical results, while error bars are not always
shown, the Mann–Whitney test was used on all relevant comparisons, and results are
significant at the 95% level unless explicitly indicated.
Notation: We use the convention that [expr] equals 1 if expr is true, and 0 otherwise.
2
Problem Definitions—Noisy Single Objective Problems
The problems studied in this article are defined on a Boolean search space of bit strings
of length n. Let N (0, σ ) denote a random number drawn from a normal distribution with
mean zero, and standard deviation σ , which will be freshly generated at each fitness
function evaluation.
2.1 Unconstrained Single-Objective Noisy Problems
The first problem is OneMax, whose fitness function is defined as,
OneMax(x) =
n(cid:2)
i=1
xi.
Evolutionary Computation Volume 31, Number 3
261
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
.
/
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Aishwaryaprajna and J.E. Rowe
The fitness function is to be maximised, so the optimum would be the all-ones string,
1n.
When the fitness evaluation is tampered with random noise, the fitness function
becomes as follows,
NoisyOneMax(x) =
n(cid:2)
i=1
xi + N (0, σ ).
The WeightedLinear problem is defined with reference to n positive weights
w1, . . . , wn as follows, where the fitness is to maximised,
WeightedLinear(x) =
n(cid:2)
i=1
xiwi,
with corresponding noisy variant,
NoisyWeightedLinear(x) =
n(cid:2)
i=1
xiwi + N (0, σ ).
In generating random problem instances, we draw the weights uniformly at ran-
dom from the range 1, . . . , 100. Thus we avoid more extreme instances such as BinVal
(in which wi = 2i−1 for each i = 1, . . . , n). The reason for this is that when the distribu-
tion of weights is highly skewed, the addition of noise is irrelevant for those bits with
very high weights, yet completely overwhelms bits with weights lower than the typi-
cal noise level. Thus most algorithms will find the more significant bits, and fail on the
remainder.
With reference to n positive weights w1, . . . , wn and a target θ , the SubsetSum prob-
lem is defined as,
SubsetSum(x) = |θ −
n(cid:2)
i=1
xiwi|.
In presence of noisy fitness evaluations, the fitness function can be written as follows,
NoisySubsetSum(x) = |θ −
n(cid:2)
i=1
xiwi| + N (0, σ ).
SubsetSum can be seen as a generalisation of the WeightedLinear problem (in
which the target is θ = 0). In our experiments, we generate instances by choosing
weights uniformly at random from 1, . . . , 100. We take the target to be two-thirds of
the sum of the weights (we have run experiments for other choices of θ and found that
they do not significantly affect the empirical observations).
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
.
/
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
2.2 Constrained Single-Objective Noisy Problems
The Knapsack problem is defined with respect to a set of positive weights w1, . . . , wn,
a capacity C and positive profits p1, . . . , pn as follows,
Knapsack(x) =
(cid:3) (cid:4)
n
i=1 xipi
(cid:4)
(cid:4)
if
n
i=1 xiwi ≤ C
C −
n
i=1 xiwi otherwise.
Random instances choose weights and profits uniformly from 1, . . . , 100, and the ca-
pacity is two-thirds of the sum of the weights. We consider two noisy variants of
262
Evolutionary Computation Volume 31, Number 3
EAs and EDAs for Solving Noisy Combinatorial Problems
the Knapsack problem. The first version simply considers posterior additive noise as
before:
NoisyKnapsackV1(x) = Knapsack(x) + N (0, σ ).
In the second version, the presence of noise in the judgement with respect to the
weights is considered,
Wσ (x) =
n(cid:2)
i=1
xiwi + N (0, σ ).
If this (noisy) weight does not exceed the capacity, we then evaluate (noisily), the profit.
Otherwise we return the excess weight:
NoisyKnapsackV2(x) =
(cid:3) (cid:4)
n
i=1 xipi + N (0, σ )
if Wσ (x) ≤ C
C − Wσ (x)
otherwise.
Note that noise is added to the weight just once, when the constraint is checked, and
the same value used to report the fitness value, in the case the constraint is violated.
The SetCover problem finds a minimal covering of m elements with a collection
of sets from n predefined subsets. A Boolean matrix aij with n-rows and m-columns is
used to define the n subsets c1, . . . , cn:
ai,j = [i ∈ cj ].
The optimal collection of the sets would have the least number of the sets needed to
cover all the m elements. The SetCover problem has several real-world applications
such as the airline crew scheduling problem.
The problem can be defined as a constrained single-objective one, as well as a single-
objective problem with a penalty term. The problem can also be defined as a multiob-
jective problem (discussed later).
The ConstrainedSetCover problem has a constraint that checks if the solution
covers each of the m elements. The optimal solution would have the least number of
sets needed to cover all the m elements. It is defined as follows,
ConstrainedSetCover(x) =
n(cid:2)
j =1
xj
subject to
n(cid:2)
j =1
xj aij ≥ 1, i ∈ 1, . . . , m.
For comparison-based algorithms, we always prefer feasible solutions instead of in-
feasible solutions. Two feasible solutions are compared by their fitness values, whereas
two infeasible solutions by their constraint violations. The noisy version of the problem
arises if the judgements regarding the number of elements uncovered and the number
of the subsets required is noisy.
NoisyConstrainedSetCover(x) =
n(cid:2)
j =1
xj + N (0, σ )
subject to
n(cid:2)
j =1
xj aij + N (0, σ ) ≥ 1, i ∈ 1, . . . , m.
Evolutionary Computation Volume 31, Number 3
263
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
/
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Aishwaryaprajna and J.E. Rowe
The fitness function of SetCover problem can also be defined by including a penalty
term such that, if elements are under-covered by the considered collection of sets, a huge
penalty μ is incurred.
PenaltySetCover(x) =
n(cid:2)
j =1
xj + μ
(cid:2)
i
(cid:5)
(cid:6)
max
0,
1 −
(cid:7)(cid:8)
aij xj
.
n(cid:2)
j =1
This gives rise to a corresponding noisy variant:
NoisyPenaltySetCover(x) =
n(cid:2)
j =1
xj + μ
(cid:2)
i
(cid:5)
(cid:6)
max
0,
1 −
n(cid:2)
j =1
(cid:7)(cid:8)
aij xj
+ N (0, σ ).
3 Algorithms Chosen for Noisy Single-Objective Optimisation
The (1 + 1)-EA
3.1
The (1 + 1)-EA uses a bitwise mutation operator that produces an offspring by flipping
each bit of the parent string independently with probability 1/n. This can be considered
as a randomised or stochastic hill-climber which considers only one point in the search
space at a time and proceeds by trying to find a point which has a superior function
value. In each iteration, only one function evaluation takes place. The expected runtime
of the (1 + 1)-EA solving the non-noisy OneMax is O(n log n). The runtime remains
polynomial in the posterior Gaussian noise case for σ 2 < 1/(4 log n), so we do not expect
this algorithm to cope with anything but the smallest noise levels (Gießen and Kötzing,
2016).
3.2 Mutation-Population Algorithm
It has long been recognised that populations can help an EA handle noise. The paper
by Goldberg et al. (1991) developed a population sizing equation and instigated the
adoption of variance-based population sizing. Rattray and Shapiro (1997) showed that
in weak selection limit, effects of Gaussian noise could be overcome by an appropriate
increase of the population size. More recently, a population-based, non-elitist EA was
analysed by Dang and Lehre to study how it optimises the noisy OneMax problem with
uniform, Gaussian and exponential posterior noise distributions (Dang and Lehre, 2015,
2016). They considered a recently developed fitness-level theorem for non-elitist pop-
ulations to estimate the expected running time for the said problems in noisy environ-
ment. In the case of additive Gaussian noise N (0, σ 2) with mutation rate χ
3σ n and
n
population size λ = bσ 2 ln n (where a and b are constants), the considered algorithm
optimises the OneMax problem in expected time O(σ 7n ln(n) ln(ln (n))). Similar results
were shown for uniform and exponential noise distributions. Note that this is poten-
tially very large, when the noise is large—in excess of n4.5 when σ =
n, although of
course this is an upper bound, and we do not know the constants.
= a
√
3.3 Compact Genetic Algorithm (cGA)
The compact GA (cGA) is an EDA, introduced by Harik et al. (1999). cGA is able to av-
erage out the noise and optimise the noisy OneMax problem in expected polynomial
time, when the noise variance σ 2 is bounded by some polynomial in n, as suggested in
Friedrich et al. (2015). The paper introduced the concept of graceful scaling in which
264
Evolutionary Computation Volume 31, Number 3
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
.
/
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
EAs and EDAs for Solving Noisy Combinatorial Problems
the runtime of an algorithm scales polynomially with noise intensity, and suggested
that cGA is capable of achieving this. It is also suggested that there is no threshold
point in noise intensity at which the cGA algorithm begins to perform poorly (by which
they mean having super-polynomial runtime). They proved that cGA is able to find
the optimum of the noisy OneMax problem with Gaussian noise of variance σ 2 after
n log n), with probability 1 − o(1). Note that
O(Kσ 2
this upper bound is in excess of n3 when σ =
n log Kn) steps when K = ω(σ 2
n.
√
√
√
3.4
Population-Based Incremental Learning (PBIL)
The algorithm PBIL, proposed by Baluja (1994), combines genetic algorithms and com-
petitive learning for optimising a function. We have included this algorithm as it is in
some ways similar to the cGA, so we might expect it to have similar performance. We
are not aware of any theoretical analysis of this algorithm on noisy problems. The run-
time of PBIL on OneMax is known to be O(n3/2 log n), for suitable choice of λ (Wu et al.,
2017).
3.5 Univariate Marginal Distribution Algorithm (UMDA)
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
/
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
The Univariate Marginal Distribution Algorithm (UMDA) proposed by Mühlenbein
(1997) belongs to the EDA schema. In some ways, it is therefore similar to cGA and PBIL.
However, it can also be viewed as generalising the genepool crossover scheme, in which
bits are shuffled across the whole population (within their respective string positions).
We have included UMDA then, to see if its behaviour is more like cGA and PBIL on the
one hand (which emphasise an evolving distribution over bit values), or like PCEA on
the other (which emphasises crossover). The UMDA algorithm initialises a population
of λ solutions, and sorts the population according to the fitness evaluation of each candi-
date solution. The best μ members of the population are selected to calculate the sample
distribution of bit values in each position. The next population is generated from this
distribution. There are two variants of UMDA, depending on whether the probabilities
are constrained to stay away from the extreme values of 0 and 1, or not. It is known
that if the population size is large enough (that is, (cid:3)(
n log n)), then this handling of
probabilities at the margins is not required (Witt, 2017). Since we will work with a large
population (to match the PCEA algorithm described below), we will not employ margin
handling, unless otherwise stated. In our experiments we will take μ = λ/2. We are not
aware of any theoretical results concerning UMDA on problems with posterior noise,
but the runtime on OneMax is known to be O(n log n) for μ = (cid:2)(
n log n)—see Witt
(2017).
√
√
3.6
Paired-Crossover EA (PCEA)
Recently, the recombination operator has been suggested to be considerably beneficial
in noisy evolutionary search. Prügel-Bennett et al. (2015) considered the problem of
√
solving OneMax with noise of order σ =
n and analysed the runtime of an evolution-
ary algorithm consisting only of selection and uniform crossover, the paired-crossover
n log n then the required num-
EA (PCEA). They show that if the population size is c
ber of generations is O
), with the proba-
bility of failure at O(1/nc ). The proof in that paper can be generalised to the case of
σ ≥
n, to give a runtime of O(σ 2 log n). It is not known what happens for lower lev-
els of noise, though it is shown that in the absence of noise, PCEA solves OneMax in
O(n(log n)2).
, giving a runtime of O(cn
(cid:10)
n log n
(cid:10)
2
log n
(cid:9)√
√
√
(cid:9)
Evolutionary Computation Volume 31, Number 3
265
Aishwaryaprajna and J.E. Rowe
Table 1: Function evaluation budgets allowed for noisy OneMax experiments with dif-
ferent noise levels.
σ
budget
σ
budget
1
38392
6
64079
2
41066
7
70736
3
44477
8
79034
4
50728
9
86078
5
56851
10
93638
4
Experiments—Simple Noisy Problems
4.1 Noisy ONEMAX
We investigate the performance of the algorithms described above, in solving the noisy
OneMax problem. In the literature, some theoretical proofs exist for the expected run-
time of specific algorithms on solving the noisy OneMax problem with additive poste-
rior Gaussian noise (Prügel-Bennett et al., 2015; Dang and Lehre, 2015; Akimoto et al.,
2015; Friedrich et al., 2017; Lucas et al., 2017; Qian et al., 2018; Dang-Nhu et al., 2018;
Doerr and Sutton, 2019; Doerr, 2020). We are interested in the algorithms’ performances
given a reasonable but fixed runtime budget across a wide range of noise levels, from
σ = 0 up to σ =
n.
√
To address the question of what constitutes a reasonable budget, we compared the
known theoretical results of our algorithms on noisy OneMax. PCEA has the lowest
proven upper bound on its runtime, compared with the other algorithms for which re-
sults exist. We therefore allowed each algorithm to have twice the number of fitness
evaluations that PCEA requires (on average) to find the optimum, as a reasonable bud-
get. The function evaluation budgets calculated in this way are given in Table 1.
√
The population size for the PCEA is taken to be 10
n log n according to the theoret-
ical proofs and empirical study by Prügel-Bennett et al. (2015). According to the proofs
by Dang and Lehre (2015), the population size λ = σ 2 log n is chosen for the mutation-
population algorithm. According to the paper by Friedrich et al. (2015), the parameter
K = 7σ 2
n log n is considered for cGA. In presence of additive posterior noise, PBIL
and UMDA have not yet been studied much. For PBIL, the population size is taken as
λ = 10n (following the theoretical requirement of Wu et al., 2017). From these, we select
the best μ = λ/2 individuals. In case of UMDA, the total number of generated candi-
n log n, so that the effective population
dates in a particular generation is chosen as 20
size is the same as for PCEA. All these parameter settings are retained for all of our
experiments in simple and constrained noisy combinatorial optimisation problems.
√
√
Figure 1 illustrates a comparison of all of the considered algorithms while solv-
ing the noisy OneMax problem for problem size n = 100. Different levels of Gaussian
additive noise with mean 0 and standard deviation σ = 1 to 10 are considered in this ex-
periment. It can be seen that PCEA and UMDA are resistant to these noise levels as they
are capable of finding the global optimum within the given budget. The runtimes for
these two algorithms are shown in Figure 2. However, (1 + 1)-EA, mutation-population
algorithm, PBIL, and cGA are not able to cope with even these small levels of noise
within the given fixed budget of function evaluations. For these experiments, we run
the algorithms until the population converges (which they will, since we do not handle
probability margins). The Mann–Whitney U-test is performed on the samples of best
results achieved and the runtimes of the algorithms, with the null hypothesis that they
are from distributions with equal medians. For each data point, the null hypothesis is
rejected at a 5% significance level.
266
Evolutionary Computation Volume 31, Number 3
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
/
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
EAs and EDAs for Solving Noisy Combinatorial Problems
Figure 1: Comparison of the algorithms while solving the noisy OneMax for different
noise levels.
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
.
/
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Figure 2: Runtime comparison of UMDA and PCEA for noisy OneMax.
4.2 Noisy WEIGHTEDLINEAR Problem
Maximising the WeightedLinear problem, as defined in Section 2, has only one global
optimum, the sum of all the weights. The OneMax problem is a special case of the
WeightedLinear problem when all the weights are units. However, optimising the
WeightedLinear problem is difficult as the bits with heavier weights get optimised
with a higher preference than the bits with lower weights.
The plot in Figure 3 illustrates the performance comparison of all of the consid-
ered algorithms while solving the noisy WeightedLinear problem for the problem size
n = 100. Random problem instances were studied with 100 randomly chosen weights
between 1 and 100. The results for a typical problem are shown in Figure 3 with aver-
ages over 100 runs. The standard deviation of the Gaussian noise is shown as multiples
of the square root of the sum of the weights. The function evaluation budget allowed
Evolutionary Computation Volume 31, Number 3
267
Aishwaryaprajna and J.E. Rowe
Figure 3: Comparison of algorithms while solving noisy WeightedLinear.
Table 2: Function evaluation budgets allowed for noisy WeightedLinear experiments.
σ
budget
σ
budget
1
47096
6
49954
2
46801
7
50876
3
47704
8
51429
4
48350
9
52794
5
48682
10
53310
to each of the algorithms are fixed at twice the average runtime of PCEA at each noise
level (see Table 2).
As evident from Figure 3, the curves of PCEA and UMDA are coincident, showing
that they can cope with the noise well and are resistant up to these levels of noise. The
runtime of UMDA and PCEA are plotted in Figure 4. However, the performance of
the (1 + 1)-EA and mutation-population algorithm worsen with increasing noise. Even
with relatively small noise levels, the cGA and PBIL are not able to solve the problem
within twice the runtime of PCEA.
It is evident from the empirical results of these simple noisy problems that uniform
crossover-based PCEA and UMDA can cope with noise significantly better than the
other algorithms. At this point, it is interesting to note that, UMDA employs a mecha-
nism similar to genepool crossover, where at each bit position, the offspring bit is obtained
by recombination of that bit across the whole parent population. It is hypothesised that
these two algorithms are therefore highly similar in operation.
5
Experiments—Noisy Combinatorial Problems
5.1 Noisy SUBSETSUM
Given the success of UMDA and PCEA on the noisy toy problems, and the failure of the
others to cope with even modest levels of noise, we now move to the second stage of
the study considering only UMDA and PCEA.
268
Evolutionary Computation Volume 31, Number 3
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
/
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
EAs and EDAs for Solving Noisy Combinatorial Problems
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
/
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Figure 4: Runtime comparison of UMDA and PCEA for noisy WeightedLinear.
For the noisy SubsetSum problem, a range of problem sizes is considered with 50,
100, 150, and 200 weights, each lying between 1 and 100, and chosen uniformly at ran-
dom. Corresponding to each problem size, 10 different problems are considered. The
target θ is considered to be two-thirds of the sum of all the weights in the set. The ad-
ditive Gaussian noise considered in the SubsetSum problem is centered at zero and is
considered to have standard deviation of integral multiples of the mean of the weights,
viz., 5 × mean(W ), 10 × mean(W ), 15 × mean(W ), and 20 × mean(W ).
The NoisySubsetSum problem being a minimisation problem, if we obtain the (non-
noisy) fitness value of zero, we obtain the global optimum. Both the algorithms are able
to find the global optimum for these problems and their corresponding noise levels.
We therefore plot the runtime (averaged over 100 runs) to find the optimum against
the standard deviation of the noise (see Figure 5). Using the Mann–Whitney U-test, it is
observed that UMDA has the better runtime.
5.2 Noisy KNAPSACK (Version 1)
For the first version of the noisy Knapsack problem, uncorrelated (see Figures 6 and 7)
and strongly correlated (see Figures 10 and 11) problem instances have been studied.
For the uncorrelated instances, 50, 100, 150, and 200 weights (randomly chosen between
1 and 100) with associated profits (in the same range) are considered. For each problem
size, 10 different problems are considered. The maximum capacity of the knapsack is
taken to be two-thirds of the sum of all the weights considered. For the strongly corre-
lated instances, 10 problems each of sizes 50 and 100, described in Pisinger (2005), are
considered.
When noise is added, neither algorithm finds the optimal solution, so we record
the best solution found (as assessed by non-noisy fitness function). For each problem
instance, we plot (in Figures 6 and 10) the best solution found (averaged over 100 runs)
as a fraction of the best solution ever encountered for that problem instance. This en-
ables us to make meaningful comparisons between problem instances. The best known
Evolutionary Computation Volume 31, Number 3
269
Aishwaryaprajna and J.E. Rowe
Figure 5: Runtime comparison of UMDA (circles) and PCEA (triangles) while solving
instances of the noisy SubsetSum problem.
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
/
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Figure 6: Solution quality of UMDA (circles) and PCEA (triangles) while solving in-
stances of NoisyKnapsackV1.
solution for each problem instance has a scaled fitness value of 1. Figures 7 and 11 show
the time taken (on average) to locate the best found solution in each case. We can ob-
serve in Figures 6 and 7, that both the algorithms can find good, though not optimal
solutions, for NoisyKnapsackV1 with significant levels of noise.
Observations from the Mann–Whitney U-test show that UMDA is slightly better
than PCEA with these parameter settings for larger noise levels.
270
Evolutionary Computation Volume 31, Number 3
EAs and EDAs for Solving Noisy Combinatorial Problems
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
.
/
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Figure 7: Comparison of runtime of UMDA (circles) and PCEA (triangles) while solving
the NoisyKnapsackV1.
5.3 Noisy KNAPSACK (Version 2)
When the measurements of the weights is uncertain, as well as the profits, this creates a
more complex noise model for the Knapsack problem. In the first stage, the total weight
of the proposed solution is compared against the capacity, and this is done with added
noise. Hence it may be thought that the proposed solution is feasible when in fact it is
not. If it is considered feasible, then the benefit (total profit) is calculated, again with
added noise.The same problem instances are considered as in the previous version of
the Knapsack problem.
Figures 8 and 10 depict how the best (non-noisy) solution varies for different prob-
lem sizes. This value is scaled with respect to the best value found when there is no
noise. The Mann–Whitney U-test shows that the best solution achieved and correspond-
ing runtime of UMDA is better than PCEA in these particular parameter settings. The
runtime required to find these values is shown in Figures 9 and 11, and we see that
UMDA finds its best solution considerably faster than PCEA.
5.4 Noisy CONSTRAINEDSETCOVER and PENALTYSETCOVER
The ConstrainedSetCover problem is solved by initially finding the feasible solutions
and then minimising the number of the selected sets. This lexicographic ordering is
achieved in the selection mechanism of the considered algorithms.
In PCEA, the child with least uncovered elements is selected. When both of the
children have the same number of uncovered elements, the child with the minimum
number of sets goes to the next population. In UMDA, the sorting of the population is
based on the above mentioned lexicographic ordering. We consider margin handling in
UMDA for all the following experiments in single objective-optimisation.
The alternative PenaltySetCover problem handles the constraint within the
penalty function, hence creating a single objective. For both versions of the noisy
Evolutionary Computation Volume 31, Number 3
271
Aishwaryaprajna and J.E. Rowe
Figure 8: Solution quality of UMDA (circles) and PCEA (triangles) while solving the
NoisyKnapsackV2.
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
/
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Figure 9: Comparison of runtime of UMDA (circles) and PCEA (triangles) while solving
the NoisyKnapsackV2.
SetCover problem, a range of 40 problem instances (10 for each problem size) are run
with 100 elements and 50, 100, 150, and 200 subsets are available to cover those elements.
The problems are created by randomly generating subsets, where the probability of
272
Evolutionary Computation Volume 31, Number 3
EAs and EDAs for Solving Noisy Combinatorial Problems
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
Figure 10: Solution quality of UMDA (circles) and PCEA (triangles) while solving the
NoisyKnapsack for strongly correlated problem instances.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
/
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Figure 11: Comparison of runtime of UMDA (circles) and PCEA (triangles) while solving
the NoisyKnapsack for strongly correlated problem instances.
Evolutionary Computation Volume 31, Number 3
273
Aishwaryaprajna and J.E. Rowe
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
/
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Figure 12: Solution quality of UMDA (circles) and PCEA (triangles) while solving the
ConstrainedSetCover.
including any element in any subset is p. This is set so that the probability of there
being cover is large:
Therefore, we take:
(1 − (1 − p)n)m = 1 − δ.
p = 1 − (1 − (1 − δ)1/m)1/n.
We have chosen δ = 0.001. All the algorithms are run until 50,000 function evaluations
are reached. An average of 30 runs are reported. Figure 12 reports the best feasible so-
lution found in the fixed budget of function evaluations. As evident from the figure,
neither of the algorithms can handle noise well. The noisy feasibility check significantly
worsens the optimum found even for small standard deviations of noise.
The parameters considered for solving the PenaltySetCover are chosen the same
as the ConstrainedSetCover. For each problem, we plot the best feasible solution
found so far in the given function evaluation budget and the runtime in Figures 13 and
14. It is interesting that both the algorithms can solve the noisy instances in a scalable
manner, with UMDA typically producing better quality solutions.
6 Noisy Combinatorial Multiobjective Problems
In this section, we empirically examine the performances of several evolutionary algo-
rithms on noisy combinatorial multiobjective problems. Much of the previous work on
multiobjective optimisation (especially in the context of noise) has concerned continu-
ous problems (Goh et al., 2010; Shim et al., 2013; Fieldsend and Everson, 2015; Falcón-
Cardona and Coello, 2020). In this article, we focus on discrete problems, but with
additive (posterior) Gaussian noise.
274
Evolutionary Computation Volume 31, Number 3
EAs and EDAs for Solving Noisy Combinatorial Problems
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
Figure 13: Best solution found in stipulated budget of function evaluations by UMDA
(circles) and PCEA (triangles) for NoisyPenaltySetCover.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
.
/
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Figure 14: Runtime of UMDA (circles) and PCEA (triangles) for best solution found
while solving NoisyPenaltySetCover.
Evolutionary Computation Volume 31, Number 3
275
Aishwaryaprajna and J.E. Rowe
A noisy multiobjective combinatorial problem in the search space of binary strings
may be defined as follows,
f (x) = (f1(x) + N (0, σ ), f2(x) + N (0, σ ), . . . , fk (x) + N (0, σ )),
where, x ∈ {0, 1}n is a candidate solution. The objectives f1(x), f2(x), . . . , fk (x) are con-
flicting in nature, so there does not necessarily exist an optimal solution that will
minimise all the objectives simultaneously. Instead, there exists a set of nondominat-
ing solutions known as the Pareto optimal solution set where none of the objectives may
be improved without worsening at least one of the other objectives. In the context of
noisy multiobjective optimisation, the goal is to find the set of Pareto optimal solutions,
as defined in the absence of noise; however, the challenge is that each time a compari-
son is made, noise is applied. This is particularly problematic for algorithms that make
use of an archive of nondominated solutions, as it is easy for a solution to be incorrectly
placed in the archive due to the noise.
In order to assess how successfully we have approximated the true Pareto optimal
set, we measure the spread of a set of nondominated solutions on the basis of the fre-
quently used hypervolume performance indicator (Zitzler and Thiele, 1998). Where we
seek to minimise each objective, this is a measure of the area (or volume) of the region
bounded below by a set of candidate solutions simultaneously and bounded above by
a reference point r in the objective space. The reference point r is chosen to be the max-
imum value each objective function can attain in each corresponding dimension of the
objective space; that is, r = (max f1, max f2, . . . , max fk ). Conversely, for maximisation
problems, we take the volume between the candidate set and a lower bounding refer-
ence point (in the case of non-negative objectives, it is common to take the origin as the
reference point). We use hypervolume of the population as an indicator of the spread
of the nondominated solutions in each generation of the considered algorithms.
In this article, we have studied two noisy multiobjective problems. The first is based
on the toy benchmark problem Counting Ones Counting Zeroes (COCZ), in which the
first objective function counts the number of ones in a string, and the second objective
function counts the number of ones in the first m bits and the number of zeroes in the
remainder. We seek to maximise both objectives.
NoisyCOCZ(x) =
(cid:11)
n(cid:2)
i=1
xi + N (0, σ ),
m(cid:2)
xi +
n(cid:2)
i=1
i=m+1
(cid:12)
(1 − xi ) + N (0, σ )
.
The Pareto optimal front consists of strings of the form 1m∗(n−m).
The second problem is a multiobjective version of SetCover problem, with the ob-
jective function and the constraint as defined in ConstrainedSetCover as the two ob-
jective functions. These objectives are conflicting in nature. The first objective minimises
the number of sets required to cover all the m elements of the target set, and the second
objective minimises the number of uncovered elements. The noisy version of the mul-
tiobjective SetCover problem is defined as follows,
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
/
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
NoisyMultiobjectiveSetCover(x) =
(cid:11)
n(cid:2)
j =1
xj + N (0, σ ),
(cid:2)
⎡
⎣
n(cid:2)
i
j =1
Recall that [expr] equals 1 when expr is true and 0 otherwise.
⎤
(cid:12)
⎦+ N (0, σ )
.
aij xj = 0
276
Evolutionary Computation Volume 31, Number 3
EAs and EDAs for Solving Noisy Combinatorial Problems
7 Algorithms Chosen for Noisy Multiobjective Combinatorial
Problems
7.1
Simple Evolutionary Multiobjective Optimiser (SEMO)
SEMO (Laumanns et al., 2004) is one of the simplest evolutionary algorithms designed
for multiobjective optimisation in discrete search space. To the best of our knowledge,
it has not previously been used to solve noisy problems. SEMO is a simple population-
based algorithm using one-bit mutation, and a variable population size (representing
the current nondominated solutions found). The algorithm starts with adding an initial
solution x ∈ {0, 1}n chosen uniformly at random to the population P . Then a solution y is
chosen randomly from P and mutated with a one-bit flip to obtain y(cid:7)
is dominated
by anything in P it is discarded. Otherwise it is added to P and all the solutions that
y(cid:7)
dominates in P are discarded. Then a new y is chosen from P and the process is
repeated. One of the great challenges SEMO will face due to noisy dominance relations
is that, often good solutions will be discarded and bad solutions will be retained in P .
. If y(cid:7)
7.2 Nondominated Sorting Genetic Algorithm—II (NSGA-II)
NSGA-II, by Deb et al. (2002), sorts the population into nondominated fronts in each
generation. Based on nondominated sorting and using a crowding heuristic to break
ties, the best half of individuals become the parent population of the next generation.
In case of noisy function evaluations, nondominated sorting will be affected and worse
solutions will appear in better nondominated fronts. We use the same algorithm struc-
ture as defined in Deb et al. (2002) except considering noisy function evaluations during
the selection process.
7.3 Variants of Multiobjective Univariate Marginal Distribution Algorithm
(moUMDA)
From our experiments in noisy single-objective combinatorial problems, UMDA and
PCEA show significantly better performance in handling noise compared with the
other algorithms we tried, with UMDA generally producing better quality solutions.
From these results, we hypothesise that a multiobjective version of UMDA (denoted
moUMDA) may be able to handle large levels of noise in noisy combinatorial multi-
objective problems if proper diversification mechanisms are employed. In order to in-
vestigate this, we have considered several versions of moUMDA in our analysis with
different diversification techniques.
Pelikan et al. (2005) introduced a version of UMDA to address multiobjective
problems which used nondominated sorting in the selection mechanism. They also
Evolutionary Computation Volume 31, Number 3
277
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
.
/
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Aishwaryaprajna and J.E. Rowe
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
/
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
experimented with clustering methods, to help the algorithm generate solutions across
the Pareto front. We have followed this idea, and studied several versions of UMDA
adapted for multiobjective problems. Where nondominated sorting and crowding are
used for selection, these are implemented identically to NSGA-II. We also consider mak-
ing use of an archive, and in using hypervolume as a criterion in selection:
moUMDA without duplicates: Uses nondominated sorting (with crowding to break
ties) for selection. Maintains diversity by disallowing duplicates when generating
the population. See Algorithm 2.
moUMDA with clustering: Uses nondominated sorting (with crowding to break
ties) for selection. Clusters the selected population members (using either K-
means or Hierarchical Agglomeration), and produces a frequency vector for each
cluster. Generates next population from these, in proportion to the number of
items within each cluster. See Algorithm 3.
278
Evolutionary Computation Volume 31, Number 3
EAs and EDAs for Solving Noisy Combinatorial Problems
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
.
/
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
moUMDA with Pareto archive: Maintains an archive of nondominated solutions
and uses this to generate the frequency vector for the next population. Uses
nondominated sorting (with crowding to break ties) for selection, and updates
the archive with the selected items. See Algorithm 4.
moUMDA with hypervolume comparison operator: Uses binary tournament se-
lection, comparing solutions initially by Pareto dominance. If neither dominates
the other, then select the one with the better hypervolume indicator value. See
Algorithm 5.
8
Experiments—Noisy Multiobjective Problems
Following the same strategy as for single objective problems, we initially choose a wide
range of evolutionary multiobjective algorithms to compare their performances on a toy
problem: noisy CountingOnesCountingZeroes (COCZ). The algorithms considered
for solving COCZ consists of SEMO, NSGA-II and several versions of multiobjective
UMDA (moUMDA) as described above. Depending on their performances on this prob-
lem, we selected a smaller set of the better performing algorithms for the multiobjective
noisy SetCover problem.
Evolutionary Computation Volume 31, Number 3
279
Aishwaryaprajna and J.E. Rowe
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
.
/
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Figure 15: Comparison of the hypervolume of population while solving the noisy COCZ
with n = 30, m = 15.
Some recent studies claim that multiobjective evolutionary approaches are useful
in solving single objective optimisation problems (Segura et al., 2016). For example,
the multiobjective version of SetCover could enable us to find good solutions to the
original single-objective version (by looking at solutions generated which do not violate
the constraints). Here, we consider whether this approach is also helpful in the context
of noise.
√
8.1 Noisy COUNTINGONESCOUNTINGZEROES (COCZ)
In this subsection, we solve a toy multiobjective problem, the noisy COCZ with n =
30, m = 15 and with additive Gaussian noise centered at zero and having standard
deviations σ = 0, 1, 3, 5, 7, 9, 11, 13, and 15. We set the parameter μ = λ/2, where
λ = 20
n log n for all the versions of moUMDA. For NSGAII, the parent population
n log n. All the algorithms are run for 50,000 function evaluations and
size is set as 10
the mean of 30 runs are reported. The best hypervolume of the population found so far
in the fixed budget of function evaluations are reported in Figure 15. The Pareto op-
timal front would contain 215 elements and the best possible hypervolume is 780. We
have used the dimension-sweep algorithm and the source code by Fonseca et al. (2006)
for hypervolume calculation in the experiments.
√
The results shown in Figure 15 show that SEMO is the worst performing algorithm,
even when there is no noise, and the performance degrades slightly as noise is increased.
The Pareto Archive algorithm (PAmoUMDA) is the next worst. Although it does not
degrade too much with added noise, it is still clearly worse than the other algorithms.
280
Evolutionary Computation Volume 31, Number 3
EAs and EDAs for Solving Noisy Combinatorial Problems
The remaining algorithms have rather similar performance, but we can still dis-
tinguish different behaviours by looking at the zoomed in section of the plot in Fig-
ure 15. The version of moUMDA that uses the hypervolume comparison operator
(moUMDAHCO) performs very well when there is little or no noise. However, its per-
formance degrades considerably as the level of noise increases. The same is true for
NSGAII. When the noise reaches a standard deviation of σ = 15, these two algorithms
are the worst of the remaining ones.
The plain moUMDA and the version forbidding duplicates in the population both
have the curious property that their performance improves with the presence of low
levels of noise, and then degrades at higher levels of noise. We speculate that low levels
of noise allow for much more diversity in the populations. At high levels of noise (σ =
15) they are the best performing algorithms, along with the two versions of moUMDA
that use clustering (moUMDA-Kmeans and moUMDA-HAC). The moUMDA with no
duplicates is marginally the best overall at this level of noise.
8.2 Noisy Multiobjective SETCOVER
In this section, we compare the performance of three of our multiobjective algorithms,
viz., NSGA-II, moUMDA with no duplicates allowed, and moUMDA employing K-
means clustering, on the noisy multiobjective SetCover problem. We have chosen these
algorithms based on their behaviours on the COCZ. These were amongst the best al-
gorithms we tried on that problem. There being little to distinguish the two different
clustering methods, we have chosen to test just one of these (K-means clustering). We
have selected the “no duplicates” version of moUMDA, as this gave a small advantage
over the plain moUMDA. And we have kept NSGAII as this is a standard algorithm for
any multiobjective problem.
All the algorithms are run for 50,000 function evaluations. The best hypervolume
of the population obtained in the fixed function evaluation budget for each of 30 runs
is shown in Figure 16. We observe that the clustering algorithm, moUMDA-Kmeans,
handles high levels of noise significantly better than other algorithms. It is evident that
the performance of NSGA-II becomes worse as the standard deviation of noise increases
and the problem size increases and indeed is the worst of the three algorithms on this
problem.
We also consider the multiobjective formulation of noisy SetCover as a means to
solve the standard single objective problem. To this end, we consider the quality of the
best feasible solutions found by each algorithm, averaged over the 30 runs. The results
are plotted in Figure 17. Again, the two versions of moUMDA perform better than NS-
GAII. A comparison with Figure 13 shows that this approach can indeed produce better
quality results than the single objective formulation.
9 Conclusion
We have empirically studied a range of evolutionary algorithms on a set of noisy prob-
lems. The (1 + 1)-EA, as expected, fails to cope with any degree of posterior noise.
Interestingly, some algorithms (the mutation-population algorithm and cGA), where
there is a theoretical polynomial runtime for noisy OneMax, fail to be useful in practice
compared with some other algorithms. PBIL performs somewhat similar to cGA. The
Paired-Crossover Evolutionary algorithm handles noise well on both the simple test
problems, and on the noisy combinatorial problems we have tried. Interestingly, UMDA
also handles these cases well, with even a slightly better performance than PCEA. This
may be due to the fact that UMDA has a strong selection method (truncation selection)
Evolutionary Computation Volume 31, Number 3
281
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
/
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Aishwaryaprajna and J.E. Rowe
Figure 16: Best hypervolume of population obtained for MultiobjectiveSetCover.
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
/
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Figure 17: Best feasible solution found while solving the noisy MultiobjectiveSet-
Cover.
than PCEA (which uses a tournament on pairs of offspring). Of course, parameter val-
ues on each could be tweaked to produce slightly different results; our key finding is
that these are the only algorithms we have tried that seem remotely practical for such
problems. It seems likely that UMDA’s performance is more due to its relationship with
282
Evolutionary Computation Volume 31, Number 3
EAs and EDAs for Solving Noisy Combinatorial Problems
crossover algorithms (such as the genepool crossover), rather than considered as an
EDA (such as PBIL).
We are not aware of any previously published results on noisy combinatorial mul-
tiobjective problems. We carefully selected a set of multiobjective algorithms on the ba-
sis of the performance on noisy COCZ and tested them on the noisy multiobjective
SetCover. We observe that multiobjective UMDA with a simple diversity mechanism
that allows no duplicate solutions in the population is effective at solving the noisy Set-
Cover problem in both constrained and multiobjective forms. UMDA can also benefit
from using a clustering approach when dealing with noisy multiobjective problems.
References
Aishwaryaprajna, and Rowe, J. E. (2019). Noisy combinatorial optimisation by evolutionary al-
gorithms. In Proceedings of the Genetic and Evolutionary Computation Conference Companion
(GECCO), pp. 139–140.
Akimoto, Y., Astete-Morales, S., and Teytaud, O. (2015). Analysis of runtime of optimization algo-
rithms for noisy functions over discrete codomains. Theoretical Computer Science, 605:42–50.
10.1016/j.tcs.2015.04.008
Anil, G., and Wiegand, R. P. (2009). Black-box search by elimination of fitness functions. In Pro-
ceedings of the Tenth ACM SIGEVO Workshop on Foundations of Genetic Algorithms, pp. 67–78.
Baluja, S. (1994). Population-based incremental learning: A method for integrating genetic search based
function optimization and competitive learning. Technical Report. Department of Computer Sci-
ence, Carnegie-Mellon University, Pittsburgh, Pennsylvania.
Bianchi, L., Dorigo, M., Gambardella, L. M., and Gutjahr, W. J. (2009). A survey on metaheuris-
tics for stochastic combinatorial optimization. Natural Computing, 8(2):239–287. 10.1007/
s11047-008-9098-4
Carraway, R. L., Schmidt, R. L., and Weatherford, L. R. (1993). An algorithm for maximizing
target achievement in the stochastic knapsack problem with normal returns. Naval Research
Logistics, 40(2):161–173. 10.1002/nav.3220400203
Dang, D., and Lehre, P. K. (2016). Runtime analysis of non-elitist populations: From classical op-
timisation to partial information. Algorithmica, 75(3):428–461. 10.1007/s00453-015-0103-x
Dang, D.-C., and Lehre, P. K. (2015). Efficient optimisation of noisy fitness functions with
population-based evolutionary algorithms. In Proceedings of the 2015 ACM Conference on
Foundations of Genetic Algorithms XIII, pp. 62–68.
Dang-Nhu, R., Dardinier, T., Doerr, B., Izacard, G., and Nogneng, D. (2018). A new analysis
method for evolutionary optimization of dynamic and noisy objective functions. In Proceed-
ings of the Genetic and Evolutionary Computation Conference, pp. 1467–1474.
Deb, K., Pratap, A., Agarwal, S., and Meyarivan, T. (2002). A fast and elitist multiobjective genetic
algorithm: NSGA-II. IEEE Transactions on Evolutionary Computation, 6(2):182–197. 10.1109/
4235.996017
Doerr, B. (2020). Exponential upper bounds for the runtime of randomized search heuristics.
In International Conference on Parallel Problem Solving from Nature, pp. 619–633. Vol. 12270.
10.1007/978-3-030-58115-2_43
Doerr, B., and Sutton, A. M. (2019). When resampling to cope with noise, use median, not mean.
In Proceedings of the Genetic and Evolutionary Computation Conference, pp. 242–248.
Droste, S. (2004). Analysis of the (1 + 1) ea for a noisy onemax. In Genetic and Evolutionary Com-
putation Conference, pp. 1088–1099.
Evolutionary Computation Volume 31, Number 3
283
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
.
/
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Aishwaryaprajna and J.E. Rowe
Falcón-Cardona, J. G., and Coello, C. A. C. (2020). Indicator-based multi-objective evolutionary
algorithms: A comprehensive survey. ACM Computing Surveys, 53(2):1–35.
Fieldsend, J. E., and Everson, R. M. (2015). The rolling tide evolutionary algorithm: A multiobjec-
tive optimizer for noisy optimization problems. IEEE Transactions on Evolutionary Computa-
tion, 19(1):103–117. 10.1109/TEVC.2014.2304415
Fonseca, C. M., Paquete, L., and López-Ibánez, M. (2006). An improved dimension-sweep algo-
rithm for the hypervolume indicator. In IEEE International Conference on Evolutionary Compu-
tation, pp. 1157–1163.
Fortz, B., Labbé, M., Louveaux, F., and Poss, M. (2013). Stochastic binary problems with sim-
ple penalties for capacity constraints violations. Mathematical Programming, 138(1):199–221.
10.1007/s10107-012-0520-4
Friedrich, T., Kötzing, T., Krejca, M. S., and Sutton, A. M. (2015). The benefit of recombination in
noisy evolutionary search. In K. Elbassioni and K. Makino (Eds.), Algorithms and computation,
pp. 140–150. Springer Berlin Heidelberg.
Friedrich, T., Kötzing, T., Krejca, M. S., and Sutton, A. M. (2017). The compact genetic algorithm
is efficient under extreme Gaussian noise. IEEE Transactions on Evolutionary Computation,
21(3):477–490.
Gießen, C., and Kötzing, T. (2016). Robustness of populations in stochastic environments. Algo-
rithmica, 75(3):462–489.
Goh, C. K., Tan, K. C., Cheong, C. Y., and Ong, Y.-S. (2010). An investigation on noise-induced fea-
tures in robust evolutionary multi-objective optimization. Expert Systems with Applications,
37(8):5960–5980. 10.1016/j.eswa.2010.02.008
Goldberg, D. E., Deb, K., and Clark, J. H. (1991). Genetic algorithms, noise, and the sizing of
populations. Urbana, 51:61801.
Harik, G. R., Lobo, F. G., and Goldberg, D. E. (1999). The compact genetic algorithm. IEEE Trans-
actions on Evolutionary Computation, 3(4):287–297. 10.1109/4235.797971
Henig, M. I. (1990). Risk criteria in a stochastic knapsack problem. Operations Research, 38(5):820–
825. 10.1287/opre.38.5.820
Laumanns, M., Thiele, L., and Zitzler, E. (2004). Running time analysis of multiobjective evolu-
tionary algorithms on pseudo-Boolean functions. IEEE Transactions on Evolutionary Compu-
tation, 8(2):170–182. 10.1109/TEVC.2004.823470
Lucas, S. M., Liu, J., and Pérez-Liébana, D. (2017). Efficient noisy optimisation with the multi-
sample and sliding window compact genetic algorithms. In IEEE Symposium Series on Com-
putational Intelligence, pp. 1–8.
Mühlenbein, H. (1997). The equation for response to selection and its use for prediction. Evolu-
tionary Computation, 5(3):303–346.
Pelikan, M., Sastry, K., and Goldberg, D. E. (2005). Multiobjective hboa, clustering, and scalability.
In Proceedings of the 7th Annual Conference on Genetic and Evolutionary Computation, pp. 663–
670.
Pisinger, D. (2005). Where are the hard knapsack problems? Computers & Operations Research,
32(9):2271–2284.
Prügel-Bennett, A., Rowe, J., and Shapiro, J. (2015). Run-time analysis of population-based evolu-
tionary algorithm in noisy environments. In Proceedings of the 2015 ACM Conference on Foun-
dations of Genetic Algorithms XIII, pp. 69–75.
284
Evolutionary Computation Volume 31, Number 3
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
.
/
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
EAs and EDAs for Solving Noisy Combinatorial Problems
Qian, C., Yu, Y., Tang, K., Jin, Y., Yao, X., and Zhou, Z.-H. (2018). On the effectiveness of sampling
for evolutionary optimization in noisy environments. Evolutionary Computation, 26(2):237–
267. 10.1162/evco_a_00201
Rattray, M., and Shapiro, J. (1997). Noisy fitness evaluation in genetic algorithms and the dy-
namics of learning. In R. K. Belew and M. D. Vose (Eds.), Foundations of Genetic Algorithms,
pp. 117–139. Morgan Kaufmann.
Ross, K. W., and Tsang, D. H. (1989). The stochastic knapsack problem. IEEE Transactions on Com-
munications, 37(7):740–747. 10.1109/26.31166
Rowe, J. E., and Aishwaryaprajna (2019). The benefits and limitations of voting mechanisms in
evolutionary optimisation. In Proceedings of the 15th ACM/SIGEVO Conference on Foundations
of Genetic Algorithms, pp. 34–42.
Segura, C., Coello, C. A. C., Miranda, G., and León, C. (2016). Using multi-objective evolution-
ary algorithms for single-objective constrained and unconstrained optimization. Annals of
Operations Research, 240(1):217–250. 10.1007/s10479-015-2017-z
Shim, V. A., Tan, K. C., Chia, J. Y., and Mamun, A. A. (2013). Multiobjective optimization with
estimation of distribution algorithm in a noisy environment. Evolutionary Computation,
21(1):149–177. 10.1162/EVCO_a_00066
Sniedovich, M. (1980). Preference order stochastic knapsack problems: Methodological issues.
Journal of the Operational Research Society, 31(11):1025–1032. 10.1057/jors.1980.189
Steinberg, E., and Parks, M. (1979). A preference order dynamic program for a knapsack problem
with stochastic rewards. Journal of the Operational Research Society, 30(2):141–147. 10.1057/
jors.1979.27
Witt, C. (2017). Upper bounds on the runtime of the univariate marginal distribution algorithm
on onemax. In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO),
pp. 1415–1422.
Witt, C. (2021). On crossing fitness valleys with majority-vote crossover and estimation-of-
distribution algorithms. In Proceedings of the 16th ACM/SIGEVO Conference on Foundations
of Genetic Algorithms, pp. 1–15.
Wu, Z., Kolonko, M., and Möhring, R. H. (2017). Stochastic runtime analysis of the cross-
entropy algorithm. IEEE Transactions on Evolutionary Computation, 21(4):616–628. 10.1109/
TEVC.2017.2667713
Zitzler, E., and Thiele, L. (1998). Multiobjective optimization using evolutionary algorithms—A
comparative case study. In International Conference on Parallel Problem Solving from Nature,
pp. 292–301. Vol. 1498.
Evolutionary Computation Volume 31, Number 3
285
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
/
e
d
u
e
v
c
o
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
1
3
2
5
9
2
1
5
5
6
3
9
e
v
c
o
_
a
_
0
0
3
2
0
p
d
/
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3