Preprint
Article

This version is not peer-reviewed.

RSG, a Method for Pareto Front Approximation and Reference Set Generation

A peer-reviewed article of this preprint also exists.

Submitted:

03 April 2025

Posted:

03 April 2025

You are already at the latest version

Abstract
In this paper, we address the problem of obtaining bias-free and complete finite size approximations of the solution sets (Pareto fronts) of multi-objective optimization problems (MOPs). Such approximations are, in particular, required for the fair usage of distance-based performance indicators, which are frequently used in evolutionary multi-objective optimization (EMO). If the Pareto front approximations are biased or incomplete, the use of these performance indicators can lead to misleading or false information. To address this issue, we propose the Reference Set Generator (RSG), which can, in principle, be applied to Pareto fronts of any shape and dimension. We finally demonstrate the strength of the novel approach on several benchmark problems.
Keywords: 
;  ;  ;  

1. Introduction

Multi-objective optimization has become an integral part of the decision-making for many real-world problems. In a multi-objective optimization problem (MOP), one is faced with the issue of concurrently optimizing k individual objectives. The set of optimal solutions is called the Pareto set. The image of the Pareto set is called the Pareto front. The latter set is, in many cases, most important for the decision maker (DM) since it provides him/her with an overview of the optimal performances for his/her project. What makes MOPs hard to deal with is that one can expect that both Pareto set and front form – at least locally and under certain assumptions on the model – objects of dimension k 1 ([1]). For the numerical treatment of MOPs, specialized evolutionary algorithms, called multi-objective evolutionary algorithms (MOEAs), have caught the interest of many researchers and practitioners during the last three decades ([2]). MOEAs are population-based and hence allow for the obtaining of a finite approximation of the entire solution set in one run of the algorithm. For the performance assessment of the outcome sets, several different indicators have been proposed so far (e.g., [3,4,5,6,7,8]). Some of these performance indicators are distance-based and require a “suitable” finite-size representation of the Pareto front. While until now, a vast variety of different MOEAs has been proposed and analyzed, it is fair to say that the generation of suitable reference sets has played a rather minor role in the evolutionary multi-objective optimization (EMO) community. It is evident that such reference sets should be complete. Further, as we show in this work, a biased representation can lead to misleading or even false information.
To fill this gap, we propose in this work the Reference Set Generator (RSG). The main steps of RSG are as follows: (i) a first approximation A y of the Pareto front is either taken or generated. This set can be, in principle, of arbitrary size, and the elements can be non-uniformly distributed along the Pareto front (i.e., biased). However, all of these elements have to be "close enough" to the set of interest. In order to obtain a bias-free approximation, (ii) component detection and (iii) a filling step is applied to A y . Finally, (iv) a reduction step is applied. The RSG is applicable to MOPs with Pareto fronts of, in principle, any shape and dimension. Further, the resulting reference set is of any desired magnitude. We will show the strength of the novel method on several benchmark problems.
The remainder of this paper is organized as follows: in Section 2, we briefly recall the background for the understanding of this work. We further discuss the related work and the performance indicators that benefit from our approach. In Section 3, we first motivate the need for bias-free complete finite-size Pareto front approximations, and propose the Reference Set Generator (RSG). In Section 4, we present some numerical results on selected benchmark problems and compare the RSG to related algorithms. Finally, we draw our conclusions in Section 5 and give possible paths for future research.

2. Background and Related Work

We consider multi-objective optimization problems (MOPs) that can be mathematically expressed via
min x Q F ( x ) .
Hereby, the map F is defined as
F : Q R k , F ( x ) = ( f 1 ( x ) , , f k ( x ) ) T ,
where we assume each of the individual objectives f i : Q R , i = 1 , k , to be continuous. We stress, however, that the method we propose in the sequel, RSG, can, in principle, also be applied to discrete problems. Q is the domain of the objective functions that is typically expressed by equality and inequality constraints.
In order to define the optimality of an MOP, one can use the concept of dominance [9].
Definition 1.
(a) Let v , w R k . Then the vector v is less than w ( v < p w ), if v i < w i for all i { 1 , , k } . The relation p is defined analogously.
(b) 
y Q is dominated by a point x Q ( x y ) with respect to (MOP) if F ( x ) p F ( y ) and F ( x ) F ( y ) .
(c) 
x Q is called a Pareto point or Pareto optimal if there is no y Q that dominates x.
(d) 
The set P Q of Pareto optimal solutions
P Q : = { x Q : y Q s . t . y x }
is called the Pareto set.
(e) 
The image F ( P Q ) of the Pareto set is called the Pareto front.
One can expect that both the Pareto set and the Pareto front form under certain conditions, at least locally objects of dimension k 1 . For details, we refer to [1]. Due to this "curse of dimensionality" It is, hence, not possible for an evolutionary multi-objective optimization algorithm (MOEA) to keep all promising candidate solutions (e.g., all non-dominated ones) during the algorithm run. It is, hence inevitable – at least for continuous problems – to select which of the promising solutions should be kept in order to obtain a "suitable" approximation of the solution set (in most cases, the Pareto front of the given MOP). Within MOEAs, this process is termed "selection". Another term, which can be used synonymously, is "archiving". The latter is typically used when the MOEA is equipped with an external archive.
Most of the existing MOEAs can be divided into three main classes: (a) MOEAs that are based on the concept of dominance (e.g., [10,11,12,13]), (b) MOEAs that are based on decompositions (e.g., [14,15,16,17,18]), and (c) MOEAs that make use of an indicator function (e.g., [19,20,21,22,23]). The selection strategies of the first generation of MOEAs of class (a) are based on a combination of non-dominated sorting and niching (e.g., [24,25,26]). Later, elite preservation was included, leading to an increased overall performance (and, as a consequence, better Pareto front approximations). This holds, e.g., for SPEA ([27]), PAES ([13]), SPEA-II ([11]), and NSGA-II ([10]). Theoretical studies on selection mechanisms have been done by the groups of Rudolph ([28,29,30,31]) and Hanne ([32,33,34,35]). All of these selection mechanisms deal with the abilities of the populations to reach the Pareto set/front. On the other hand, the distributions of the individuals along the Pareto sets/fronts have not been considered. The selection mechanisms of the MOEAs within classes (b) and (c) follow directly from the construction of the algorithms: selection in a MOEA of class (b) is done by considering the values of the chosen scalarization functions. Analogously, the selection in a MOEA of class (c) is done by considering the indicator contributions.
Existing (external) archiving strategies can also be divided into three classes: (a) unbounded archivers, (b) implicitly bounded archivers, and (c) bounded archivers. Unbounded archivers store all promising solutions during the algorithm run. The magnitudes of such archives can exceed any given threshold if the algorithm is run long enough. Unbounded archivers have, e.g., been used and analyzed in [36,37,38,39,40,41,42]. ϵ -dominance ([43]) can be viewed as a weaker concept of dominance. This relation allows single solutions to “cove” entire parts of the Pareto front of a given MOP, which is the basis for most implicitly bounded archivers. Such strategies have first been considered in the context of evolutionary multi-objective optimization (EMO) by Laumanns et al. ([44]). Later works proposed and analyzed different approximations (such as gap-free approximations of the Pareto front) and deal with different sets of interest (e.g., the consideration of all nearly optimal or all ϵ -locally optimal solutions), ([40,41,45,46,47,48,49,50]). Finally, bounded archivers have, e.g., been proposed in [51,52] where adaptive grid selections have been utilized. Bounded archivers tailored to the use of the Hypervolume indicator have been suggested in [53,54]). Laumanns and Zenklusen have proposed two bounded archivers that aim for ϵ -approximations of the Pareto front ([55]).
All of the selection/archiving strategies mentioned above have in common that they aim for a "best approximation" of the set of interest out of the given (finite) data. A related but slightly different problem is to generate a "suitable" (in particular complete and bias-free) finite size approximation of the set of interest S for the sake of comparisons, even if S is known approximately or even analytically but not “trivia” (e.g., a line segment). Tian et al. [56] describe the procedure of PlatEMO [57], which uses uniform sampling on the simplex and then maps these points to the particular Pareto front of each problem. However, an analytical expression or characterization of the Pareto front is required, and in some cases, the obtained set is not completely uniform. A method related to RSG can be found in [58] which has the aim of guiding the iterates of a particular Newton method toward the Pareto front. In this work, we extend this idea for the purpose of generating complete and bias-free Pareto front approximations of relatively large magnitudes (in particular compared to population sizes used in EMO).
Finally, reference sets as the ones generated by RSG are helpful for the evaluation of the performance qualities of candidate sets (populations) in EMO. More precisely, such sets are required for all distance-based indicators. The earliest such indicators are the Generational Distance (GD, [3]) and the Inverted Generational Distance. (IGD, [5]). Later, the indicator Δ p ([6,59]) has been proposed that is a combination of slight variants of GD and IGD, and that can be viewed as an averaged version of the Hausdorff distance d H . So far, there exist several extensions of these performance indicators. For instance, the consideration of continuous sets – either only the Pareto front or also the candidate solution set – has been done in [8,60] leading to modifications of I G D and Δ p . The indicators IGD+ ([7]) and DOA ([61]) are modifications of IGD that are Pareto compliant.

3. Reference Set Generator (RSG)

In this section, we first motivate the need for complete and bias-free finite size Pareto front approximations, and then propose the Reference Set Generator (RSG) that targets for such sets.

3.1. Motivation

Distance-based indicators require a “suitable” finite-size approximation of the Pareto front in order to give a "correct" value for the approximation quality of the considered candidate solution set. This holds in particular for the above-mentioned indicators GD, IGD, and Δ p , together with their variants. Such representations are ideally uniformly spread along the entire Pareto front ([60]). This, however, is a non-trivial task unless the Pareto front is given analytically and has a relatively simple form (e.g., linear or perfectly spherical). Regrettably, this is the case for only a few test problems (e.g., DTLZ1 and DTLZ2). On the other hand, there exist quite a few benchmark MOPs where the shape of the Pareto set is relatively simple. For such problems, it is tempting to choose uniform samples from the Pareto set (i.e., X = { s 1 , , s m } , where all s i P Q ), and to use the respective image Y = F ( X ) to represent of the Pareto front. The following discussion shows, however, that this approach has to be handled with care since it can induce unwanted biases in the approximations that, in turn, may result in misleading indicator values.
As the first example, consider the one-dimensional bi-objective problem
F ( x ) = 1 1 x 1 x .
Let the domain be given by Q = [ 0.1 , 3 ] , then the Pareto set is identical to Q, and the Pareto front is the line segment that connects the points a = ( 9,10 ) and b = ( 2 / 3,1 / 3 ) . Figure 1 shows the result when using N equally spaced points along the Pareto set. As it can be seen for N = 50 and N = 500 , there is a clear bias of the images toward the right lower end of the Pareto front. For N = 10,000 , the Pareto front approximation is "complete" (at least from the practical point of view) and appears to be perfect. However, it possesses the same bias. To see the impact of the reference set on the performance indicators consider the two hypothetical outcomes (e.g., possible results from different MOP solvers):
A : = a + 2 i 1 10 ( b a ) | i { 1 , , 5 } , and B : = a + i 10 ( b a ) | i { 6 , , 10 } .
Figure 2 shows the two sets together with the Pareto front. Note that A is the perfect 5-element approximation of the Pareto front: the elements are equally distributed along the Pareto front, and the extreme points are shifted “half-way in” ([60]). The set B is certainly not perfect as it, e.g., misses to "cover" more than half of the front. Table 1 shows values I ( O , R ) for different distance-based indicators, the outcomes O { A , B } , and different representations R of the Pareto front. For R = R 10,000 x (the one shown in Fig. 1 (c)), all indicators – except d H = Δ – yield lower values for B than for A, indicating (erroneously) that B is better than A. This is not the case for d H since the Hausdorff distance is determined by the maximum of the considered distances and not by an average of those (however, d H has other disadvantages in the context of EMO, most prominently that it punishes single outliers [6]). The situation changes when selecting R = R 10,000 y as a representation of the front. This representation also contains N = 10,000 elements, but these are chosen uniformly along the Pareto front. Now, there is a tie for the two GD variants, and for all other indicators, A leads to better values than B. These values are indeed very close to the `correct’ values: all exact GD values are equal to zero since A and B are contained in the Pareto front. In order to compute the exact IGD values, a particular integral has to be solved [60]. When using R 10000 y as representation, the computation of the IGD values can be interpreted as a Riemann sum with N = 10,000 equally sized sub-intervals leading to perfect solutions, at least from the practical point of view.
We repeat the process, but now using only N = 100 elements for the representation (see Table 1). We can see the same trend, i.e., that B appears to be better for R 100 x while A appears to be better when using R 100 y . Further, we see that the indicator values for R 100 y are already quite close to the exact values (i.e., when using R 10000 y ). While the proper choice of N may not be an issue for bi-objective problems ( k = 2 ), this may get important for a larger number of objectives due to the “curse of dimensionality”: at least for continuous problems, one can expect that the Pareto front forms under certain (mild) assumptions a manifold of dimension k 1 ([1]).
We next consider further test problems. Figure 4, Figure 5 and Figure 6 show the Pareto front approximations for the commonly used test problems DTLZ1, ZDT1, and ZDT3, respectively, where the representation has been obtained via uniform sampling in decision variable space along the Pareto set. In all cases, the Pareto front representations appear to be perfect for large enough values of N. However, certain biases can be observed for lower values of N. For all test problems, we selected two points out of the representation and showed their nearest neighbors. These values differ by one order of magnitude which confirms that the solutions are not uniformly distributed along the fronts. Using such representations, the same issues can arise as discussed above.
To conclude, the suitable representation of the Pareto front of a given MOP is crucial when considering distance-based performance indicators that use an average of the distances considered. Such representations are ideally equally distributed over the front. If the representation contains a bias, this may result in misleading indicator values, leading, in turn, to a wrong evaluation of the obtained results. In particular, the approach to performing the sampling along the Pareto set is, though tempting, not appropriate for such indicator-based indicators. In the sequel, we will propose a method that aims to achieve a uniform Pareto front representations.
Figure 3. Distribution resulting from uniformly sampling the Pareto set of DTLZ1 with N points. In (a) d ( y 1 , N N y 1 ) = 0.068181240846835 , d ( y 2 , N N y 2 ) = 0.008729713347982 .
Figure 3. Distribution resulting from uniformly sampling the Pareto set of DTLZ1 with N points. In (a) d ( y 1 , N N y 1 ) = 0.068181240846835 , d ( y 2 , N N y 2 ) = 0.008729713347982 .
Preprints 154643 g003
Figure 4. Distribution resulting from uniformly sampling the Pareto set of DTLZ2 with N points. In (a) d ( y 1 , N N y 1 ) = 0.174311485495316 , d ( y 2 , N N y 2 ) = 0.030268871802677 .
Figure 4. Distribution resulting from uniformly sampling the Pareto set of DTLZ2 with N points. In (a) d ( y 1 , N N y 1 ) = 0.174311485495316 , d ( y 2 , N N y 2 ) = 0.030268871802677 .
Preprints 154643 g004
Figure 5. Distribution resulting from uniformly sampling the Pareto set of ZDT1 with N points. In (a) d ( y 1 , N N y 1 ) = 0.101010101010101 , d ( y 2 , N N y 2 ) = 0.011440746020984 .
Figure 5. Distribution resulting from uniformly sampling the Pareto set of ZDT1 with N points. In (a) d ( y 1 , N N y 1 ) = 0.101010101010101 , d ( y 2 , N N y 2 ) = 0.011440746020984 .
Preprints 154643 g005
Figure 6. Distribution resulting from uniformly sampling the Pareto set of ZDT3 with N points. In (a) d ( y 1 , N N y 1 ) = 0.066834581606972 , d ( y 2 , N N y 2 ) = 0.004486990395919 .
Figure 6. Distribution resulting from uniformly sampling the Pareto set of ZDT3 with N points. In (a) d ( y 1 , N N y 1 ) = 0.066834581606972 , d ( y 2 , N N y 2 ) = 0.004486990395919 .
Preprints 154643 g006

3.2. RSG

In the following, we assume we are interested in a Pareto front approximation of size N for a given MOP. Further, we assume that we are given a set A y of (in principle) arbitrary size of non-dominated, possibly non-uniformly distributed points that are "close enough" to the PF. Below, we will discuss different strategies to obtain A y . Given this data, the Reference Set Generation (RSG) process consists of three main steps: component detection, filling, and reduction. The idea is to fill the gaps between the points within each connected point of A y . This leads to a more complete set F with a higher cardinality than A y , which can then be reduced to obtain a uniform reference set of size N. In general, PFs can be disconnected, and if we simply fill the gaps in A y , we may introduce points that do not belong to the PF (Figure 7(b)). Therefore, component detection must be performed before applying the filling process to each detected component (Figure 7(a)).
The general procedure of RSG is presented in Algorithm 1, and each of the main steps will be explained in the following subsections. Figures 8, 12, 9, and 13 illustrate all the steps for the problems ZDT3, CONV3-4, DTLZ7 and CONV4-2F. It is important to note that the cases for k = 2 and k 3 objectives exhibit slight variations in the component detection step and follow completely different procedures for the filling step.
Algorithm 1 Reference Set Generation (RSG)
Input: 
Starting set A y , filling size N f , output size N.
Output: 
PF reference Z.
1:
C = { C 1 , , C n c } ComponentDetection( A y )
2:
for  i = 1 , , n c  do               ▹ for each component
3:
      F i Filling( C i , N f · | C i | | A y | )
4:
end for
5:
F i F i
6:
Z reduction( F , N )
7:
return Z

3.3. Component Detection

Since the PF might be disconnected, we first apply a component detection on A y . We use DBSCAN [62] in the objective space for this purpose for three main reasons: (i) the number of components does not need to be known a priori, (ii) the method detects outliers, and (iii) we have observed that a density-based approach works better than a distance-based one (e.g., k-means) for component detection. DBSCAN has two parameters: m i n p t s and r. To make the component detection process “parameter-free”, we compute a small grid-search to find the optimal values of m i n p t s and r based on the weakest link function defined in [63]. The default parameter values are m i n p t s { 2,3 } and r { 0.10 d ¯ , 0.11 d ¯ , , 0.15 d ¯ } for bi-objective problems, and m i n p t s { 3,4 } and r { 0.19 d ¯ , 0.20 d ¯ , , 0.23 d ¯ } otherwise, where d ¯ is the average pairwise distance between all points. However, these parameters can be adjusted when calling RSG. A summary of the component detection process is presented in Algorithm 2. In the following, we describe the remaining steps for a single connected component. If multiple components exist, the procedures must be repeated analogously for each component C i identified by Algorithm 2.
Algorithm 2 Component Detection
Input: 
starting set A y = { p 1 , , p } , number of objectives k
Output: 
Number of clusters n c , clusters C = { C 1 , , C n c } .
1:
Set d ¯ 2 ( 1 ) p i p j | p i p j |
2:
if  k = 2  then
3:
    Set r _ r a n g e { 0.1 d ¯ , 0.11 d ¯ , , 0.16 d ¯ }
4:
    Set m i n p t s _ r a n g e { 2,3 }
5:
else
6:
    Set r _ r a n g e { 0.19 d ¯ , 0.20 d ¯ , , 0.23 d ¯ }
7:
    Set m i n p t s _ r a n g e { 3,4 }
8:
end if
9:
Set w l m i n
10:
for  m i n p t s in m i n p t s _ r a n g e  do
11:
    for  r in r _ r a n g e  do
12:
         C t , n c t DBSCAN( A y , r , m i n p t s )
13:
         w l WeakestLink( C t )
14:
        if  w l w l m i n  then
15:
            C C t
16:
            n c n c t
17:
            w l m i n w l
18:
        end if
19:
    end for
20:
end for
21:
return  C = { C 1 , , C n c } , n c

3.4. Filling

Even if we know the PS a priori, a uniform sampling of the PS will not result in a uniform sampling of the PF. We assume that we have a set of points A y that is not uniformly distributed. However, if we fill the gaps and select N points from the filled set, we can obtain a more uniform set, leading to better IGD approximations when selecting points from these filled sets. The idea behind the filling step is to create a set that is as uniform as possible, so that the reduction step (in particular, k-means) does not get stuck in non-uniform local optima, which would lead to non-uniform final sets. The next task is, therefore, to compute N f solutions that are ideally uniformly distributed along A y .
This process is performed differently for k = 2 and k 3 objectives:
  • For k = 2 , we sort the points of A y = { p 1 , , p } in increasing order of f 1 , i.e., the first objective. Then, we consider the piecewise linear curve formed by the segments between p 1 and p 2 , p 2 and p 3 , and so on. The total length of this curve is given by | L | = i = 1 1 | L i | , where | L i | = p i p i + 1 2 . To perform the filling, we arrange the N f desired points along the curve L such that the first point is p 1 and the subsequent points are distributed equidistantly along L. This is achieved by placing each point at a distance of δ = | L | / ( N f 1 ) from the previous one along L. See Algorithm 3 for details.
  • The filling process for k 3 consists of several intermediate steps that must be described first; see Algorithm 4 for a general outline of the procedure. The procedure is as follows: First, to better represent A y (particularly for the filling step), we triangulate this set in k 1 dimensional space. This is done because the PF for continuous MOPs forms a set whose dimension is at most k 1 . To achieve this, we compute a “normal vector” η to A y using equation 9, and then we project it onto the k 1 hyperplane normal to η , obtaining the projected set P k 1 . After this, we compute the Delaunay triangulation [64] of P k 1 , which provides a triangulation T that can be used in the original k-dimensional space. For concave PFs, the triangulation may include triangles (or simplex for k > 3 ) that extend beyond A y (Figure 12(d)), so a removal strategy is applied to eliminate these triangles and obtain the final triangulation T. Finally, each triangle t i T is uniformly filled at random with a number of points proportional to its area (or volume for k > 3 ), resulting in the filled set F of size N f .
    We will now describe each step in more detail in the following:
    Computing “normal vector” ( η ) . Since the front is not known, we compute the normal direction η orthogonal to the convex hull defined by the minimal elements of A y . More precisely, we compute η as follows: if A y = { p 1 , , p } , choose
    y ( i ) m arg min j = 1 , , N p j , i , i = 1 , , k ,
    where p j , i denotes the i-th element of p j , and set
    M : = ( y ( 2 ) m y ( 1 ) m , y ( 3 ) m y ( 1 ) m , , y ( k ) m y ( 1 ) m ) R k × ( k 1 ) .
    Next, compute a QR-factorization of M, i.e.,
    M = Q R = ( q 1 , , q k ) R ,
    where Q R k × k is an orthogonal matrix with column vectors q i , and R R k × ( k 1 ) is a right upper triangular matrix. Then, the vector
    η = sgn ( q k , 1 ) q k q k 2
    is the desired shifting direction. Since Q is orthogonal, the vectors v 1 : = q 1 , , v k 1 : = q k 1 form an orthonormal basis of the hyperplane that is orthogonal to η . That is, these vectors can be used for the construction of P k 1 .
    k 1 Projection ( P k 1 ) . We use η as the first axis of a new coordinate system ( η , v 1 , , v k 1 ) , where the vectors v i are defined as above. In this coordinate system, the orthonormal vectors v 1 , , v k 1 form the basis of a hyperplane orthogonal to η . The projection of A y onto this hyperplane ( P k 1 ) is achieved by first expressing A y in this new coordinate system as P = β η + β 1 v 1 + + β k 1 v k 1 , and then removing the first coordinate, yielding P k 1 = β 1 v 1 + + β k 1 v k 1 .
    Delaunay Triangulation ( T ) . Compute the Delaunay triangulation of P k 1 . This returns T , a list of size δ containing the indices of P k 1 that form the triangles (or simplices for k > 3 ). The list T serves as the triangulation for the k-dimensional set A y , which is possible because T consists of indices, making it independent of the dimension. We use δ to denote the number of triangles obtained, { T ( i , 1 ) , , T ( i , k ) } the indices of the vertices forming triangle i and { p T ( i , 1 ) , , p T ( i , k ) } to denote the corresponding vertices of triangle i.
    Triangle Cleaning ( T ) . We identify three types of unwanted triangles: those with large sides, those with large areas, and those where the matrix containing the coordinates of the vertices has a large condition number. The type of cleaning applied depends on the problem; however, the procedure remains the same for any problematic triangle case and is outlined in Algorithm 5. First, the property ρ i (area, largest side, or condition number) is computed for all the triangles i = 1 , , δ . Next, triangles i with ρ i > τ are removed. The parameter τ is also problem-dependent, and the specific values used for each problem will be detailed in the results section.
    Triangle Filling ( F ) . For each triangle t i T with area a i , we generate a i A N f points uniformly at random inside triangle t i , following the procedure described in [65]. That is, the number of points is proportional to the area (or volume) of each triangle (or simplex). Here, A = i = 1 δ a i is the total area of the triangulation.
Algorithm 3 Filling ( k = 2 Objectives)
Input: 
starting set A y = { p 1 , , p } , filling size N f
Output: 
Filled set F = { y 1 , , y N f }
1:
X = { x 1 , , x } sort A y according to its first objective f 1
2:
L i x i x i + 1 i = 1 , , 1
3:
L i = 1 1 L i
4:
δ L / ( N f 1 )
5:
d i s t _ l e f t ( 0 , 0 , , 0 ) R
6:
for  i = 1 : 1  do            ▹ compute number of points per segment
7:
     r a t i o ( | L i | + d i s t _ l e f t ( i ) ) / δ
8:
     p o i n t s _ p e r _ s e g m e n t ( i ) r a t i o
9:
     d i s t _ l e f t ( i + 1 ) ( r a t i o r a t i o ) · δ
10:
end for
11:
c o u n t = 1
12:
for  i = 1 : 1  do              ▹ for each line segment
13:
    if  p o i n t s _ p e r _ s e g m e n t ( i ) > 0  then      ▹ check if a point lands in segment L i
14:
         ν i : = ( x i + 1 x i ) / L i
15:
         y c o u n t = x i + ( δ d i s t _ l e f t ( i ) ) · ν i
16:
         c o u n t c o u n t + 1
17:
        for  j = 2 : p o i n t s _ p e r _ s e g m e n t ( i )  do        ▹ if L i has more than one point
18:
            y c o u n t y c o u n t 1 + δ · ν i
19:
            c o u n t c o u n t + 1
20:
        end for
21:
    end if
22:
end for
23:
return  F = { y 1 , , y N f }
Algorithm 4 Filling ( k 3 Objectives)
Input: 
starting set A y = { p 1 , , p } , filling size N f
Output: 
Filled set F = { y 1 , , y N f }
1:
η normal_vector ( A y )
2:
P k 1 projection ( A y , η )
3:
T DelaunayTriangulation ( P k 1 )
4:
T TriangleCleaning ( T )
5:
F TriangleFilling ( T , A y )
6:
return  F = { y 1 , , y N f }
Algorithm 5: Triangle Cleaning
Input: 
starting set A y = { p 1 , , p l } , triangulation T R k × δ , number of triangles δ , parameter threshold τ
Output: 
cleaned triangulation T of size δ
1:
if chosen property is area then
2:
     ρ i 1 k ! det p T ( i , 2 ) T p T ( i , 1 ) T p T ( i , 3 ) T p T ( i , 1 ) T p T ( i , k ) T p T ( i , 1 ) T p T ( i , 2 ) T p T ( i , 1 ) T p T ( i , 3 ) T p T ( i , 1 ) T p T ( i , k ) T p T ( i , 1 ) T T 1 / 2 i = 1 , , δ
3:
else if chosen property is largest side then
4:
     ρ i largest side of simplex with vertices { p T ( i , 1 ) , , p T ( i , k ) }
5:
else if chosen property is condition number then
6:
     ρ i κ p T ( i , 1 ) p T ( i , 2 ) p T ( i , k )         ▹ κ ( A ) is the condition number of matrix A
7:
end if
8:
ρ i = 1 δ ρ i
9:
T T
10:
if  ρ i > τ · ρ  then
11:
    remove triangle i from T
12:
end if
13:
return T, number of triangles δ

3.5. Reduction

Once we have computed the filled set F, we need to select N points that are ideally evenly distributed along F. To this end, we use k-means clustering with N clusters, as there is a strong relationship between k-means and the optimal I G D subset selection ([66,67]). The resulting N cluster centroids form the PF reference set Z. Note that this reduction method can be modified if needed to obtain an ideal reference according to other types of indicators (i.e., those that are not distance-based).

3.6. Obtaining A y

RSG requires an initial approximation A y of the Pareto front. Note that by construction of the algorithm, this set can have small imperfections (which can be removed by the filling step) and can also have biases in the approximation (reduction step). However, it is desired that A y "captures" the shape of the entire Pareto front. The computation of such an approximation is certainly problem-dependent. For our computations, we have used the following three main procedures to get A y :
  • sampling: For some benchmark problems, either the Pareto set or the Pareto front is given in analytic form. If a sampling can be performed in objective space (e.g., for linear fronts, the remaining steps of the RSG may not be needed to further improve the quality of the solution set. If the sampling is performed in decision variable space, the elements of the resulting image A y may not be uniformly distributed along the Pareto front as discussed above. However, in that case, the filling and reduction step will help to remove biases.
  • archiving: The result of an MOEA or any other MOP solver can, of course, be taken. This could be either the final archive of the population, via merging several populations of the same or several runs ([58]), or via using external (unbounded) archives ([50]). Note that this includes taking a reference set from a given repository. We have used archiving, e.g., for the test problems WFG3-9, DTLZ1-4, DTLZ7, ZDT1-6, CONV3, CONV3-4, and CONV4-2F.
  • continuation: An alternative to the above mentioned techniques is to make use of multi-objective continuation methods, probably in combination with the use of several different starting points. In particular, we have used the Pareto Tracer (PT, [68,69,70]), a state-of-the-art continuation method that is able to treat problems of in principle any dimensions (both n and k), can handle general constraints and that can even detect local degeneration of the solution set. We have used PT, e.g., for the test problem WFG1, WFG2, DTLZ5, and DTLZ6.

3.7. Complexity Analysis

The overall complexity is O ( γ 2 + τ 2 N k N f ) for k = 2 (regardless of the number of components), and O ( γ 2 + δ k 2 + τ 2 N k N f ) for k 3 , where is the size of the initial approximation A y , k is the number of objectives, δ is the number of triangles in the Delaunay triangulation, τ 2 is the number of iterations of k-means (bounded to 500 in this work), N is the desired size of the reference set Z, and N f is the size of the filling. This assumes k and that the triangle-cleaning method used is based on the longest side (which was the method applied to all the references presented in this work). Typically, obtaining a decent approximation requires a large value of N f , making the clustering step the dominant one and thus reducing the overall complexity to O ( τ 2 N k N f ) for any k. We now present the complexity analysis in detail for each step separately, considering a single component. We now present the complexity analysis in detail for each step separately, considering a single component.
  • Component Detection. The time complexity is O ( ( ) 2 + γ ( log ( ) + ( ) 2 ) ) which accounts for the computation of the average distance, plus the size of the grid search ( γ ) multiplied by the sum of the complexities of DBSCAN and the WeakestLink computation. Here, is the size of A y , and γ represents the number of parameter combinations of the grid search, with γ = 14 for k = 2 and γ = 10 for k 3 using the default values. If it is previously known that the Pareto front is connected, then the parameters of DBSCAN can be correctly adjusted, and γ can be set to 1.
  • Filling. The time complexity depends on the number of objectives:
    For k = 2 the time complexity is O ( + k N f ) , which accounts for sorting and placing the N f points along the line segments.
    For k 3 the time complexity is O ( [ k + k ( k 1 ) 2 ] + k 2 + log + [ ( k 1 ) δ + k N f ] ) due to the computations involved in determining the normal vector η , changing coordinates and projecting, performing the Delaunay triangulation, and filling the triangles. Here, δ represents the size of the cleaned Delaunay triangulation, i.e., the number of triangles. Additionally, triangle cleaning must be considered, though its complexity depends on the method used. It is given by O ( δ k 3 ) when cleaning is based on area or the condition number (due to determinant computation), or O ( δ k 2 ) when cleaning based on the longest side.
  • Select Reference Set T. The time complexity is O ( τ 2 N k N f ) due to the k-means clustering algorithm. Here, τ 2 is the number of iterations of k-means.

4. Results

In this section, we show the strength of the novel approach to selected test problems. We further show – as far as possible – comparisons to related methods.
First, we show all the steps of the RSG on five representative problems: Figure 8 shows these steps for ZDT3 ( k = 2 and disconnected), Figure 9 for DTLZ7 ( k = 3 and disconnected), Figure 11 for CONV3 (defined in Appendix A, k = 3 ), Figure 12 for CONV3-4 (defined in Appendix A, k = 3 ) and Figure 13 for CONV4-2F (defined in Appendix A, k = 4 and disconnected).
In Figure 8 and Figure 9, we have used starting sets A y that have a slight bias. The last subfigures show the results of RSG for different values of N. For Figure 11, Figure 12 and Figure 13, we demonstrate the universality of RSG, as it does not need any analytical information of the PF since it utilizes the result of a MOEA and an archiver to generate a reference of any desired size.
Additionally, a comparison between RSG, PlatEMO [57], Pymoo, and the filling step is presented in Figure 14 and Figure 15 for the WFG2, ZDT1, ZDT3, DTLZ2, Convex DTLZ2 (CDTLZ2), C2-DTLZ2, and DTLZ7 test problems. Although the reference sets provided by PlatEMO and Pymoo are of high quality, they still exhibit some bias in certain bi-objective problems (such as ZDT1 and ZDT3), and especially in three-objective problems such as WFG2, and CDTLZ2. Furthermore, for problems like WFG2, and DTLZ7, the number of points in the reference sets of PlatEMO and Pymoo is limited to a fixed set, in contrast to RSG, which can generate any desired number of points.
Upon acceptance of this paper, we will provide reference sets on our website http://neo.cinvestav.mx:3005/ for the problems considered in this work as well as other problems (in the future, we intend to provide data for all commonly used benchmark test functions). The website will allow users to request reference sets of any desired size for any of the aforementioned problems. Further, the codes of RSG will be found on our GitHub page https://github.com/aerfangel/RSG.

5. Conclusions and Future Work

In this paper, we have addressed the problem of obtaining bias-free and complete finite-size approximations of the solution sets (Pareto fronts) of multi-objective optimization problems (MOPs). Such approximations are, in particular, required for the fair usage of distance-based performance indicators, which are frequently used in evolutionary multi-objective optimization (EMO). If the Pareto front approximations are biased or incomplete, the use of these performance indicators can lead to misleading or false information. To address this issue, we have proposed the Reference Set Generator (RSG). This method starts with an initial (probably biased) approximation of the Pareto front. An unbiased approximation is then computed via component detection, filling, and a reduction to the desired size. The RSG can be applied to Pareto fronts of any shape and dimension. We have finally demonstrated the strength of the novel approach on several benchmark problems.
In the future, we intend to use the RSG on the Pareto fronts of all commonly used continuous test problems. Special attention has to be paid to problems with degenerated fronts (i.e., problems where the Pareto front does locally not form an object of dimension k 1 ). In the current approach, we handle degeneracy using the PT to obtain a filled set and, from there, use the reduction step. For future work, we will explore if the projection can be modified to fill such sets, which may lead to a more general approach to handling degeneration.

Author Contributions

All authors contributed equally to this work.

Funding

O. Schütze acknowledges support from CONAHCYT project CBF2023-2024-1463.

Data Availability Statement

The code and data presented in the study will be openly available upon acceptance in GitHub at https://github.com/aerfangel/RSG and http://neo.cinvestav.mx:3005/ respectively.

Acknowledgments

Angel E. Rodriguez-Fernandez acknowledges support from the SECIHTI to pursue his postdoc fellowship at the CINVESTAV-IPN.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
A y Current PF approximation
Size of A y , the starting PF approximation
N Desired size of approximation
Z RSG result: Reference set of size N
F Filled set
N f Size of Filling
C i i-th detected component
C Set of all detected components
L Total length of 2D curve
T Delaunay triangulation
δ Number of triangles in T
T Cleaned Triangulation
δ Number of triangles in T
η Normal vector
P k 1 Projected A y
ρ Selected cleaning property (area/volume, largest side, or condition number)
ρ i Value of property ρ for triangle i
τ Threshold for removing triangles
a i Area/volume of triangle/simplex i
A Total area/volume of the triangulation
r Radius of DBSCAN

Appendix A Function Definitions

  • CONV3
    F : [ 3 , 3 ] 3 R 3 F ( x ) = ( f 1 ( x ) , f 2 ( x ) , f 3 ( x ) ) T where : f i ( x ) = x a i 2 a 1 = ( 1 , 1 , 1 ) , a 2 = ( 1,1,1 ) , a 3 = ( 1,1 , 1 )
  • CONV3-4
    F : [ 3 , 3 ] 3 R 3 F ( x ) = ( f 1 ( x ) , f 2 ( x ) , f 3 ( x ) ) T where : f 1 ( x ) = ( x 1 a 1 1 ) 4 + ( x 2 a 2 1 ) 2 + ( x 3 a 3 1 ) 2 f 2 ( x ) = ( x 1 a 1 2 ) 2 + ( x 2 a 2 2 ) 4 + ( x 3 a 3 2 ) 2 f 3 ( x ) = ( x 1 a 1 3 ) 2 + ( x 2 a 2 3 ) 2 + ( x 3 a 3 3 ) 4 a 1 = ( 1 , 1 , 1 ) , a 2 = ( 1,1,1 ) , a 3 = ( 1,1 , 1 )
  • CONV4-2F
    F : [ 3 , 3 ] 4 R 4 F ( x ) = ( f 1 ( x ) , f 2 ( x ) , f 3 ( x ) , f 4 ( x ) ) T where : f i ( x ) = x + 1 a i 2 3.5 σ if x < ( 0 , 0 , 0 , 0 ) x a i 2 otherwise ϕ 1 = ( 0 , a 1 a 2 2 , a 1 a 3 2 , a 1 a 4 2 ) ϕ 4 = ( a 4 a 1 2 , a 4 a 2 2 , a 4 a 3 2 , 0 ) σ = ϕ 4 ϕ 1 a 1 = ( 1,0,0,0 ) , a 2 = ( 0,1,0,0 ) a 3 = ( 0,0,1,0 ) , a 4 = ( 0 , 0 , 0 , 1 ) , 1 = ( 1 , 1 , 1 , 1 ) ,

References

  1. Hillermeier, C. Nonlinear Multiobjective Optimization: A Generalized Homotopy Approach; Vol. 135, Springer Science & Business Media, 2001.
  2. Coello Coello, C.A.; Goodman, E.; Miettinen, K.; Saxena, D.; Schütze, O.; Thiele, L. Interview: Kalyanmoy Deb Talks about Formation, Development and Challenges of the EMO Community, Important Positions in His Career, and Issues Faced Getting His Works Published. Mathematical and Computational Applications 2023, 28. [Google Scholar] [CrossRef]
  3. Veldhuizen, D.A.V. Multiobjective evolutionary algorithms: classifications, analyses, and new innovations. Technical report, Air Force Institute of Technology, 1999.
  4. Zitzler, E.; Thiele, L.; Laumanns, M.; Fonseca, C.M.; Grunert, V.D.F. Performance assessment of multiobjective optimizers: An analysis and review. IEEE Transactions on Evolutionary Computation 2003, 7, 117–132. [Google Scholar] [CrossRef]
  5. Coello, C.A.C.; Cruz, N.C. Solving Multiobjective Optimization Problems Using an Artificial Immune System. Genetic Programming and Evolvable Machines 2005, 6, 163–190. [Google Scholar] [CrossRef]
  6. Schütze, O.; Esquivel, X.; Lara, A.; Coello Coello, C.A. Using the averaged Hausdorff distance as a performance measure in evolutionary multi-objective optimization. IEEE Transactions on Evolutionary Computation 2012, 16, 504–522. [Google Scholar] [CrossRef]
  7. Ishibuchi, H.; Masuda, H.; Nojima, Y. A Study on Performance Evaluation Ability of a Modified Inverted Generational Distance Indicator, New York, NY, USA, 2015; p. 695–702. [CrossRef]
  8. Bogoya, J.M.; Vargas, A.; Cuate, O.; Schütze, O. A (p,q)-Averaged Hausdorff Distance for Arbitrary Measurable Sets. Mathematical and Computational Applications 2018, 23. [Google Scholar] [CrossRef]
  9. Deb, K.; Ehrgott, M. On Generalized Dominance Structures for Multi-Objective Optimization. Mathematical and Computational Applications 2023, 28. [Google Scholar] [CrossRef]
  10. Deb, K.; Pratap, A.; Sameer, S.A.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. Evolutionary Computation, IEEE Transactions on 2002, 6, 182–197. [Google Scholar] [CrossRef]
  11. Zitzler, E.; Laumanns, M.; Thiele, L. SPEA2: Improving the Strength Pareto Evolutionary Algorithm for Multiobjective Optimization. In Proceedings of the Evolutionary Methods for Design, Optimisation and Control with Application to Industrial Problems (EUROGEN 2001); Giannakoglou, K.; et al., Eds. International Center for Numerical Methods in Engineering (CIMNE), 2002, pp. 95–100.
  12. Fonseca, C.M.; Fleming, P.J. An overview of evolutionary algorithms in multiobjective optimization. Evolutionary Computation 1995, 3, 1–16. [Google Scholar] [CrossRef]
  13. Knowles, J.D.; Corne, D.W. Approximating the nondominated front using the Pareto Archived Evolution Strategy. Evolutionary Computation 2000, 8, 149–172. [Google Scholar] [CrossRef]
  14. Zhang, Q.; Li, H. MOEA/D: A Multi-objective Evolutionary Algorithm Based on Decomposition. IEEE Transactions on Evolutionary Computation 2007, 11, 712–731. [Google Scholar] [CrossRef]
  15. Deb, K.; Jain, H. An evolutionary many-objective optimization algorithm using reference-point-based nondominated sorting approach, part I: Solving problems with box constraints. Transactions on Evolutionary Computation 2014, 18, 577–601. [Google Scholar] [CrossRef]
  16. Jain, H.; Deb, K. An Evolutionary Many-Objective Optimization Algorithm Using Reference-Point Based Nondominated Sorting Approach, Part II: Handling Constraints and Extending to an Adaptive Approach. IEEE Transactions on Evolutionary Computation 2014, 18, 602–622. [Google Scholar] [CrossRef]
  17. Zuiani, F.; Vasile, M. Multi Agent Collaborative Search based on Tchebycheff decomposition. Computational Optimization and Applications 2013, 56, 189–208. [Google Scholar] [CrossRef]
  18. Moubayed, N.A.; Petrovski, A.; McCall, J. (DMOPSO)-M-2: MOPSO Based on Decomposition and Dominance with Archiving Using Crowding Distance in Objective and Solution Spaces. Evolutionary Computation 2014, 22. [Google Scholar] [CrossRef] [PubMed]
  19. Beume, N.; Naujoks, B.; Emmerich, M.T.M. SMS-EMOA: Multiobjective selection based on dominated hypervolume. Eur. J. Oper. Res. 2007, 181, 1653–1669. [Google Scholar] [CrossRef]
  20. Zitzler, E.; Thiele, L.; Bader, J. SPAM: Set Preference Algorithm for multiobjective optimization. In Proceedings of the Parallel Problem Solving From Nature PPSN X, 2008, pp. 847–858.
  21. Wagner, T.; Trautmann, H. Integration of Preferences in Hypervolume-based multiobjective evolutionary algorithms by means of desirability functions. IEEE Transactions on Evolutionary Computation 2010, 14, 688–701. [Google Scholar] [CrossRef]
  22. Schütze, O.; Domínguez-Medina, C.; Cruz-Cortés, N.; de la Fraga, L.G.; Sun, J.Q.; Toscano, G.; Landa, R. A scalar optimization approach for averaged Hausdorff approximations of the Pareto front. Engineering Optimization 2016, 48, 1593–1617. [Google Scholar] [CrossRef]
  23. Sosa-Hernández, V.A.; Schütze, O.; Wang, H.; Deutz, A.; Emmerich, M. The Set-Based Hypervolume Newton Method for Bi-Objective Optimization. IEEE Transactions on Cybernetics 2020, 50, 2186–2196. [Google Scholar] [CrossRef]
  24. Fonseca, C.M.; Fleming, P.J. Genetic algorithms for multiobjective optimization: formulation, discussion, and generalization. In Proceedings of the 5-th International Conference on Genetic Algorithms; 1993; pp. 416 – 423.
  25. Srinivas, N.; Deb, K. Multiobjective optimization using nondominated sorting in genetic algorithms. Evolutionary Computation 1994, 2, 221–248. [Google Scholar] [CrossRef]
  26. Horn, J.; Nafpliotis, N.; Goldberg, D.E. A niched Pareto genetic algorithm for multiobjective optimization. In Proceedings of the Proceedings of the First IEEE Conference on Evolutionary Computation, IEEE World Congress on Computational Computation. IEEE Press, 1994, pp. 82 – 87.
  27. Zitzler, E.; Thiele, L. Multiobjective evolutionary algorithms: A comparative case study and the strength Pareto approach. IEEE Transactions on Evolutionary Computation 1999, 3, 257–271. [Google Scholar] [CrossRef]
  28. Rudolph, G. Finite Markov Chain results in evolutionary computation: A Tour d’Horizon. Fundamenta Informaticae 1998, 35, 67–89. [Google Scholar] [CrossRef]
  29. Rudolph, G. On a multi-objective evolutionary algorithm and its convergence to the Pareto set. In Proceedings of the IEEE International Conference on Evolutionary Computation (ICEC 1998). IEEE Press, 1998, pp. 511 – 516.
  30. Rudolph, G.; Agapie, A. Convergence Properties of Some Multi-Objective Evolutionary Algorithms. In Proceedings of the Evolutionary Computation (CEC), 2011 IEEE Congress on. IEEE Press, 2000.
  31. Rudolph, G. Evolutionary Search under Partially Ordered Fitness Sets. In Proceedings of the Proceedings of the International NAISO Congress on Information Science Innovations (ISI 2001). ICSC Academic Press, Sliedrecht, The Netherlands, 2001, pp. 818 – 822.
  32. Hanne, T. On the convergence of multiobjective evolutionary algorithms. European Journal of Operational Research 1999, 117, 553–564. [Google Scholar] [CrossRef]
  33. Hanne, T. Global multiobjective optimization with evolutionary algorithms: Selection mechanisms and mutation control. In Proceedings of the Evolutionary Multi-Criterion Optimization, First International Conference, EMO 2001, Zurich, Switzerland. Springer Berlin, 2001, pp. 197 – 212.
  34. Hanne, T. A multiobjective evolutionary algorithm for approximating the efficient set. European Journal of Operational Research 2007, 176, 1723–1734. [Google Scholar] [CrossRef]
  35. Hanne, T. A Primal-Dual Multiobjective Evolutionary Algorithm for Approximating the Efficient Set. In Proceedings of the Evolutionary Computation (CEC), 2007 IEEE Congress on. IEEE Press, 2007, pp. 3127 – 3134.
  36. Brockhoff, D.; Tran, T.D.; Hansen, N. Benchmarking numerical multiobjective optimizers revisited. In Proceedings of the Proceedings of the 2015 Annual Conference on Genetic and Evolutionary Computation, 2015, pp. 639–646.
  37. Wang, R.; Zhou, Z.; Ishibuchi, H.; Liao, T.; Zhang, T. Localized weighted sum method for many-objective optimization. IEEE Transactions on Evolutionary Computation 2016, 22, 3–18. [Google Scholar] [CrossRef]
  38. Pang, L.M.; Ishibuchi, H.; Shang, K. Algorithm Configurations of MOEA/D with an Unbounded External Archive. arXiv preprint arXiv:2007.13352 2020.
  39. Nan, Y.; Shu, T.; Ishibuchi, H. Effects of External Archives on the Performance of Multi-Objective Evolutionary Algorithms on Real-World Problems. In Proceedings of the 2023 IEEE Congress on Evolutionary Computation (CEC), 2023, pp. 1–8. [CrossRef]
  40. Rodriguez-Fernandez, A.E.; Schäpermeier, L.; Hernández, C.; Kerschke, P.; Trautmann, H.; Schütze, O. Finding ϵ-Locally Optimal Solutions for Multi-Objective Multimodal Optimization. IEEE Transactions on Evolutionary Computation 2024, pp. 1–1. [CrossRef]
  41. Schütze, O.; Rodriguez-Fernandez, A.E.; Segura, C.; Hernández, C. Finding the Set of Nearly Optimal Solutions of a Multi-Objective Optimization Problem. IEEE Transactions on Evolutionary Computation 2024. [Google Scholar]
  42. Nan, Y.; Ishibuchi, H.; Pang, L.M. Small Population Size is Enough in Many Cases with External Archives. In Proceedings of the Evolutionary Multi-Criterion Optimization; et al., H.S., Ed. Springer Nature Singapore, 2025, pp. 99–113.
  43. Author, P. Placehodler Title. Journal of Placeholder 1984, 42, 4200. [Google Scholar]
  44. Laumanns, M.; Thiele, L.; Deb, K.; Zitzler, E. Combining convergence and diversity in evolutionary multiobjective optimization. Evolutionary Computation 2002, 10, 263–282. [Google Scholar] [CrossRef] [PubMed]
  45. Schütze, O.; Laumanns, M.; Coello Coello, C.A.; Dellnitz, M.; Talbi, E.G. Convergence of Stochastic Search Algorithms to Finite Size Pareto Set Approximations. Journal of Global Optimization 2008, 41, 559–577. [Google Scholar] [CrossRef]
  46. Schütze, O.; Laumanns, M.; Tantar, E.; Coello Coello, C.A.; Talbi, E.G. Computing gap free Pareto front approximations with stochastic search algorithms. Evolutionary Computation 2010, 18, 65–96. [Google Scholar] [CrossRef]
  47. Schütze, O.; Lara, A.; Coello, C.A.C.; Vasile, M. On the Detection of Nearly Optimal Solutions in the Context of Single-Objective Space Mission Design Problems. Journal of Aerospace Engineering 2011, 225, 1229–1242. [Google Scholar] [CrossRef]
  48. Schütze, O.; Vasile, M.; Coello, C.A.C. Computing the Set of Epsilon-Efficient Solutions in Multiobjective Space Mission Design. Journal of Aerospace Computing, Information, and Communication 2011, 8, 53–70. [Google Scholar] [CrossRef]
  49. Schütze, O.; Hernández, C.; Talbi, E.G.; Sun, J.Q.; Naranjani, Y.; Xiong, F.R. Archivers for the Representation of the Set of Approximate Solutions for MOPs. Journal of Heuristics 2019, 5, 71–105. [Google Scholar] [CrossRef]
  50. Schütze, O.; Hernández, C. Archiving Strategies for Evolutionary Multi-objective Optimization Algorithms; Springer, 2021.
  51. Knowles, J.D.; Corne, D.W. Properties of an adaptive archiving algorithm for storing nondominated vectors. IEEE Transactions on Evolutionary Computation 2003, 7, 100–116. [Google Scholar] [CrossRef]
  52. Knowles, J.D.; Corne, D.W. Bounded Pareto archiving: Theory and practice. In Proceedings of the Metaheuristics for Multiobjective Optimisation. Springer, 2004, pp. 39 – 64.
  53. Knowles, J.D.; Corne, D.W.; Fleischer, M. Bounded archiving using the Lebesgue measure. In Proceedings of the Proceedings of the IEEE Congress on Evolutionary Computation. IEEE Press, 2003, pp. 2490 – 2497.
  54. López-Ibáñez, M.; Knowles, J.D.; Laumanns, M. On Sequential Online Archiving of Objective Vectors. In Proceedings of the Evolutionary Multi-Criterion Optimization (EMO 2011). Springer, Berlin, Heidelberg, 2011, pp. 46 – 60.
  55. Laumanns, M.; Zenklusen, R. Stochastic convergence of random search methods to fixed size Pareto front approximations. European Journal of Operational Research 2011, 213, 414–421. [Google Scholar] [CrossRef]
  56. Tian, Y.; Xiang, X.; Zhang, X.; Cheng, R.; Jin, Y. Sampling Reference Points on the Pareto Fronts of Benchmark Multi-Objective Optimization Problems. In Proceedings of the 2018 IEEE Congress on Evolutionary Computation (CEC), 2018, pp. 1–6. [CrossRef]
  57. Tian, Y.; Cheng, R.; Zhang, X.; Jin, Y. PlatEMO: A MATLAB platform for evolutionary multi-objective optimization. IEEE Computational Intelligence Magazine 2017, 12, 73–87. [Google Scholar] [CrossRef]
  58. Wang, H.; Rodriguez-Fernandez, A.E.; Uribe, L.; Deutz, A.; na, O.C.P.; Schütze, O. A Newton Method for Hausdorff Approximations of the Pareto Front Within Multi-objective Evolutionary Algorithms. IEEE Transactions on Evolutionary Computation 2024, pp. 1–1. [CrossRef]
  59. Bogoya, J.M.; Vargas, A.; Schütze, O. The Averaged Hausdorff Distances in Multi-Objective Optimization: A Review. Mathematics 2019, 7. [Google Scholar] [CrossRef]
  60. Rudolph, G.; Schütze, O.; Grimme, C.; Domínguez-Medina, C.; Trautmann, H. Optimal averaged Hausdorff archives for bi-objective problems: Theoretical and numerical results. Comput Optim Appl 2016, 64, 589–618. [Google Scholar] [CrossRef]
  61. Dilettoso, E.; Rizzo, S.A.; Salerno, N. A Weakly Pareto Compliant Quality Indicator. Mathematical and Computational Applications 2017, 22. [Google Scholar] [CrossRef]
  62. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A Density-Based Algorithm for Discovering Clusters in Large Spatial Databases with Noise. In Proceedings of the KDD; et al., E.S., Ed. AAAI Press, 1996, pp. 226–231.
  63. Ben-David, S.; Ackerman, M. Measures of Clustering Quality: A Working Set of Axioms for Clustering. In Proceedings of the Advances in Neural Information Processing Systems; et al., D.K., Ed. Curran Associates, Inc., 2008, Vol. 21.
  64. Delaunay, B. Sur la sphère vide. Bulletin de l’Acadeémie des Sciences de l’URSS. Classe des sciences mathématiques et na 1934, 1934, 793–800. [Google Scholar]
  65. Smith, N.A.; Tromble, R.W. Sampling uniformly from the unit simplex 2004.
  66. Uribe, L.; Bogoya, J.M.; Vargas, A.; Lara, A.; Rudolph, G.; Schütze, O. A Set Based Newton Method for the Averaged Hausdorff Distance for Multi-Objective Reference Set Problems. Mathematics 2020, 8. [Google Scholar] [CrossRef]
  67. Chen, W.; Ishibuchi, H.; Shang, K. Clustering-Based Subset Selection in Evolutionary Multiobjective Optimization. In Proceedings of the 2021 IEEE International Conference on Systems, Man, 2021, and Cybernetics (SMC); pp. 468–475. [CrossRef]
  68. Martín, A.; Schütze, O. Pareto Tracer: A predictor-corrector method for multi-objective optimization problems. Engineering Optimization 2018, 50, 516–536. [Google Scholar] [CrossRef]
  69. Beltrán, F.; Cuate, O.; Schütze, O. The Pareto Tracer for General Inequality Constrained Multi-Objective Optimization Problems. Mathematical and Computational Applications 2020, 25. [Google Scholar] [CrossRef]
  70. Schütze, O.; Cuate, O. The Pareto Tracer for the treatment of degenerated multi-objective optimization problems. Engineering Optimization 2024, pp. 1–26.
Figure 1. Pareto front representations of MOP (4) when using N equally distributed samples along the Pareto set.
Figure 1. Pareto front representations of MOP (4) when using N equally distributed samples along the Pareto set.
Preprints 154643 g001
Figure 2. Representation R 10000 x of the Pareto front of MOP (4) together with the two hypothetical outcomes A and B.
Figure 2. Representation R 10000 x of the Pareto front of MOP (4) together with the two hypothetical outcomes A and B.
Preprints 154643 g002
Figure 7. Filling with (a) and without (b) component detection for ZDT3. Note that points not in the PF are included if the component detection step is omitted (b).
Figure 7. Filling with (a) and without (b) component detection for ZDT3. Note that points not in the PF are included if the component detection step is omitted (b).
Preprints 154643 g007
Figure 8. RSG process for ZDT3. The starting set A y is taken from PlatEMO. We set N f = 10 , 000 .
Figure 8. RSG process for ZDT3. The starting set A y is taken from PlatEMO. We set N f = 10 , 000 .
Preprints 154643 g008
Figure 9. RSG process for DTLZ7. The starting set A y is taken from Pymoo. We set N f = 100 , 000 . Figure (l) shows the result of RSG with N = 300 over the same range for all variables, displaying uniformity in the solution.
Figure 9. RSG process for DTLZ7. The starting set A y is taken from Pymoo. We set N f = 100 , 000 . Figure (l) shows the result of RSG with N = 300 over the same range for all variables, displaying uniformity in the solution.
Preprints 154643 g009
Figure 10. RSG process for WFG2. The starting set A y is obtained using the PT and a non-dominance test. We set N f = 3 , 500 , 000 .
Figure 10. RSG process for WFG2. The starting set A y is obtained using the PT and a non-dominance test. We set N f = 3 , 500 , 000 .
Preprints 154643 g010
Figure 11. RSG process for CONV3. The starting set A y is obtained by sampling 100,000 points uniformly at random near the PS and then applying a nondominance test. We set N f = 10 , 000 .
Figure 11. RSG process for CONV3. The starting set A y is obtained by sampling 100,000 points uniformly at random near the PS and then applying a nondominance test. We set N f = 10 , 000 .
Preprints 154643 g011
Figure 12. RSG process for CONV3-4. The starting set A y is obtained by sampling 100,000 points uniformly at random near the PS and then applying a non-dominance test. We set N f = 10 , 000 .
Figure 12. RSG process for CONV3-4. The starting set A y is obtained by sampling 100,000 points uniformly at random near the PS and then applying a non-dominance test. We set N f = 10 , 000 .
Preprints 154643 g012
Figure 13. RSG process for CONV4-2F. The starting set A y is obtained by merging the final populations from 30 independent runs of NSGA-III, each run consisting of 400 generations with a population size of 500, and then applying a non-dominance test to the merged set. We set N f = 100 , 000 .
Figure 13. RSG process for CONV4-2F. The starting set A y is obtained by merging the final populations from 30 independent runs of NSGA-III, each run consisting of 400 generations with a population size of 500, and then applying a non-dominance test to the merged set. We set N f = 100 , 000 .
Preprints 154643 g013
Figure 14. The first part of the comparisons between RSG (first column), PlatEMO (second column), Pymoo (third column), and the filling step F of RSG (fourth column) are shown. A reference set of size 100 was used for bi-objective problems and 300 for three-objective problems. For bi-objective problems, all methods produced exactly 100 points. However, for three-objective problems, PlatEMO and Pymoo do not always yield exactly 300 points—for example, PlatEMO in WFG1, WFG2, and DTLZ7, and Pymoo in WFG1, WFG2, DTLZ7, CDTLZ2, and C2-DTLZ2, the last two being unobtainable (as far as the authors could determine) in Pymoo.
Figure 14. The first part of the comparisons between RSG (first column), PlatEMO (second column), Pymoo (third column), and the filling step F of RSG (fourth column) are shown. A reference set of size 100 was used for bi-objective problems and 300 for three-objective problems. For bi-objective problems, all methods produced exactly 100 points. However, for three-objective problems, PlatEMO and Pymoo do not always yield exactly 300 points—for example, PlatEMO in WFG1, WFG2, and DTLZ7, and Pymoo in WFG1, WFG2, DTLZ7, CDTLZ2, and C2-DTLZ2, the last two being unobtainable (as far as the authors could determine) in Pymoo.
Preprints 154643 g014
Figure 15. The second part of the comparisons between RSG (first column), PlatEMO (second column), Pymoo (third column), and the filling step F of RSG (fourth column) are shown. A reference set of size 100 was used for bi-objective problems and 300 for three-objective problems. For bi-objective problems, all methods produced exactly 100 points. However, for three-objective problems, PlatEMO and Pymoo do not always yield exactly 300 points—for example, PlatEMO in WFG1, WFG2, and DTLZ7, and Pymoo in WFG1, WFG2, DTLZ7, CDTLZ2, and C2-DTLZ2, the last two being unobtainable (as far as the authors could determine) in Pymoo.
Figure 15. The second part of the comparisons between RSG (first column), PlatEMO (second column), Pymoo (third column), and the filling step F of RSG (fourth column) are shown. A reference set of size 100 was used for bi-objective problems and 300 for three-objective problems. For bi-objective problems, all methods produced exactly 100 points. However, for three-objective problems, PlatEMO and Pymoo do not always yield exactly 300 points—for example, PlatEMO in WFG1, WFG2, and DTLZ7, and Pymoo in WFG1, WFG2, DTLZ7, CDTLZ2, and C2-DTLZ2, the last two being unobtainable (as far as the authors could determine) in Pymoo.
Preprints 154643 g015
Table 1. Indicator values I ( O , R ) for different indicators, the outcomes O { A , B } , and the representations R { R 100 x , R 100 y , R 10 , 000 x , R 10 , 000 y } of the Pareto front.
Table 1. Indicator values I ( O , R ) for different indicators, the outcomes O { A , B } , and the representations R { R 100 x , R 100 y , R 10 , 000 x , R 10 , 000 y } of the Pareto front.
G D 1 G D 2 I G D 1 I G D 2 I G D + Δ 1 Δ 2 Δ
I ( A , R 100 x ) 0.5118 0.7384 0.9084 0.9873 0.6423 0.9084 0.9873 1.3671
I ( B , R 100 x ) 0.0698 0.1002 0.4522 1.0744 0.3198 0.4522 1.0744 8.2024
I ( A , R 100 y ) 0.0684 0.0684 0.6835 0.7883 0.4833 0.6835 0.7883 1.2987
I ( B , R 100 y ) 0.0684 0.0684 2.5974 3.6765 1.8367 2.5974 3.6765 8.1341
I ( A , R 10,000 x ) 0.0028 0.0032 0.8968 0.9776 0.6341 0.8968 0.9776 1.3671
I ( B , R 10,000 x ) 0.0008 0.0010 0.4117 0.8792 0.2911 0.4117 0.8792 8.2024
I ( A , R 10,000 y ) 0.0007 0.0007 0.6835 0.7893 0.4833 0.6835 0.7893 1.3664
I ( B , R 10,000 y ) 0.0007 0.0007 2.5974 3.6767 1.8367 2.5974 3.6767 8.2018
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated