1 Introduction

Uncertainties exist in the real world, such as in the measurement system or final control element. The search for optimal solutions in the presence of uncertainties is often referred to as robust optimization in the literature [3, 4, 9, 11]. However, the term robust optimization can be used to encompass uncertainties in both the decision variables and the parameters of the problem, as discussed by Ben-Tal, El Ghaoui, and Nemirovski in their book on Robust Optimization [2]. This includes considering uncertainty in the input of the mathematical model of the optimization problem. In this article, our focus is specifically on uncertainties in the decision variables, referred to as perturbations within the context of robust optimization. On the other hand, uncertainties in the objective functions, which pertain to the output of the process, are referred to as noise within the context of noisy optimization. Figure 1 illustrates these definitions and the distinctions between noisy and robust optimization.

Fig. 1.
figure 1

Robust Optimization (perturbation in the decision variables/parameters) versus Noisy Optimization (noise in the function output).

Real optimization problems can have different types of uncertainties. These uncertainties can prevent the implementation of practical solutions if not taken into account. The designer (optimizer) therefore faces the challenge of finding solutions that are less sensitive to uncertainties. Optimization with uncertainty is a relatively new and rapidly growing research field that has gained significant attention in the past decade [10]. Some works with applications using optimization under uncertainties are [1, 7, 12, 17, 20, 25]. Examples of uncertainties include noise, model inaccuracies, time variations, measurement inaccuracies, disturbances, and other uncontrolled factors that can degrade the performance of the designed solutions [19].

Recently, Shaaban Sahmoud and Haluk Rahm [22] worked on noisy and dynamic optimization. We aim to fill a gap in the literature by conducting a study that tests evolutionary algorithms in the presence of both types of uncertainties; in decision variables and objective functions. In order to answer this research question, we extend a test function generator available in the literature for multi-objective optimization. The goal is to provide an examination of the behavior of these algorithms under such conditions.

The generator of benchmark problems from [18] uses a bottom-up approach and generates problems with various features, the objective space, and the decision space separately. It allows the creation of scalable and customizable problems for any number of objectives and Pareto fronts with different shapes and topologies, and can incorporate different features such as dissimilarity, robustness, and modality. The objective space can always be represented as a vector \(({\textbf {x}}_p, {\textbf {x}}_d)\), where \({\textbf {x}}_p\) responsible for the spatial location of the points in the objective space and \({\textbf {x}}_d\) governs the convergence.

The extension introduces different types of noise (Gaussian, Uniform, and Cauchy) for the objective functions. These noises have different properties, such as being additive or multiplicative, and can have different intensities. The noise intensities were taken from the work of [8]. Proposed intensities were implemented, along with a third intensity we specifically proposed for the Gaussian and Uniform noises.

Then, three evolutionary algorithms designed for robust or noisy optimization are selected: RNSGA-II, RMOEA/D, and C-RMOEA/D. The results of the tests show that the algorithms were not able to handle simultaneous uncertainties in decision variables and objective functions, leading to a degradation in the quality of the solutions compared to only one type of uncertainty (either perturbation or noise). Thus, we conclude that there is a need for designing algorithms specifically for handling simultaneous uncertainties.

In this work, the preliminary definitions and concepts such as uncertainties and robustness measures are presented in Sect. 2. In Sect. 3 presents the extension of the function generator proposed in this work. The results of the computational experiment, including a brief description of the three algorithms used and the experimental setup, are presented in Sect. 4. Finally, the conclusion of the work is presented in Sect. 5.

2 Preliminary Concepts

Some of the main concepts covered in this section include types of uncertainties and robustness measures. These concepts provide the foundation for the later discussions in the paper about robust and noisy multi-objective optimization.

2.1 Types of Uncertainties

There are several classifications of uncertainties in the literature. Three prevalent classifications of uncertainties are presented in Table 1. The parameter \(\delta \) represents uncertainties in the decision variables (perturbation), \(f({\textbf {x}}+\delta )\), or in the objective functions (noise), \(f({\textbf {x}})+\delta \).

Table 1. Classifications of types of uncertainties, according to [4, 13, 21].

According to [4], uncertainties are categorized based on their sources, with some sources of different types of uncertainties being identified in the design and optimization process. Ong et al. [21] consider which elements of the model are affected, such as the objective function, variables, or environment. Finally, Jin and Branke [13] classify uncertainties in optimization problems according to how they impact the optimization process.

We observed that there is an equivalence between the classifications of uncertainties mentioned in the literature. This equivalence is depicted in Table 2. To illustrate this equivalence, consider Type B, which classifies the source of uncertainty as production tolerance. For example, a final control element has uncertainty (a hypothetical example would be a drill that aims to drill 5 m, but may have a \({\pm }10\%\) error). This error in the final control element generates a perturbation in the decision variables, thus making it both Type II and Category II.

Table 2. Equivalence between classifications of uncertainties, according to [4, 13, 21].

2.2 Robustness Measurements

Mathematically, there are different ways to quantify the uncertainties classified above. One paper that defined robustness measures was by [4]. According to this work, uncertainties can basically be modeled deterministically, probabilistically, or possibilistically:

  1. 1.

    The deterministic type defines domains of uncertainty parameters. An example is the worst-case measure:

    $$\begin{aligned} R({\textbf {x}},\delta ) = \max _{\delta \in \varDelta }f({\textbf {x}},\delta ) \end{aligned}$$
    (1)
  2. 2.

    The probabilistic type defines probability measures that describe the probability by which a given event can occur. An example is the Monte Carlo integration followed by the mean:

    $$\begin{aligned} \hat{R}({\textbf {x}},\delta ) = \frac{1}{k} \sum _{i=1}^{k} [f({\textbf {x}}+\delta ^i)] \end{aligned}$$
    (2)
  3. 3.

    The possibilistic type defines fuzzy rules that describe the degree of adherence by which a given event may be acceptable. An example is the treatment of uncertainty sets using fuzzy logic and the theory of evidence [14].

3 Function Generator Extension - Robust Optimization And/or Noisy Optimization

A contribution of this work is the proposal of an extension to the function generator described in [18]. The main aspect of this extension is the integration of concepts from robust and noisy optimization. The Gaussian, Uniform, and Cauchy noise models present in the BBOB functions [8] are incorporated into the function generator, resulting in an optimization problem formulation with uncertainty in both the decision variables and the objective functions. Mathematically:

$$\begin{aligned} f({\textbf {x}},\delta _x, \delta _f) = f({\textbf {x}} + \delta _x) + \delta _f \end{aligned}$$
(3)

where \(\delta _x\) indicates the uncertainties in the decision variables and \(\delta _f\) indicates the uncertainties in the objective functions. In Eq. 3, the noise is considered additive, but it can also be multiplicative:

$$\begin{aligned} f({\textbf {x}},\delta _x, \delta _f) = f({\textbf {x}} + \delta _x) \times \delta _f \end{aligned}$$
(4)

The main objective of the proposed extension is to evaluate the behavior of algorithms in problems with uncertainty in both the decision variables and the objective functions. To date, the algorithms have only been tested in either robust optimization or noisy optimization. They were designed specifically for one type of uncertainty, so the question is whether they will perform well in the presence of two types of uncertainty simultaneously.

The test function used in this study is called GPD and is taken from [18]. This problem allows for changing the format of the Pareto front through a parameter p. The parameter p defines the norm (\(p \ge 1\)) or quasi-norm (\(0 < p < 1\)) used in the function \(h({\textbf {x}}) = || T({\textbf {x}}) ||_p\). If \(0 < p < 1\), the Pareto front is convex. If \(p=1\), it is a hyperplane, and if \(p > 1\), it is concave. In this article, it was defined as follows: \(p = 0.5\) or \(p = 1\) or \(p = 2\).

The focus of this work is to assess the robustness and noise tolerance of the algorithms. To achieve this, unrestricted and bi-objective problems were chosen. The problem GPD is evaluated according to the methodology described in [18].

As previously stated, this work utilizes three models of noise: Gaussian, Uniform, and Cauchy (\(\delta _f\) of Eqs. 3 and 4). These noises will be incorporated into the function generator. These noise models were adopted from the work of [8] (BBOB Functions). Finck et al. [8] defined two levels of intensity for the noise: moderate and severe. The mathematical description of the three noise models is presented below.

  • Gaussian Noise

    $$\begin{aligned} f_{GN}(f,\beta ) = f \times \exp (\beta \mathcal {N}(0,1)) \end{aligned}$$
    (5)

    where \(\beta \) controls the intensity of the noise and \(\mathcal {N}(0,1)\) is a random variable with a normal distribution with a mean of 0 and a variance of 1.

  • Uniform noise

    $$\begin{aligned} f_{UN}(f,\alpha ,\beta ) = f \times \mathcal {U}(0,1)^\beta \max \left( \begin{array}{c} 1,(\frac{10^9}{f + \epsilon })^{\alpha \mathcal {U}(0,1)} \end{array} \right) \end{aligned}$$
    (6)

    where \(\mathcal {U}(0,1)\) represents a random, uniformly distributed number in the interval (0, 1) and \(\alpha \) is a parameter that controls the intensity of the noise in conjunction with the parameter \(\beta \), resulting in two random factors. The first factor is uniformly distributed in the interval [0,1] for \(\beta = 1\). The second factor (\(\max \)) is greater than or equal to 1. Finally, the factor \(\epsilon \) is introduced to avoid division by zero, with a value of \(10^{-99}\) being adopted. The Uniform noise is considered to be more severe than Gaussian noise.

  • Cauchy Noise

    $$\begin{aligned} f_{CN}(f,\alpha ,\text {p}) = f + \alpha \left( \begin{array}{c} 1000 + \mathbb {I}_{\{\mathcal {U}(0,1)<\text {p}\}} \frac{\mathcal {N}(0,1)}{|\mathcal {N }(0,1)| + \epsilon } \end{array} \right) \end{aligned}$$
    (7)

    where \(\alpha \) defines the intensity of the noise and p determines the frequency of the noise disturbance. The value of \(\epsilon \) is \(10^{-199}\). This noise model has two important characteristics. First, only a small percentage of the function values are affected by noise. Second, the noise distribution is considered to be unusual, with outliers occurring from time to time.

The Gaussian and Uniform noises are multiplicative, so the intensity remains constant regardless of the range of values in the objective function. For example, if the maximum noise value is 5, the effect on the objective function will be to increase its value by 5 times, whether the value ranges from [0,1] or [0,100]. Cauchy’s noise differs from the previous two in that it is additive. In this case, the scale of the objective function must be considered. For example, if the range of values for one problem is [0,1] and for another problem is [0,100], and the noise peak has an intensity of 1, the intensity for the first problem will be \(100\%\) since the noise is additive, whereas, for the second problem, it will only be \(1\%\). Therefore, in this article, Eq. 7 will not include the 1000 summation term. In this article, the intensity values from [8] are kept.Footnote 1

4 Computational Experiment

In this section, the computational experiments will be presented. A brief description of the three selected algorithms and the testing setup will be provided. Subsequently, the solutions that the algorithms found for problems with perturbation and/or noise will be displayed. Finally, a discussion of the results will be made.

4.1 Algorithms

Typically, robustness measures (discussed in Sect. 2.2) are employed to account for uncertainties, resulting in a new objective function. As a result, existing algorithms in the literature are transformed into robust versions. For instance, the RMOEA/D employs MOEA/D [27] and the RNSGA-II employs NSGA-II [5], with the addition of Monte Carlo sampling in the uncertainty space. The average of this sampling is then calculated [6]. The RMOEA/D selects a set of vectors with the best weighting.

C-RMOEA/D [19] is a coevolutionary algorithm for robust multi-objective optimization that employs the worst-case minimization methodology and the decomposition/aggregation strategy. The decomposition approach facilitates the implementation of a coevolutionary approach [19]. The algorithm uses a competitive strategy between two populations of candidate solutions, with X being a set of vectors from the decision space and \(\varDelta \) being a set of perturbation vectors.

The RNSGA-II and RMOEA/D employ a probabilistic robustness measure, whereas C-RMOEA/D is a coevolutionary algorithm that uses a deterministic robustness measure. Other recent strategies for robust optimization or noisy optimization can be found at [15, 16, 26].

4.2 Experimental Setup

The Function Generator Extension (Sect. 3) will be utilized in the experiments. The Pareto front will be tested with convex, linear, and concave formats with simultaneous uncertainties. Problems with only perturbations or noise will be demonstrated using the concave Pareto front. The primary aim of this study is to evaluate algorithms in the presence of simultaneous uncertainties. To achieve this, the tests will be conducted with bi-objective and unconstrained scenarios. As mentioned earlier, the function generator is adaptable to these parameters. Tests with perturbations only (Robust Optimization) are conducted as shown in [18].

The following parameters were used for all tests: a population size of 210; a total of 24 decision variables; the stopping criterion is the maximum number of evaluations of the objective function, equal to 150,000. The RNSGA-II and RMOEA/D algorithms used 100 samples for averaging. The C-RMOEA/D used the PBIFootnote 2 (Penalty Boundary Intersection) aggregation function and a subpopulation size of 15. The RMOEA/D also utilized a subpopulation size of 15.

4.3 Results

The results will be presented in three parts: (i) robust optimization (with only perturbation); (ii) noisy optimization (with only noise); (iii) robust and noisy optimization (with both perturbation and noise - simultaneous uncertainties).

Robust Optimization. Figure 2 shows the results for \(\delta _x = 0.1\), which is considered a high perturbation as it represents an uncertainty of \(\pm 10\%\). The RNSGA-II and RMOEA/D (using a mean-based robustness measure) mapped the Robust Front. The algorithm based on a worst-case measure, C-RMOEA/D, also mapped the Robust Front. The results with convex and linear fronts are not displayed due to page limitations. The results are similar to those obtained with a concave front.

Fig. 2.
figure 2

Result with \(\delta _x = 0.1\).

Noisy Optimization. The C-RMOEA/D does not incorporate coevolution with noise only. Hence, it becomes the MOEA/D algorithm, without a second evolutionary cycle. The results were compared between MOEA/D and RMOEA/D and between NSGA-II and RNSGA-II. This comparison was made between the original algorithm without any uncertainty handling, and the algorithm that uses sampling followed by the mean metric.

Figure 3 presents the solutions obtained from RNSGA-II and NSGA-II. The results were similar for Gaussian and Uniform noise. The algorithms showed good convergence and dispersion from the Global Front with moderate intensity. However, they did not produce any mapping for severe intensity. Finally, both RNSGA-II and NSGA-II mapped the Global Front well for Cauchy noise (single additive noise) at both intensities. Due to page limitations, the results of MOEA/D and RMOEA/D will not be shown. These results were similar to those of NSGA-II and RNSGA-II from the Fig. 3.

Fig. 3.
figure 3

Result for function generator extension with noise for NSGAII and RNSGAII.

Robust and Noisy Optimization. Figure 4 displays the outcome of extending the function generator with Gaussian noise and \(\delta _x=0.1\). The three algorithms failed to converge for high-intensity noise. This outcome was predictable because the severe noise greatly impacted the original function. However, the C-RMOEA/D solutions exhibited a tendency towards the Robust Front. For moderate-intensity noise, the algorithms achieved convergence and good dispersion. Nonetheless, the RMOEA/D had some poor solutions, but, generally, the solutions converged. The same can be said for the C-RMOEA/D algorithm with concave PF.

Fig. 4.
figure 4

Result for function generator extension - Gaussian noise and \(\delta _x = 0.1\).

Figure 5 displays the results of extending the function generator with Uniform noise and \(\delta _x=0.1\). The analysis of results for Uniform noise is similar to Gaussian noise. The three algorithms did not converge for high-intensity noise. The algorithms converged and had good dispersion for moderate-intensity noise. However, the RMOEA/D and C-RMOEA/D had some poor solutions. RNSGA-II had some poor solutions for concave PF. As a result, it becomes evident that these algorithms with simultaneous uncertainties begin to deteriorate (even with small noise).

Fig. 5.
figure 5

Result for function generator extension - Uniform noise and \(\delta _x = 0.1\).

Figure 6 displays the result of extending the function generator with Cauchy noise and \(\delta _x=0.1\). This noise differs from the previous ones in that it is additive and not multiplicative. The C-RMOEA/D and RNSGA-II showed good convergence and dispersion for all PF formats and intensities. For high-intensity noise, the C-RMOEA/D had some solutions without convergence, and for the convex PF it failed to map the robust PF at the extremes. The RMOEA/D algorithm showed good convergence and dispersion for moderate-intensity noise. Only some final solutions were poor. Finally, RMOEA/D produced some solutions that mapped small parts of the PF to high-intensity noise. Therefore, the solutions did not have good dispersion and most of them did not converge. The worst convergence result was for convex PF.

Fig. 6.
figure 6

Result for function generator extension - Cauchy noise and \(\delta _x = 0.1\).

As stated in the previous results, the original function is severely contaminated by high Gaussian and Uniform noise. Therefore, we created an intermediate-intensity noise for further testing. The value was \(\beta =0.5\) for Gaussian noise. For Uniform noise, the parameters were \(\beta =0.5\) and \(\alpha =0.2*(0.49+1/D)\). The effect of intermediate-intensity noise showed that the trend of the original function remained. Consequently, EAs must be able to eliminate the effect of noise at this intensity.

For this test set (intermediate-intensity), 30 simulations were performed for each front and algorithm. The IGD (Inverted Generational Distance) evaluation metric [24] was calculated for each simulation, using the Robust Front as the reference set. The solution set that performed best according to the IGD metric (closest to zero) was chosen, and Fig. 7 shows these results for tests of the function generator extension with intermediate noise intensity and \(\delta _x=0.1\).

The algorithms did not achieve good convergence, regardless of the geometry of the PF (convex, linear, or concave) or the type of noise (Gaussian or uniform). This indicates that, when the noise has a greater effect (intermediate intensity), the tested algorithms were unable to eliminate its effect with perturbation.

Specifically, RNSGA-II started to converge at the extremes and in the middle of the PF for Gaussian noise and convex PF. The same happened for linear PF, but with a smaller number of solutions. Some RMOEA/D solutions also converged in the central region of the robust PF (linear PF and Gaussian noise). These algorithms did not show convergence when the PF was concave. The solutions of the C-RMOEA/D algorithm showed a trend of robust PF shapes for Gaussian noise, mainly the convex PF.

Uniform noise had a more significant detrimental effect (compared to Gaussian) on the original function. The results (Fig. 7) were consistent for all PF formats and Uniform noise. The algorithms did not converge. The C-RMOEA/D showed a tendency towards solutions on the border.

Fig. 7.
figure 7

Result for the extension of the function generator - intermediate noise and \(\delta _x = 0.1\) (perturbation in the objective functions and decision variables).

Table 3 contains the average of the IGD metric for the 30 simulations, considering the Robust Front as a reference. The previous results qualitatively demonstrated that the algorithms were not able to deal with simultaneous uncertainties. The quantitative results (Table 3) reinforce this conclusion. The average IGD was high for all algorithms (keep in mind that a lower IGD value indicates better performance), except for the CRMOEA/D algorithm with convex shape. The CRMOEA/D can be considered the superior algorithm regardless of the format of the Pareto Front. Only for the Concave Front and Uniform noise, the RNSGA-II algorithm showed slightly better performance.

Table 3. Results of the IGD metric for intermediate intensity noise and \(\delta _x = 0.1\).

Table 3 displays the variability of the IGD metric across 30 simulations, using the Robust Front as the reference, as indicated by the standard deviation. The RNSGA-II algorithm demonstrated superiority in this test across different noise and Border Shape settings, with consistently low standard deviation. The standard deviation test indicates that the RNSGA-II algorithm exhibited minimal variations in IGD across the 30 simulations. The CRMOEA/D demonstrated superior performance compared to the average IGD. However, it is worth noting that the standard deviation was high, suggesting the presence of simulations with exceptionally poor results. Conversely, the RMOEA/D exhibited unsatisfactory outcomes in terms of both the mean and standard deviation.

4.4 Discussion

Algorithms with a robustness measure had no difficulties in finding the Robust Front when the problem has only perturbation.

The evolutionary algorithms were able to eliminate the effect of noise for moderate intensity (noisy optimization), including algorithms without any robustness measure (NSGA-II and MOEA/D). However, the algorithms did not achieve convergence when the effect of Gaussian and Uniform noise was high (severe intensity). One possible explanation is that the intensity of these noises severely deteriorates the original function, causing the function’s trend to be lost.

Cauchy noise is different from the previous ones. This noise has peaks of high intensity. As a result, the trend of the original function remains even with severe Cauchy noise. The RMOEA/D algorithm had difficulties in mapping the Global Front to severe Cauchy noise. The reason is that noise has contaminated the average value (noise spikes have increased the average). The original MOEA/D algorithm managed to map the Global Front for this noise. One explanation is that the evolutionary process eliminated the effect of noise on solutions. The NSGA-II and RNSGA-II algorithms achieved excellent mapping of the Global Front for Cauchy noise. This shows that the characteristics of the NSGA-II (archive, elitism, dominance) eliminated the effect of this noise, even when the average was used (RNSGA-II).

Finally, for robust noisy problems, the results demonstrate that the tested algorithms had difficulties in this new class of problems. Some solutions started to deteriorate with moderate noise and perturbation. We performed some tests by increasing the noise intensity and observed that the algorithms presented greater difficulty for problems with simultaneous uncertainties. Thus, we created an intermediate noise. The algorithms did not solve the problems with perturbation and intermediate noise.

To confirm the conclusion presented in this discussion, results are provided for Gaussian noise with a value of \(\beta = 0.25\), as depicted in Fig. 8. The RNSGA-II algorithm successfully mapped the Global Frontier when the problem contained only noise. However, when both perturbation and noise were present, the majority of algorithm solutions did not converge.

Fig. 8.
figure 8

Result for the extension of the function generator - \(\beta =0.25\) Gaussian Noise).

5 Conclusion

In the real world, simultaneous uncertainties are common. A practical example is a drilling process, where the sensor measuring the depth is subject to uncertainty (resulting in noise in the objective function), and the drill (actuator) is also subject to uncertainty (resulting in perturbations in the decision variables). The authors identified a deficiency in the field of optimization with uncertainty, as no previous research had dealt with simultaneous uncertainties.

The results indicated that the algorithms tested (RNSGA-II, RMOEA/D, and CRMOEA/D) were unable to effectively handle simultaneous uncertainties. Consequently, there is a need to formulate algorithms that are capable of handling both robust and noisy environments concurrently.