Accelerating Convergence Towards the Optimal ... - Semantic Scholar

Report 2 Downloads 128 Views
Accelerating Convergence Towards the Optimal Pareto Front Mohsen Davarynejad∗ , Jafar Rezaei∗ , Jos Vrancken∗ , Jan van den Berg∗ and Carlos A. Coello Coello† ∗ Faculty

of Technology, Policy and Management, Delft University of Technology, the Netherlands Email: {M.Davarynejad; J.Rezaei; J.L.M.Vrancken; J.vandenBerg}@tudelft.nl † CINVESTAV-IPN, Departmento de Computaci´ on (Evolutionary Computation Group) Av. IPN No. 2508, Col. San Pedro Zacatenco, M´exico D.F. 07300, M´exico Email: [email protected]

Abstract—Evolutionary algorithms have been very popular optimization methods for a wide variety of applications. However, in spite of their advantages, their computational cost is still a prohibitive factor in certain real-world applications involving expensive (computationally speaking) fitness function evaluations. In this paper, we depart from the observation that nature’s survival of the fittest is not about exact measures of fitness; rather it is about rankings among competing peers. Thus, by exploiting this natural tolerance for imprecision, we propose here a new, fuzzy granules-based approach for reducing the number of necessary function calls involving time consuming real-world problems. Our proposed approach is compared with respect to the standard NSGA-II, using the Set Coverage, Hypervolume and Generational Distance performance measures. Our results indicate that our proposed approach is a very promising alternative for dealing with multi-objective optimization problems involving expensive fitness function evaluations.

reduce the computational cost of MOEAs (see [25]). In this paper, we aim to contribute to this area by introducing a fuzzy granules-based approach for reducing the number of objective function evaluations required by a MOEA. The remainder of this paper is organized as follows. Section II provides some basic multi-objective optimization concepts. The previous related work is discussed in Section III. Section IV presents the approach proposed in this paper. To illustrate the efficiency of the proposed method, the performance results on ZDT1-6 test problem is presented in Section V. Finally, in Section VI we provide our conclusions.

I. I NTRODUCTION Optimization using metaheuristics has become a very popular research topic in the last few years. Real-world problems, however, frequently have two or more (possibly conflicting) objectives that we aimed to optimize at the same time. Such problems are called multi-objective and have been intensively studied using metaheuristics (particularly, evolutionary algorithms) in the last few years [2]. As opposed to single-objective optimization problems in which we aim to find a single optimum solution, in multiobjective optimization problems (MOPs) the notion of optimality changes, since there is normally no single solution that is the best for all the criteria. The aim in this case is then to find a set of solutions for which no objective can be improved without worsening another. This set of solutions is known as the Pareto optimal set and their vectors are said to be nondominated. When plotted in objective function space, these solutions are collectively known as the Pareto front. A wide variety of multi-objective evolutionary algorithms (MOEAs) have been proposed since the inception of this field in the mid-1980s [7], [2]. However, MOEAs are known to be computationally expensive, since they normally require a high number of objective function evaluations in order to produce a reasonably good approximation of the Pareto front of the problem being solved. Nevertheless, relatively little research has been reported so far on the development of techniques that

minimize f⃗(⃗x) := [f1 (⃗x), f2 (⃗x), . . . , fk (⃗x)]

II. BASIC C ONCEPTS We are interested in solving problems of the type1 : (1)

subject to:

where ⃗x = variables, fi functions and the constraint

gi (⃗x) ≤ 0

i = 1, 2, . . . , m

(2)

hi (⃗x) = 0

i = 1, 2, . . . , p

(3)

T

[x1 , x2 , . . . , xn ] is the vector of decision : IRn → IR, i = 1, ..., k are the objective gi , hj : IRn → IR, i = 1, ..., m, j = 1, ..., p are functions of the problem.

To describe the concept of optimality in which we are interested, we will introduce next a few definitions. Definition 1. Given two vectors ⃗x, ⃗y ∈ IRk , we say that ⃗x ≤ ⃗y if xi ≤ yi for i = 1, . . . , k, and that ⃗x dominates ⃗y (denoted by ⃗x ≺ ⃗y ) if ⃗x ≤ ⃗y and ⃗x ̸= ⃗y . Definition 2. We say that a vector of decision variables ⃗x ∈ X ⊂ IRn is nondominated with respect to X , if there does not exist another ⃗x′ ∈ X such that f⃗(⃗x′ ) ≺ f⃗(⃗x). Definition 3. We say that a vector of decision variables ⃗x∗ ∈ F ⊂ IRn (F is the feasible region) is Pareto-optimal if 1 Without

loss of generality, we will assume only minimization problems.

it is nondominated with respect to F. Definition 4. The Pareto Optimal Set P ∗ is defined by: P ∗ = {⃗x ∈ F |⃗x is Pareto-optimal} Definition 5. The Pareto Front PF ∗ is defined by: PF ∗ = {f⃗(⃗x) ∈ IRk |⃗x ∈ P ∗ } We thus wish to determine the Pareto optimal set from the set F of all the decision variable vectors that satisfy (2) and (3). Note however that in practice, not all the Pareto optimal set is normally desirable (e.g., it may not be desirable to have different solutions that map to the same values in objective function space) or achievable. III. P REVIOUS R ELATED W ORK Evolutionary algorithms usually require such a large number of function calls that this frequently makes them computationally prohibitive in some real-world applications. When dealing with MOPs, this issue becomes more critical, because more objectives are involved and this multiplies the computational cost, while also making the search more difficult. For dealing with expensive objective functions, it is relatively common to rely on approximate models that allow us to simplify the representation of real-world complex behaviors. 2 Approximation techniques may estimate each of the individuals’ fitness value on the basis of previously observed objective function values of neighboring individuals. A wide range of approximation and meta-model techniques have been adopted in combination with evolutionary algorithms, including Kriging [24], artificial neural networks [26], and radial-basis-function networks [18]. Other authors have adopted fitness inheritance [21], cultural algorithms [15] and other fitness approximation techniques [14] for the same purpose. Next, we will briefly review the most representative work on the use of mechanisms for handling expensive objective functions with MOEAs reported in specialized literature. Fitness inheritance, a popular class of fitness approximation method, was originally introduced by Smith et al. [27] and is a very simple technique that works as follows: when assigning fitness to an individual, some times the objective function is evaluated as usual, but the rest of the time, the fitness assigned to the individual is the average (or a weighted average) of the fitness of its parents. This fitness assignment scheme operates based on the assumption of similarity between an offspring and its parents. Clearly, fitness inheritance cannot be applied all the time, since some true fitness function values are required in order to obtain enough information to guide the search. This approach uses a parameter called inheritance proportion, which regulates how many times the fitness has to be approximated. Very few authors have reported the use of fitness inheritance in MOPs. Ducheyne et al. [11] tested 2 This is based on the assumption that approximate models require small computational resources compared to the cost of complex simulations, which is normally the case when considering real-world problems.

the performance of both average and weighted average fitness inheritance approaches and concluded that the usefulness of this technique was limited to cases in which the Pareto front is convex and continuous. Ducheyne et al. [10] also concluded that for non-convex Pareto fronts, fitness inheritance produces a slower convergence to the true Pareto front than when the approach is not adopted. Other authors, however, have successfully applied fitness inheritance to more complicated test problems having non-convex Pareto fronts (see [21]). Another approach for dealing with expensive objective functions is based on learning and interpolation from representative small datasets of the true objective functions values in the desired design space which is known as functional approximation [14]. Function approximation methods provide a mapping between design space and objective functions space that may be multi-dimensional. The accuracy of these models depends greatly on the number of sample data points used and their location in the multi-dimensional space. Some examples of this sort of approach are the following: the response surface methodology that uses low-order polynomials and the least square estimations [17], [13], [12] and Gaussian processes (also known as Kriging) that build probability models by exploiting information recorded and use them to estimate the function values of new candidate solutions [3]. Artificial Neural networks (ANNs) can also be used for dealing with expensive objective functions. In fact, ANNs can be considered one of the best approaches to approximate a generic ℜn ⇒ ℜm function3 , where n and m represent the number of design variables and number of objectives, respectively. Although nonlinear interpolation can be used, it is shown that with a number of design variables higher than 10, the interpolation problem becomes almost not tractable [20]. ANNs are successfully used for building approximate models in a number of complex multiplicative optimization problems. As an example, in [1], a generic supersonic aircraft configuration with two main goals (maximization of the total range of the aircraft and minimization of the ground sonic boom) and a number of buildability and mission constraints (such as structural integrity of the aircraft, take-off and landing field length) is optimized using ANNs to generate inexpensive surrogates. The approximation is used only where this is warranted. Using Latin Hypercube Sampling (LHS), 300 sample data were generate via CFD (Computational Fluid Dynamics) simulation are fitted using a single hidden layer perceptron with sigmoid activation functions to provide a general nonlinear higher fidelity model. In another study, Poloni et al. [20] used a combination of GAs and ANNs with a modified backpropagation algorithm, and a local search method to optimize the design of a sailing yacht fin keel which is a complex design problem in fluid dynamics. The ANN acted as a model for 3D Navier-Stokes simulation of the fin keel while cruising. For more information on approaches for dealing with ex3 If they are provided with sufficient structural complexity and a rich training data set.

pensive objective functions in the context of multi-objective optimization, interested readers should refer to [25]. A. Final Remarks on Fitness Approximation In most of the fitness approximation models currently available, the main problem is the lack of sufficient training data and hence the failure to reach a model with sufficient approximation accuracy. Since the evaluation of the original fitness function, in many practical problems, is obtained by some sort of analysis (i.e., fluid mechanics analysis, thermodynamic analysis) that is computationally expensive, the approximate model may be of low fidelity. Furthermore, if the training data does not cover the full domain range, large errors may occur due to extrapolation. Errors may also occur when the set of training points is not sufficiently dense and uniform. Here, we adopt the concept of information granulation as an attempt to address these difficulties. IV. A DAPTIVE F UZZY F ITNESS G RANULATION Granular computing is regarded as the processing of granules of information that are aggregated due to their indistinguishability, similarity, proximity or functionality in some context [30]. It is a vehicle for handling information, as well as a lack of it (uncertainty), at a level of coarseness that can solve problems appropriately and efficiently [6]. In problems with incomplete, uncertain or vague information, the practical necessity; and in problems with huge detailed information, the simplicity are the main reasons of popularity, respectively, of granular computing. It is widely used in many fields including interval analysis, Dempster-Shafer theory of belief functions, cluster analysis, optimization and problem solving [23], machine learning, bioinformatics, among other fields [19]. The concept of information granulation was proposed by Zadeh [31] (in the context of fuzzy set theory) as a technique by which a class of points (objects) is partitioned into granules. The fuzziness of granules and their attributes is characteristic of the ways by which human concepts and reasoning are formed, organized and manipulated. The concept of a granule is more general than that of a cluster, potentially giving rise to several conceptual structures in various fields of science as well as mathematics. In the present paper, with the aim to reducing the computational cost of MOPs, the concept of information granulation and approximation in the context of rough set theory is studied to exploit the natural tolerance of EAs in fitness function computations. Nature’s survival of the fittest is not about exact measures of fitness; rather it is about rankings among competing peers. By exploiting this natural tolerance for imprecision and aiming to exploit this uncertainty [16], optimization performance can be preserved by computing fitness only selectively and only to keep this ranking among individuals in a given population. In the proposed algorithm, a pool of solutions with exact fitness values are maintained. Based on the maximum similarity of a new candidate solution to this pool, the fitness of individuals will be either approximated or calculated explicitly.

Fig. 1: A number of gaussian granules with different widths in a 2-D solution space. Once a new individual is sufficiently similar to a granule in the granule pool, then that granules’ fitness is used instead as a crude estimate. Otherwise, that individual is added to the pool as a new fuzzy granule. Each granules’ radius of influence is determined based on equation (7).

If a new individual is sufficiently similar to a known fuzzy granule, then that granules’ fitness is used instead as a crude estimate. Otherwise, that individual is added to the pool as a new granule. In this fashion, regardless of the competitions’ outcome, the fitness of the new individual is always a physically realizable one, even if it is a crude estimate and not an exact measurement. The pool size as well as each granules’ radius of influence depends on the utility of each granule [5].Furthermore, to prevent the pool from growing too large, pool members are competing for survival and members with lower life index are gradually replaced by new granules. By splitting up the pool into two parts, the new granules are given a chance to survive a number of steps [4]. A. Algorithm’s Structure The preceding section provided a general overview of our approach. Going in more detail now, the algorithm’s computation steps are as follows: Step 1: Create a random parent population P1 = {X11 , X21 , . . . , Xj1 , . . . , Xt1 } of decision vectors, where, Xji = {xij,1 , xij,2 , . . . , xij,r , . . . , xij,m } is the jth individual in the ith generation, xij,r the rth component of Xji , m the number of components of decision vector and t is the population size. Step 2: Define a multi-set G of fuzzy granules (Ck , σk , Lk ) according to G = {(Ck , σk , Lk )|Ck ∈ ℜm , σk ∈ ℜ, Lk ∈ ℜ, k = 1, . . . , l}. G is initially empty (i.e., l = 0). Ck is an m-dimensional vector of centers, σk is the width of membership functions (WMFs) of the kth fuzzy granule, and Lk is the granule’s life index. A number of granules with different widths are shown in Figure 1. Step 3: • •

Choose the phenotype of chromosomes, Xji , as the center of granules, Ck . Rank P1 and goto step 8.

Step 4: Define the membership µk,r of each xij,r to each granule member by a Gaussian similarity neighborhood function according to (

µk,r xij,r

)

( = exp

( )2 ) − xij,r − ck,r 2

(σk )



, k = 1, 2, . . . , l , (4)

where l is the number of fuzzy granules. Step 5: Compute the average similarity of the new decision vector Xji = {xij,1 , xij,2 , . . . , xij,r , . . . , xij,m } to each granule Gk using equation (5) m ∑

µj,k =

often non-commensurable and conflicting, dominancebased ranking is used. The spread of granules grow as their rank among granule members increases, with a rate of grσ . Here, grσ is set to 0.1 and σmin ∈ {2n |n ∈ Z}. Goto step 4.

( ) µk,r xij,r

r=1

m

(5)

Step 6: Either calculate the exact fitness value of Xji or estimate it by associating it to one of the granules in the pool in case there is a granule in the pool with similarity value higher than a predefined threshold, i.e.,  max {µj,k } > θi , ( i ) f (Ck ) if k∈{1,2,...,l} f Xj = (6) f (X i ) otherwise. j where f (Cx ) is the fitness function value of the fuzzy granule and f (Xji ) is the real fitness calculation of the individual. Remark: θi is a predefined (time-varying) threshold that controls the minimum similarity a solution has to have with a pool member to be approximated. Here, θi is considered as a constant value for all simulations, and is set to 0.9. In general, as the population matures steadily, the algorithm needs to be more selective (to calculate the exact fitness more often), suggesting the need for a gradual increase of θi . Alternatively, if max {µj,k } < θi k∈{1,2,...,l}

Xji

is chosen as a newly created granule. Step 7: If the population size is not completed, repeat Steps 4 to 7. Step 8: When termination/evolution control criteria are not met: • Create offspring population. • Rank the granule pool. • Assign σk based on equation (7). σk = σmin ∗ ((1 − grσ ) + grσ ∗ rank(k))

(7)

where rank(k) is the rank of the granule k among the granule set, and σmin ∈ R>0 is a proportional constant that defines the minimum spread of granules. σmin is a problem dependent design parameter. Remark: σk , the distance measurement parameter that controls the degree of similarity between two individuals, controls the radius of influence of each granule. Instead of drawing the radius directly from the fitness (as in the single-objective optimization case [4]), as objectives are

B. Controlling the granule pool length and protecting new pool members (innovation) through speciation As the evolutionary procedures are applied, it is inevitable that new granules are generated and added to the pool. Depending on the complexity of the problem, the size of this pool can be excessive and become a computational burden itself. To prevent such unnecessary computational effort, a life index is introduced in order to appropriately decrease the size of the pool. In other words, it is better to remove granules that do not win new individuals, thereby producing a bias against individuals that have low fitness and were likely produced by a failed mutation attempt. Lk is initially set to 0 and subsequently updated as below, { Lk + M if k = K , Lk = (8) Lk otherwise , where M is the life reward of the granule and K is the index of the winning granule for each individual at generation i. Here, M is set at 1. At each table update, only the NG granules with the highest Lk index are kept, and the others are discarded. In [6], an example has been provided that illustrates the competitive granule pool update law. Adding a new granule to the granule pool and assigning a life index to it, is a simple way of controlling the size of the granule pool, since the granules with the lowest life index will be removed from the pool. However, it may happen that the new granule is removed, even though it was just inserted into the pool. In order to prevent this, the pool is split into two parts with sizes εNG and (1 − ε)NG . The first part is a FIFO (First In, First Out) queue and new granules are added to this part. If it grows above εNG , then the top of the queue is moved to the other part. Removal from the pool takes place only in the (1−ε)NG part. In this way, new granules have a good chance to survive a number of steps. In all of the simulations that are conducted here, ε is set to 0.1. V. N UMERICAL RESULTS In order to validate our proposed approach, we adopted the Zitzler-Deb-Thiele (ZDT) test problems [33] and compared our results with respect to those obtained with the standard NSGA-II [8]. The following parameters were adopted for our experiments: • • • • •

Population size = 50. Crossover rate = 0.9. (SBX) Binary tournament selection. Mutation rate of 1/L, L = number of decision variables. Distribution indices for crossover and mutation: ηc = 20 and ηm = 20.

Problem ZDT1 ZDT2 ZDT3 ZDT4 ZDT6

σmin 2−4 2−5 2−5 2−6 2−5

NG 100 100 100 100 100

Reference point [1.1, 3.5] [1.1, 5.0] [1.1, 6.0] [1.1, 140] [1.1, 9.0]

Problem ZDT1 ZDT2 ZDT3 ZDT4 ZDT6

TABLE I: AFFG-NSGA-II utilized parameter values and reference points used for calculating IH . Problem ZDT1 ZDT2 ZDT3 ZDT4 ZDT6

AFFG-NSGA-II mean, σ 0.010165, 0.005744 0.018143, 0.008509 0.098656, 0.022421 11.160124, 4.239201 0.768217, 0.143028

NSGA-II mean, σ 0.102095, 0.029859 0.716683, 0.365823 0.236176, 0.048486 20.191547, 11.658247 1.328310, 0.224595

AFFG-NSGA-II mean, σ 1.000000, 0.000000 1.000000, 0.000000 0.995745, 0.023307 0.613805, 0.455574 0.891819, 0.134759

NSGA-II-AFFG mean, σ 0.000000, 0.000000 0.000000, 0.000000 0.003401, 0.018630 0.324147, 0.427815 0.033209, 0.081798

TABLE IV: Mean and standard deviation of the SC performance measure.

3.5

TABLE II: Mean and standard deviation of the GD performance measure.

3

2.5

For assessing our results we adopted three performance measures: (1) Generational Distance (GD) [28], which measures how far the given solutions are, on average, from the true Pareto front, (2) the Hypervolume indicator IH (also known as Lebesgue measure or S-metric) [32], which measures the volume of the dominated portion of the objective space which is enclosed by the reference set and (3) Set Coverage (SC) [33], which measures the percentage of solutions from one algorithm that are covered by the solutions of the other. To measure the Hypervolume, a single reference point, R = r ∈ ℜn was considered in all cases. This point corresponds to the worst value in each dimension of the fronts. The reference values we used here are given in Table I. The performance measures to assess the results are presented in Tables II, III and IV. The measures are evaluated by conducting 30 independent runs per test problem per algorithm. Each run is restricted to 1,000 fitness function evaluations. Each table displays the average and standard deviation of each of the performance measures. Figures 2 to 11 present results of 30 independent runs of the standard NSGA-II and the AFFG-NSGA-II, respectively, adopting the test problems ZDT1, ZDT2, ZDT3, ZDT4 and ZDT6 (ZDT5 is a binary problem and was, therefore, omitted here), with a budget of only 1,000 fitness function evaluations. Each color corresponds to single run. The results clearly show that the proposed AFFG approach outperforms the standard NSGA-II. According to the

2

1.5

1

0.5

0

0

0.2

0.4

0.6

0.8

1

Fig. 2: 30 independent runs of NSGA-II for the ZDT1 test problem using 1,000 real fitness function evaluations.

3.5

3

2.5

2

1.5

1

Problem ZDT1 ZDT2 ZDT3 ZDT4 ZDT6

AFFG-NSGA-II mean, σ 3.408204, 0.052768 4.524421, 0.110119 6.106243, 0.198963 108.878924, 10.460062 3.229885, 0.896935

NSGA-II mean, σ 2.689226, 0.164173 2.227951, 0.350130 4.516725, 0.267211 100.619288, 9.466605 1.178803, 0.176150

TABLE III: Mean and standard deviation of the IH performance measure.

0.5

0

0

0.2

0.4

0.6

0.8

1

Fig. 3: 30 distinct runs of AFFG-NSGA-II over ZDT1 problem using 1,000 real fitness function evaluations.

5

6

4.5 5 4 4

3.5 3

3

2.5 2

2 1.5

1

1 0 0.5 0

0

0.2

0.4

0.6

0.8

1

Fig. 4: 30 independent runs of NSGA-II over ZDT2 problem using 1,000 real fitness function evaluations.

5

−1

0

0.2

0.4

0.6

0.8

1

Fig. 6: 30 independent runs of NSGA-II over ZDT3 problem using 1,000 real fitness function evaluations.

6

4.5 5

4 3.5

4

3 3 2.5 2

2

1.5 1 1 0

0.5 0

0

0.2

0.4

0.6

0.8

1

Fig. 5: 30 distinct runs of AFFG-NSGA-II over ZDT2 problem using 1,000 real fitness function evaluations.

Wilcoxon rank-sum test, the results of our proposed approach are better with a significance level of 5%. To further investigate the convergence speed of the proposed approach, in Figure 12, the changes in hypervolume metric is poltted against the number of fitness function evaluations, for the ZDT1 problem. VI. C ONCLUSIONS AND F UTURE W ORK By combining the concepts of survival of the fittest and fuzzy granulation, which enables a faster convergence without degrading the estimated set of solutions, this paper presents an approach to speed up convergence towards the Pareto optimal front of multi-objective optimization problems. With the proposed approach, we can exploit the information obtained from our previous objective function evaluations. Our

−1

0

0.2

0.4

0.6

0.8

1

Fig. 7: 30 distinct runs of AFFG-NSGA-II over ZDT3 problem using 1,000 real fitness function evaluations.

results indicate that the proposed approach is very promising, since it can achieve a faster convergence than the standard NSGA-II in the test problems adopted. However, a more thorough validation is still required (adopting other problems such as the DTLZ test problems [9]). It is also desirable to perform comparisons with respect to other fitness approximation methods such as curve fitting, fitness inheritance and artificial neural networks. As part of our future work, we are interested in studying the effect of the number of granules on the convergence rate. Additionally, in order to further test the robustness of our proposed approach, we want to study its sensitivity to its parameters and its scalability when increasing the number of decision variables and objectives. Adaptively changing θi and being more selective as population matures

140

9 8

120

7 100 6 80

5

60

4 3

40 2 20

0

1

0

0.2

0.4

0.6

0.8

1

Fig. 8: 30 independent runs of NSGA-II over ZDT4 problem using 1,000 real fitness function evaluations.

140

0 0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

Fig. 10: 30 independent runs of NSGA-II over ZDT6 problem using 1,000 real fitness function evaluations.

9 8

120

7 100 6 80

5 4

60

3 40 2 20 1 0

0

0.2

0.4

0.6

0.8

1

Fig. 9: 30 distinct runs of AFFG-NSGA-II over ZDT4 problem using 1,000 real fitness function evaluations.

(to calculate the exact fitness more often), is indeed part of our ongoing research. Finally, we wish to apply our proposed approach to real-world problems in the field of network design problems [29] and supplier selection [22]. ACKNOWLEDGMENT This research received funding from the European Community’s Seventh Framework Programme (FP7/2007-2013) under grant agreement no. INFSO-ICT-223844, the Next Generation Infrastructures Research Program of Delft University of Technology and the Mexican CONACyT Project No. 103570. R EFERENCES [1] J.J. Alonso, P. LeGresley, and V. Pereyra. Aircraft design optimization. Mathematics and Computers in Simulation, 79(6):1948–1958, 2009.

0 0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

Fig. 11: 30 distinct runs of AFFG-NSGA-II over ZDT6 problem using 1,000 real fitness function evaluations.

[2] C.A. Coello Coello, G.B. Lamont, and D.A. Van Veldhuizen. Evolutionary Algorithms for Solving Multi-Objective Problems. Springer, New York, second edition, September 2007. ISBN 978-0-387-33254-3. [3] S. D’Angelo and E.A. Minisci. Multi-objective evolutionary optimization of subsonic airfoils by kriging approximation and evolutionary control. In 2005 IEEE Congress on Evolutionary Computation (CEC’2005), volume 2, pages 1262–1267, Edinburg, Scotland, September 2005. [4] M. Davarynejad, C. W. Ahn, J. L. M. Vrancken, J. van den Berg, and C. A. Coello Coello. Evolutionary hidden information detection by granulation-based fitness approximation. Applied Soft Computing, 10(3):719–729, 2010. [5] M. Davarynejad, J. Vrancken, J. van den Berg, and C.A. Coello Coello. A Fitness Granulation Approach for Large-Scale Structural Design Optimization. In Raymond Chiong, Thomas Weise, and Zbigniew Michalewicz, editors, Variants of Evolutionary Algorithms for RealWorld Applications. Springer-Verlag, Berlin, 2011. [6] Mohsen Davarynejad. Fuzzy Fitness Granulation in Evolutionary Algorithms for Complex Optimization. Master’s thesis, Ferdowsi University of Mashhad, June 2007.

3.6 3.4

Hypervolume

3.2 3 2.8 2.6 2.4 2.2 2

NSGA−II AFFG−NSGA−II 0

1000

2000 3000 Number of fitness evaluation

4000

5000

Fig. 12: Convergence of hypevolume metric over ZDT1 problem (30 distinct runs).

[7] K. Deb. Multi-Objective Optimization using Evolutionary Algorithms. John Wiley & Sons, Chichester, UK, 2001. ISBN 0-471-87339-X. [8] K. Deb, A. Pratap, Sameer Agarwal, and T. Meyarivan. A Fast and Elitist Multiobjective Genetic Algorithm: NSGA–II. IEEE Transactions on Evolutionary Computation, 6(2):182–197, April 2002. [9] K. Deb, L. Thiele, M. Laumanns, and E. Zitzler. Scalable Test Problems for Evolutionary Multiobjective Optimization. In Ajith Abraham, Lakhmi Jain, and Robert Goldberg, editors, Evolutionary Multiobjective Optimization. Theoretical Advances and Applications, pages 105–145. Springer, USA, 2005. [10] E.I. Ducheyne, B. De Baets, and R.R. De Wulf. Fitness inheritance in multiple objective evolutionary algorithms: A test bench and real-world evaluation. Applied Soft Computing, 8(1):337–349, 2008. [11] E.I. Ducheyne, B. De Baets, and R. De Wulf. Is Fitness Inheritance Useful for Real-World Applications? In Carlos M. Fonseca, Peter J. Fleming, Eckart Zitzler, Kalyanmoy Deb, and Lothar Thiele, editors, Evolutionary Multi-Criterion Optimization. Second International Conference, EMO 2003, pages 31–42, Faro, Portugal, April 2003. Springer. Lecture Notes in Computer Science. Volume 2632. [12] T. Goel, R. Haftka, W. Shyy, N. Queipo, R. Vaidyanathan, and K. Tucker. Response surface approximation of pareto optimal front in multiobjective optimization. Computer Methods in Applied Mechanics and Engineering, 196(4-6):879–893, 1 January 2007. [13] T. Goel, R. Vaidyanathan, R. Haftka, W. Shyy, N. Queipo, and K. Tucker. Response surface approximation of pareto optimal front in multiobjective optimization. Technical Report 2004-4501, AIAA, 2004. [14] Y. Jin. A comprehensive survey of fitness approximation in evolutionary computation. Soft Computing, 9(1):3–12, 2005. [15] R. Landa Becerra and C.A. Coello Coello. Solving Hard Multiobjective Optimization Problems Using ε-Constraint with Cultured Differential Evolution. In Thomas Philip Runarsson, Hans-Georg Beyer, Edmund Burke, Juan J. Merelo-Guerv´os, L. Darrell Whitley, and Xin Yao, editors, Parallel Problem Solving from Nature - PPSN IX, 9th International Conference, pages 543–552. Springer. Lecture Notes in Computer Science Vol. 4193, Reykjavik, Iceland, September 2006. [16] D. Lim, Y. Jin, Y.S. Ong, and B. Sendhoff. Generalizing surrogateassisted evolutionary computation. IEEE Transactions on Evolutionary Computation, 14(3):329–355, 2010. [17] J.I. Madsen, W. Shyy, and R.T. Haftka. Response surface techniques for diffuser shape optimization. AIAA journal, 38(9):1512–1518, 2000. [18] H. Nakayama, M. Arakawa, and K. Washino. Optimization for black-box objective functions. In Panos M. Pardalos, Ider Tseveendorj, and Rentsen Enkhbat, editors, Optimization and Optimal Control, pages 185–210. World Scientific, Singapore, 2003. [19] W. Pedrycz, A. Skowron, and V. Kreinovich. Handbook of granular computing. Wiley-Interscience New York, NY, USA, 2008.

[20] C. Poloni, A. Giurgevich, L. Onesti, and V. Pediroda. Hybridization of a multi-objective genetic algorithm, a neural network and a classical optimizer for a complex design problem in fluid dynamics. Computer Methods in Applied Mechanics and Engineering, 186(2-4):403–420, 2000. [21] M. Reyes Sierra and C.A. Coello Coello. A Study of Fitness Inheritance and Approximation Techniques for Multi-Objective Particle Swarm Optimization. In 2005 IEEE Congress on Evolutionary Computation (CEC’2005), volume 1, pages 65–72, Edinburgh, Scotland, September 2005. IEEE Service Center. [22] J. Rezaei and M. Davoodi. Multi-objective models for lot-sizing with supplier selection. International Journal of Production Economics, 130(1):77–86, 2011. [23] A. Rowhanimanesh and M. Akbarzadeh-T. Perception-based evolutionary optimization: Outline of a novel approach to optimization and problem solving. In IEEE International Conference on Systems Man and Cybernetics, pages 4270–4275. IEEE Press, 2010. [24] J. Sacks, W.J. Welch, T.J. Mitchell, and H.P. Wynn. Design and analysis of computer experiments. Statistical science, 4(4):409–423, 1989. [25] L.V. Santana-Quintero, A. Arias Monta˜no, and C.A. Coello Coello. A Review of Techniques for Handling Expensive Functions in Evolutionary Multi-Objective Optimization. In Yoel Tenne and Chi-Keong Goh, editors, Computational Intelligence in Expensive Optimization Problems, pages 29–59. Springer, Berlin, Germany, 2010. ISBN 978-3-642-107009. [26] M. Smith. Neural Networks for Statistical Modeling. von Nostrand, Reinhold, New York, USA, 1993. [27] R.E. Smith, B. A. Dike, and S. A. Stegmann. Fitness inheritance in genetic algorithms. In SAC ’95: Proceedings of the 1995 ACM symposium on Applied computing, pages 345–350, New York, NY, USA, 1995. ACM Press. [28] D.A. Van Veldhuizen. Multiobjective Evolutionary Algorithms: Classifications, Analyses, and New Innovations. PhD thesis, Department of Electrical and Computer Engineering. Graduate School of Engineering. Air Force Institute of Technology, Wright-Patterson AFB, Ohio, May 1999. [29] L.J.J. Wismans, E.C. van Berkum, and M.C.J. Bliemer. Multi objective optimization of traffic systems using dynamic traffic management measures. In Proceedings of ITS-conference Models and Technologies for Intelligent Transportation Systems, June 2009. [30] Y.Y. Yao. Information granulation and rough set approximation. International Journal of Intelligent Systems, 16(1):87–104, 2001. [31] L. A. Zadeh. Fuzzy sets and information granularity. In Advances in Fuzzy Set Theory and Applications, pages 3–18. North Holland, New York, USA, 1979. [32] E. Zitzler. Evolutionary Algorithms for Multiobjective Optimization: Methods and Applications. PhD thesis, Swiss Federal Institute of Technology (ETH), Zurich, Switzerland, November 1999. [33] E. Zitzler, K. Deb, and L. Thiele. Comparison of Multiobjective Evolutionary Algorithms: Empirical Results. Evolutionary Computation, 8(2):173–195, Summer 2000.