Hybrid krill herd algorithm with differential evolution for global ...

Report 3 Downloads 101 Views
Neural Comput & Applic (2014) 25:297–308 DOI 10.1007/s00521-013-1485-9

ORIGINAL ARTICLE

Hybrid krill herd algorithm with differential evolution for global numerical optimization Gai-Ge Wang • Amir H. Gandomi Amir H. Alavi • Guo-Sheng Hao



Received: 9 December 2012 / Accepted: 10 September 2013 / Published online: 1 October 2013 Ó Springer-Verlag London 2013

Abstract In order to overcome the poor exploitation of the krill herd (KH) algorithm, a hybrid differential evolution KH (DEKH) method has been developed for function optimization. The improvement involves adding a new hybrid differential evolution (HDE) operator into the krill, updating process for the purpose of dealing with optimization problems more efficiently. The introduced HDE operator inspires the intensification and lets the krill perform local search within the defined region. DEKH is validated by 26 functions. From the results, the proposed methods are able to find more accurate solution than the KH and other methods. In addition, the robustness of the DEKH algorithm and the influence of the initial population size on convergence and performance are investigated by a series of experiments. Keywords Global optimization problem  Krill herd (KH)  Hybrid differential evolution (HDE) operator

G.-G. Wang (&)  G.-S. Hao School of Computer Science and Technology, Jiangsu Normal University, Xuzhou, Jiangsu 221116, China e-mail: [email protected] G.-S. Hao e-mail: [email protected] A. H. Gandomi Department of Civil Engineering, The University of Akron, Akron, OH 44325, USA e-mail: [email protected]; [email protected] URL: http://www.gozips.uakron.edu/*ag72/ A. H. Alavi Department of Civil and Environmental Engineering, Michigan State University, Engineering Building, East Lansing, MI 48824, USA e-mail: [email protected]; [email protected] URL: http://www.egr.msu.edu/*alavi/

1 Introduction In the current world, the resources are becoming less and less. Due to the limited resources, all the people are striving for making most profit with least cost and resources. These problems can be modeled into maximum/minimum problem, which are also called optimization problems. And then, various technologies are well capable of solving these optimization problems. Generally speaking, there are various ways to divide these technologies; however, a simple way is to look at the property of the techniques. These techniques can be classified into two categories: traditional classical methods and modern intelligent algorithms. The prior has strict step and always generate the same solution with the same starting value. Contrarily, modern intelligent algorithms, based on some stochastic distribution, cannot generate the same solutions under any conditions. In most cases, both can find the satisfactory solution. Recently, nature-inspired metaheuristic methods have emerged, and they perform better than traditional methods in solving optimization problems. These robust metaheuristic methods have been designed to solve complicated problems, such as cluster [1], truss structure design [2], UCAV path planning [3], permutation flow shop scheduling [4], knapsack problem [5], task assignment problem [6], and engineering optimization [7, 8]. These kinds of metaheuristic methods can take advantage of the useful information from the whole population. Therefore, they have a higher probability of searching for the final satisfactory solution. After the emergence of the genetic algorithms (GAs) [9–11] during the 1960s and 1970s, serials of techniques have been developed in the field of optimization, such as differential evolution (DE) [12–14], harmony search (HS) [15–17], artificial bee colony (ABC) [18], particle swarm optimization (PSO) [19–21], evolutionary strategy (ES) [22, 23], ant colony optimization (ACO) [24], firefly algorithm (FA) [25], artificial

123

298

plant optimization algorithm (APOA) [26], artificial physics optimization [27], biogeography-based optimization (BBO) [28], genetic programming (GP) [29], cuckoo search (CS) [30–32], animal migration optimization (AMO) [33], probability-based incremental learning (PBIL) [34], big bang–big crunch algorithm [35–38], bat algorithm (BA) [39, 40], charged system search (CSS) [41], and the KH method [42]. According to the idealization of the social phenomenon of krill, KH has been developed for optimizing complicated functions [42]. In KH, the objective function can be idealized into the distances from food source and its herd density. The position of the krill involves three components. A major advantage of KH is that it requires few control variables compared with other methods. In some cases, KH may fail to find the best solution on multimodal functions. Herein, an effective hybrid differential evolution krill herd (DEKH) method is proposed so as to accelerate convergence speed. In DEKH, firstly, we use the KH to shrink the search apace and select a good promising candidate solution set. Thereafter, for making the krill proceed to the best solution, a more focused hybrid differential evolution (HDE) operator is introduced into the algorithm. This added operator is utilized to search promising region to get optimal solutions for solving optimization problems. Some solutions would be updated by HDE operator around the optimal krill, and this will accelerate the convergence speed. Proposed method is validated on twenty-six functions. Experimental results indicate that the DEKH is able to find the better solution than the KH, other eight methods.

2 A review of hybrid metaheuristic methods It is well-known that a hybrid method can take merits of the methods that it was originated from. Some most representative hybrid methods are reviewed here. El-Abd [43] investigated the capabilities of two swarm intelligence algorithms, namely ABC and PSO and combined into one method, called ABC–PSO. ABC–PSO is a component-based technique where the PSO is expanded with an ABC to enhance the optimal particles. Wang and Guo [44] enhanced the diversity of the bat swarm by adding pitch adjustment operation in HS to the BA method. Based on this mutation operator, a hybrid HS/BA method is proposed and tested on various benchmarks. Duan et al. [45] proposed a hybrid method by merging the BBO together with the TS (tabu search) to deal with the test-sheet composition problem. HS cannot always find a satisfactory solution. In order to overcome this limitation, Gao et al. [46] developed a new HS– PBIL method through the combination of HS and PBIL. Geem [47] proposed an improved HS incorporating particle swarm concept. Gong et al. [48] introduced a hybrid DE and BBO method to merge the exploration of DE together with the

123

Neural Comput & Applic (2014) 25:297–308

exploitation of BBO. In order to cope with the difficulty of setting the number of clusters in most of the clustering methods, Kuo [49] established a novel clustering method based on PSO and GA (DCPG). DCPG can cluster data by means of testing of the data without a cluster number. In 2012, a hybrid co-evolutionary cultural algorithm based on PSO, namely CECBPSO, is developed by Sun et al. [50]. In [51], two new models are proposed for short-term electricity load forecasting. These models use ACO and hybridization of GA and ACO for feature selection and multi-layer perceptron for load prediction. Marichelvam [52] developed an improved hybrid CS algorithm to deal with the permutation flow shop scheduling problems.

3 The proposed DEKH method KH [42] is a heuristic search technique for solving function optimization, which is the idealization of the krill swarm in the process of hunting for the food and communicating with each other. The position of a krill can be idealized into three actions: 1. 2. 3.

movement influenced by other krill; foraging action; physical diffusion. KH used the Lagrangian model in Eq. (1).

dXi ¼ Ni þ Fi þ Di dt

ð1Þ

where Ni is the motion induced by other krill individuals; Fi is the foraging motion, and Di is the physical diffusion of the ith krill individual. The direction of the first movement, ai, is estimated by the target effect, local effect, and a repulsive effect. For a krill individual, this movement can be formulated as: Ninew ¼ N max ai þ xn Niold

ð2Þ

And Nmax is the maximum induced speed, xn is the inertia weight, Niold is the last motion induced. The second movement is affected by the two factors: the current food location and the previous experience. For the ith krill, it can be given below: Fi ¼ Vf bi þ xf Fiold

ð3Þ

where bi ¼ bfood þ bbest i i

ð4Þ

and Vf is the foraging speed, xf is the inertia weight, Fiold is the last foraging motion. The third movement can be considered to be a random process. It can be expressed according to a diffusion speed and a directional factor. It can be formulated as follows:

Neural Comput & Applic (2014) 25:297–308

Di ¼ Dmax d

299

ð5Þ

where Dmax is the maximum diffusion speed and d is the random directional vector, and its arrays are values in [-1, 1]. Herein, the position of a krill in the Dt is formulated by the Eq. (6): Xi ðt þ DtÞ ¼ Xi ðtÞ þ Dt

dXi dt

ð6Þ

For more details on this method, readers are advised to refer to Gandomi and Alavi [42]. For the third motion of the KH method, how a krill moves toward next position is mainly based on random walks; therefore, the satisfactory solution cannot always be found every time. In order to remove this disadvantage, some kind of genetic reproduction mechanisms have been added to KH approach [42]. For most cases, KH is well capable of making full use of the population knowledge and has proven good performance on some benchmarks. However, sometimes, KH may not proceed to the optimal solutions on some complex problems. Here, in order to further enhance the exploitation of KH, a more focused hybrid mutation operator performing local search, called hybrid differential evolution (HDE) operator, is incorporated into the KH to form a novel differential evolution KH (DEKH) method. This more focused local search technique can make the population not converge prematurely. The greedy selection strategy is used in this operator. In other words, the newly generated krill replace the previous one only when its fitness is better than before. Therefore, in this way, the HDE operator works better than mutation. The HDE operator used in DEKH algorithm can be represented in Algorithm 1.

problem, if the newly generated krill has a smaller fitness, it will be in place of the old one. In other words, it can be considered as krill i in the next generation. In DEKH, standard KH algorithm is utilized to search globally for the purpose of making most krill move toward a more promising area. After exploration stage, HDE operator is utilized to search locally with a small step to get the final best solution. Based on the mainframe of DEKH, the KH emphasizes the diversification at the beginning of the search with a big step to explore the search space extensively and evade trapping into local optima, while later HDE operator focuses on the intensification and lets the individuals move toward the best solution at the later stage of the optimization process. In addition, this method is able to cope with the conflict global and local search effectively. In addition, except HDE operator, elitism strategy is also added to the DEKH. Like other methods, some kind of elitism is introduced with the aim of keeping the optimal krill in the population, which can forbid the optimal krill from being ruined by incorrect position updating operator. This elitism strategy is able to make all the krill always proceed to the population with better fitness than before. By incorporating the above-mentioned HDE operator and focused elitism strategy into the KH algorithm, the DEKH has been developed, as described in Fig. 1. Here, d is dimension of the function. NP is the parent population size. FW is the mutation scaling factor. CR is a crossover constant. Xi(j) is the j-dimension for Xi. Ui, Vi are the offspring. r1, r2, and r3 are random integers in [1, NP].

4 The results



In the present work, various test functions, as shown in Table 1, are employed to evaluate our DEKH method. In order to get fair results, all the experiments were done under the same conditions as shown in [53]. More details about these test functions can be found in recent literatures [28, 54, 55]. 4.1 The robustness of DEKH

In Algorithm 1, the mutation and crossover operators are used to generate new position for krill i. For a minimum

As mentioned ‘‘robust’’ several times, it is required to show the robust feature of the DEKH method here. To do this, firstly select a certain benchmark problem (F24 Sum function). And then, we carry out this 100 times to get its statistic characteristics of the final solutions (mean, standard deviation of each variable in the final solution sets), as shown in Table 2. The relation between the evolutions of the fitness and number of generations is represented in Fig. 2. From Table 2, DEKH is able to get the optimal value for F24 Sum function among ten methods. For the

123

300

Neural Comput & Applic (2014) 25:297–308 Table 1 Benchmark functions No.

Name

F01

Ackley

F02

Alpine

F03

Brown

F04

Dixon & Price

F05

Fletcher-Powell

F06

Griewank

F07

Holzman 2 function

F08

Levy

F09

Pathological function

F10

Penalty #1

F11 F12

Penalty #2 Perm #1

F13

Perm #2

F14

Powel

F15

Quartic with noise

F16

Rastrigin

F17

Rosenbrock

F18

Schwefel 2.26

F19

Schwefel 1.2

F20

Schwefel 2.22

F21

Schwefel 2.21

F22

Sphere

F23

Step

F24

Sum function

F25

Zakharov

F26

Wavy1

worst values, the values of ABC, DE, DEKH, GA, KH, and PSO are slightly different, which are better than others. Similar to best value, DEKH is well capable of getting the minimum on the average. In addition, the Std of DEKH is the minimum. It is demonstrated that the values obtained by DEKH have little difference from different implementations. 4.2 The superiority of DEKH In order to prove the merits of DEKH, it was compared with nine methods, which are ABC [18], ACO [24], DE [12, 56, 57], ES [23], GA [9], HS [15], KH [42], PBIL [34], and PSO [19, 58, 59]. In addition, by combination of different genetic operators, four variants of KHs are developed, as described below [42]: KH KH KH KH Fig. 1 Flowchart of DEKH algorithm

123

I. KH without any genetic operators; II. KH with crossover operator; III. KH with mutation operator; IV. KH with crossover and mutation operators [42].

Through various experiments on benchmark problems, Gandomi and Alavi [42] have claimed that the KH II is the

Neural Comput & Applic (2014) 25:297–308

301

Table 2 The best, worst, mean, and standard of the F24 Sum function ABC

ACO

DE

DEKH

ES

GA

HS

KH

PBIL

PSO

Best

140.04

317.93

112.65

0.18

730.56

152.52

999.80

29.22

1.11E3

303.92

Worst

906.99

1.48E3

906.99

906.99

2.05E3

906.99

2.30E3

906.99

2.78E3

906.99

Mean

422.13

964.83

192.05

1.44

2.05E3

467.83

2.05E3

84.93

2.37E3

735.84

Std

133.74

288.96

41.39

1.00

374.75

172.91

355.51

26.67

364.61

176.15

4000 ABC ACO DE DEKH ES

benchmark function value

3500 3000

GA HS KH PBIL PSO

2500 2000 1500 1000 500 0

0

5

10

15

20

25

30

35

40

45

50

Number of generations

Fig. 2 The evolutions of the objective function for ten methods

best method among four variants of KHs. Herein, in the present work, KH II is considered as one of the comparative methods. In other words, in the next texts, when KH is mentioned, it means KH II. For KH and DEKH, the same parameters will be used: the Vf = 0.02, Dmax = 0.005, Nmax = 0.01, FW = 0.1, and CR = 0.4. For others, we set the same parameters as [53]. Note that, for each function, it has 20 variables. To get representative performances for each algorithm, we did 150 implementations on test functions. The results are recorded in Tables 3 and 4, which illustrate the average and mean minima, respectively. The function minimum is highlighted through bold font. The values in the following tables are normalized according to its own scale, the detailed process of which can be found in [55]. From Tables 3 and 4, it can be seen that, DEKH can find function minimum on twenty-four of the twenty-six benchmarks. GA performs the best on the function F19. Furthermore, ES and ACO overtakes all other methods on F09 for the best values and F24 for the average values, respectively. In addition, to provide more proof to testify the advantage of the DEKH, optimization process of ten

methods on some most typical benchmarks are also provided in this section, as shown in Figs. 3, 4, 5 and 6. The solutions are the true average objective function minimum. Figure 2 clearly indicates DEKH is the best method for solving F02 Alpine function from start to finish. For other methods, KH is simply inferior to the DEKH and it eventually finds the function value close to DEKH; while other methods except KH and DEKH, do not search for the satisfactory global minimum. Here, all the methods start the optimization process form almost same value; however, DEKH outperforms them soon. For this problem, DEKH outperforms all others from start to finish. Other methods do not find the satisfactory solution. Clearly, KH converges to the value that is only inferior to DEKH’s. For others, DE can search for the function value that is almost the same with ABC. Table 3 and Fig. 4 show that ABC performs slightly better than DE for this case. For this case, DEKH outperforms all other methods from start to finish. Other methods fail to find the satisfactory solution under certain conditions. Moreover, for other methods, similarly, ABC can search for the function value that is almost the same with KH. Table 2 and Fig. 5 also shows ABC is slightly superior to KH for this function. KH performs very well and ranks 3 among ten approaches. For this case, DEKH is the best method that can search for the global minimum and outperforms all others. ACO performs very well and ranks 2 among ten approaches. 4.3 The difference of initial population size The above comparisons are based on same number of initial population. If we use more individuals in the initial population in the methods, in other words, the diversity of methods is increased. Here, the influence of initial population size will be showed through various simulations on benchmark F01–F14. In the present work, the initial population size NP = 50, 100, 150, and 200 is selected. Because the initial population size NP = 50 is used in the experiments conducted in Sects. 4.1–4.2, the results can be found in the first 15 rows of the Tables 3 and 4 for

123

302

Neural Comput & Applic (2014) 25:297–308

Table 3 Mean normalized function values ABC

ACO

DE

DEKH

ES

GA

HS

KH

PBIL

PSO

F01

11.92

13.81

10.76

1.00

16.84

15.14

17.19

3.87

17.44

14.50

F02

5.63

9.12

10.47

1.00

19.63

8.91

17.40

5.90

20.49

13.36

F03

68.31

121.42

19.76

1.00

947.04

62.02

516.97

96.27

192.64

453.86

F04

754.82

1.2E3

222.33

1.00

8.9E3

467.10

9.0E3

96.78

1.0E4

1.2E3

F05

2.73

10.83

4.15

1.00

10.43

4.23

10.17

4.06

9.43

8.36

F06

38.24

12.10

21.33

1.00

98.94

40.45

197.31

6.11

218.71

81.30

F07

1.3E3

2.4E3

543.81

1.00

1.7E4

1.5E3

1.7E4

161.62

2.1E4

3.8E3

F08

22.43

37.70

24.73

1.00

119.08

32.84

108.72

13.03

130.50

68.53

F09

2.93

3.12

1.69

2.47

1.00

2.45

3.71

3.54

3.22

2.64

F10

2.6E5

1.2E7

5.7E4

1.00

7.6E6

8.3E4

1.3E7

2.7E3

1.6E7

1.4E6

F11 F12

3.5E5 1.1E11

7.5E6 3.8E10

9.5E4 1.1E6

1.00 1.00

4.5E6 2.6E8

2.8E5 1.3E10

6.7E6 1.7E8

9.7E3 1.6E11

8.7E6 7.4E9

1.3E6 1.3E8

F13

4.6E12

5.6E12

8.4E7

1.00

4.2E10

5.0E12

3.6E10

8.7E12

3.6E12

4.2E10

F14

14.81

90.22

36.61

1.00

128.34

20.83

98.18

12.90

107.42

42.65

F16

1.7E3

1.3E3

579.54

1.00

1.7E4

1.3E3

1.8E4

168.57

2.4E4

4.2E3

F17

1.38

2.68

2.34

1.00

3.63

2.45

3.46

1.45

3.74

2.72

F18

13.06

70.94

11.23

1.00

95.61

19.47

68.56

4.98

79.93

22.45

F19

1.60

1.04

2.03

1.32

2.49

1.00

3.09

1.94

3.16

3.03

F20

6.11

5.31

7.55

1.00

8.88

5.95

7.87

3.64

8.67

6.05

F21

7.65

21.34

9.23

1.00

31.92

15.68

25.97

12.01

26.01

18.86

F22

10.48

6.49

8.22

1.00

9.97

8.37

10.45

1.77

10.97

8.25

F23

615.61

1.6E3

320.20

1.00

3.4E3

1.2E3

3.1E3

97.12

3.5E3

1.2E3

F24

492.15

204.77

242.95

1.00

1.6E3

469.39

2.3E3

58.49

2.6E3

911.76

F25

262.03

594.33

129.14

1.00

1.4E3

296.50

1.3E3

57.32

1.6E3

476.61

F26

2.15

9.6E3

2.54

1.00

2.87

2.20

2.77

1.61

2.52

2.48

F16

9.24

7.28

10.00

1.00

20.83

11.53

26.21

9.10

27.29

17.28

Total

0

0

0

24

1

1

0

0

0

0

DE

DEKH

ES

GA

HS

KH

PBIL

PSO

Table 4 Best normalized function values ABC

ACO

F01

75.56

95.18

70.44

1.00

128.99

103.93

137.49

14.37

136.79

108.85

F02

43.86

55.28

81.07

1.00

140.49

49.28

140.24

36.34

176.68

104.85

F03

48.72

233.26

39.27

1.00

733.71

97.60

504.38

100.76

338.32

367.58

F04

446.05

1.3E3

438.99

1.00

1.6E4

87.45

2.2E4

144.87

2.3E4

2.7E3

F05

3.47

15.24

7.16

1.00

18.08

5.93

15.44

5.19

6.98

12.98

F06

12.52

5.49

12.11

1.00

60.99

13.58

129.45

2.47

148.09

54.87

F07

934.13

1.0E4

1.6E3

1.00

1.9E5

4.0E3

1.7E5

1.2E3

2.5E5

1.4E4

F08

16.39

34.15

36.15

1.00

153.36

15.45

167.48

14.92

209.22

62.20

F09

105.68

132.64

48.46

1.00

36.64

105.86

178.00

158.05

143.20

69.69

F10

716.91

1.05

6.3E3

1.00

5.8E6

5.59

1.0E7

11.00

1.1E7

1.3E5

F11 F12

3.8E4 8.2E16

1.00 4.6E21

8.1E4 4.5E12

1.31 1.00

8.7E6 1.1E16

3.7E3 4.6E21

1.8E7 4.0E15

5.6E3 1.9E8

2.1E7 4.6E21

2.3E6 2.6E16

F13

4.1E14

1.8E17

3.5E6

1.00

4.2E11

1.8E17

6.6E12

1.6E10

1.8E17

1.5E9

F14

19.95

84.67

74.08

1.00

201.05

25.93

257.77

18.68

288.11

56.54

F16

771.76

3.9E3

4.2E3

1.00

1.3E5

1.7E3

1.7E5

956.73

1.4E5

2.3E4

123

Neural Comput & Applic (2014) 25:297–308

303

Table 4 continued ABC

ACO

DE

DEKH

ES

GA

HS

KH

PBIL

PSO 4.08

F17

1.48

3.98

3.58

1.00

4.82

2.76

5.28

1.61

6.15

F18

6.95

41.32

8.17

1.00

72.96

8.00

44.77

3.11

46.99

14.35

F19

3.48

1.70

4.75

2.82

6.12

1.00

7.98

3.95

7.04

7.06 8.37

F20

7.36

3.47

8.14

1.00

14.75

3.91

8.85

3.25

11.63

F21

21.07

63.02

27.79

1.00

70.68

42.49

86.58

33.27

92.47

52.61

F22

17.83

8.77

12.55

1.00

15.57

7.62

17.65

2.11

20.41

13.31

F23

1.3E3

4.6E3

769.86

1.00

9.5E3

1.9E3

1.2E4

182.22

1.4E4

4.5E3

F24

1.9E3

920.00

1.1E3

1.00

1.2E4

1.3E3

1.6E4

195.00

1.3E4

6.0E3

F25

660.01

1.5E3

607.12

1.00

5.3E3

795.14

7.7E3

167.34

7.7E3

1.6E3

F26

3.87

1.48

3.50

1.00

4.32

1.63

3.52

1.77

4.63

2.06

F16

11.78

9.14

14.53

1.00

30.63

11.15

43.95

11.53

37.71

24.70

Total

0

1

0

24

0

1

0

0

0

0

300

40 35

benchmark function value

250

benchmark function value

30 25 20 15 10

ABC ACO DE DEKH ES

5 0

0

5

10

GA HS KH PBIL PSO

15

20

200

150

50

25

30

35

40

45

50

ABC ACO DE DEKH ES

100

0

5

10

140 GA HS KH PBIL PSO

80

60

40

20

0

5

10

15

20

25

30

35

40

45

Number of generations

Fig. 4 Performance comparison for the F08 Levy function

30

35

40

45

50

1200 1000 800 600 400

ABC ACO DE DEKH ES

200

0

25

1400

benchmark function value

benchmark function value

100

20

Fig. 5 Performance comparison for the F16 Rastrigin function

Fig. 3 Performance comparison for the F02 Alpine function

120

15

Number of generations

Number of generations

ABC ACO DE DEKH ES

GA HS KH PBIL PSO

50

0

0

5

10

GA HS KH PBIL PSO

15

20

25

30

35

40

45

50

Number of generations

Fig. 6 Performance comparison for the F26 Wavy1 function

123

304

Neural Comput & Applic (2014) 25:297–308

Table 5 Best normalized function values with initial population size NP = 100 ABC

ACO

DE

DEKH

ES

GA

HS

KH

PBIL

PSO

F01

88.33

109.24

86.04

1.00

147.50

110.69

147.68

14.70

148.08

120.55

F02

4.26

7.24

10.53

1.00

20.18

4.83

18.04

5.57

17.12

11.67

F03

24.94

80.01

18.45

1.00

272.27

37.04

397.37

47.70

132.06

219.58

F04

93.63

546.99

46.19

1.00

5.0E3

58.28

1.5E3

83.16

3.6E3

361.46

F05

3.19

12.22

5.27

1.00

11.79

1.72

10.27

6.34

9.11

9.93

F06

16.50

3.78

17.06

1.00

72.98

9.82

145.94

7.77

160.34

34.58

F07

258.72

697.47

252.52

1.00

7.9E3

36.31

5.5E3

199.76

9.6E3

675.49

F08

7.62

23.15

13.84

1.00

65.21

7.41

43.19

8.91

55.23

14.65

F09

3.65

4.19

1.32

2.61

1.00

2.78

5.00

4.34

3.98

3.22

F10

486.74

2.06

8.6E3

1.00

9.6E6

4.36

1.8E7

1.6E3

1.9E7

9.2E5

F11 F12

4.2E6 1.1E14

1.00 3.2E16

3.9E6 1.2E9

260.70 1.00

1.0E9 8.6E12

1.4E4 3.2E16

1.3E9 4.4E11

8.4E6 4.6E7

1.1E9 3.2E16

3.0E7 6.0E12

F13

1.8E12

1.9E14

2.8E6

1.00

1.3E7

1.9E14

8.9E7

4.4E10

1.9E14

2.5E9

F14

4.07

44.32

14.92

1.00

39.67

1.70

36.98

6.59

49.61

18.32

Total

0

1

0

12

1

0

0

0

0

0

Table 6 Mean normalized function values with initial population size NP = 100 ABC

ACO

DE

DEKH

ES

GA

HS

KH

PBIL

PSO

F01

54.69

65.22

52.11

1.00

82.79

70.34

82.51

13.39

82.69

69.40

F02

2.74

6.19

6.06

1.00

11.05

4.48

10.03

3.63

10.62

7.58

F03

34.17

51.60

10.56

1.00

428.17

27.11

320.45

52.51

79.64

169.13

F04

166.12

443.20

51.75

1.00

2.6E3

78.17

1.8E3

68.20

2.1E3

243.24

F05

1.73

7.16

2.88

1.00

8.54

2.17

6.10

3.41

5.97

6.06

F06

23.44

5.45

17.05

1.00

125.43

23.81

144.32

9.07

148.93

50.33

F07

244.23

783.74

120.62

1.00

5.4E3

167.08

3.6E3

126.13

4.5E3

412.07

F08

11.61

30.13

15.28

1.00

74.43

10.21

60.02

10.22

67.07

26.02

F09

3.44

3.86

1.90

2.95

1.00

2.80

4.47

4.15

3.76

3.25

F10

1.1E5

1.1E7

3.3E4

1.00

7.5E6

3.9E4

8.2E6

1.8E4

9.4E6

9.0E5

F11 F12

1.9E4 5.1E9

1.5E6 9.6E9

1.0E4 1.2E5

1.00 1.00

9.7E5 7.1E7

7.2E3 5.7E9

9.6E5 2.3E7

8.4E3 1.2E10

1.0E6 5.7E9

1.2E5 1.3E8

F13

2.3E10

6.9E10

1.4E6

1.00

5.4E8

5.7E10

2.9E8

7.4E9

5.7E10

3.1E7

F14

4.52

46.29

14.05

1.00

47.80

4.32

38.85

8.37

39.53

16.10

Total

0

0

0

13

1

0

0

0

0

0

NP = 50. For the initial population size NP = 100, 150, and 200, the normalized values are shown in Tables 5, 6, 7, 8, 9, and 10. From Tables 5, 6, 7, 8, 9, and 10, DEKH is always better than other nine methods on different initial population size for most functions. For other approaches, ABC, ACO, ES, and GA are able to get optimal value on one or two functions in initial population size. This indicates that the DEKH method is well capable of taking the use of the population information, and it is insensitive to the initial population size NP. It has been demonstrated that DEKH is a more robust method than other methods.

123

The experiments carried out in Sects. 4.1–4.3 and above analyses about the Figs. 1, 2, 3 and 4 show that, in most cases, the DEKH and KH are the first and second best method on most benchmark problems among ten approaches, respectively. ABC, ACO, DE, and GA are the third best methods that are simply inferior to the DEKH and KH.

5 Conclusions In order to overcome the disadvantage of the KH, a HDE operator has been incorporated into the KH to generate

Neural Comput & Applic (2014) 25:297–308

305

Table 7 Best normalized function values with initial population size NP = 150 ABC

ACO

DE

DEKH

ES

GA

HS

KH

PBIL

PSO

F01

63.32

92.47

64.50

1.00

133.78

99.00

127.66

13.12

128.83

100.10

F02

2.99

4.68

5.67

1.00

12.74

3.80

9.73

4.49

9.92

6.71

F03

5.98

28.76

4.85

1.00

231.23

8.91

160.03

25.94

58.26

62.11

F04

51.64

250.90

46.83

1.00

2.7E3

37.96

1.9E3

65.15

1.9E3

141.22

F05

1.96

5.73

3.90

1.00

7.46

1.94

7.00

4.38

8.51

7.57

F06

6.05

4.36

11.74

1.00

106.39

6.52

90.71

7.44

118.46

32.96

F07

61.52

210.98

50.56

1.00

2.7E3

4.59

1.8E3

104.31

2.3E3

174.51

F08

12.23

15.72

11.48

1.00

56.09

8.71

61.09

11.06

67.29

32.47

F09

3.86

3.96

1.32

2.69

1.00

2.36

5.06

4.80

3.86

2.31

F10

2.5E3

1.49

2.3E4

3.54

1.6E7

1.00

5.9E6

3.0E4

4.9E6

2.1E5

F11 F12

6.4E4 3.3E14

1.28 9.5E15

5.3E4 2.5E9

1.00 1.00

8.8E6 7.4E11

799.21 9.5E15

9.1E6 9.6E11

1.1E5 3.6E8

5.7E6 9.5E15

6.5E5 7.0E11

F13

1.4E10

1.9E14

7.0E6

1.00

3.6E8

1.9E14

1.4E10

1.1E11

1.9E14

4.9E8

F14

2.34

26.76

11.93

1.00

53.88

1.28

28.12

5.97

31.48

9.32

Total

0

0

0

12

1

1

0

0

0

0

Table 8 Mean normalized function values with initial population size NP = 150 ABC

ACO

DE

DEKH

ES

GA

HS

KH

PBIL

PSO

F01

58.97

74.87

61.00

1.00

106.18

89.12

103.64

13.59

104.20

82.46

F02 F03

2.24 20.71

4.08 21.51

4.32 4.70

1.00 1.00

9.02 298.94

3.31 14.93

8.09 149.08

3.58 33.53

8.38 48.49

5.34 80.79

F04

66.65

200.85

29.61

1.00

1.6E3

38.33

1.3E3

58.71

1.5E3

134.77

F05

1.02

4.17

1.79

1.00

4.45

1.57

3.61

2.24

3.97

3.70

F06

16.74

5.00

12.55

1.00

117.68

14.73

126.96

10.53

127.21

39.33

F07

96.01

563.16

42.31

1.00

2.4E3

64.45

1.6E3

101.11

1.8E3

162.67

F08

9.58

25.77

10.35

1.00

48.17

9.93

45.55

9.57

52.49

25.25

F09

3.37

3.75

1.72

3.06

1.00

2.44

4.32

4.17

3.64

2.83

F10

9.9E3

1.7E6

1.6E4

1.07

5.0E6

1.00

3.4E6

1.8E4

2.8E6

2.1E5

F11

4.2E3

2.9E4

1.7E3

1.00

2.6E5

412.97

2.2E5

3.6E3

2.2E5

1.9E4

F12

1.3E11

2.1E11

1.0E6

1.00

7.7E8

2.1E11

2.9E8

1.1E8

2.1E11

1.1E9

F13

1.4E9

1.2E10

7.4E4

1.00

5.5E7

1.2E10

6.3E7

3.0E8

1.2E10

3.6E6

F14

2.24

26.76

9.42

1.00

42.00

2.33

19.73

4.70

23.32

8.55

Total

0

0

0

12

1

1

0

0

0

0

an improved DEKH approach for function optimization. In DEKH, KH is utilized to perform global search that make most krill move toward a more promising area. Thereafter, the HDE operator searches locally to get better solutions. By using two techniques, DEKH can balance global and local search and cope with multimodal benchmarks. From the results of the ten methods on the functions discussed in this paper, the DEKH’s performance is better than, or at least comparable with, the other nine algorithms.

For DEKH, there are various issues that still deserve further study. Firstly, the influence of the number of elitism, crossover factor CR, and weighting factor FW is required to be investigated through a series of experiments. Through this, DEKH can be used to solve problems more efficiently. Secondly, the capability of this hybrid approach to improve the performance of other population-based optimization methods is worthy of further study. Thirdly, the performance of DEKH is validated only by comparing with some basic methods. Thus, for more verification of the

123

306

Neural Comput & Applic (2014) 25:297–308

Table 9 Best normalized function values with initial population size NP = 200 ABC

ACO

DE

DEKH

ES

GA

HS

KH

PBIL

PSO

F01

89.72

101.10

84.41

1.00

148.85

110.05

142.21

6.26

141.45

108.23

F02

2.33

5.18

5.41

1.00

9.09

2.79

9.67

3.62

8.92

6.96

F03

12.26

19.82

4.56

1.00

95.79

10.95

163.39

21.38

41.02

54.36

F04

42.63

211.26

16.32

1.00

1.1E3

8.50

761.96

55.62

801.46

75.63

F05

1.00

7.85

3.67

2.52

8.64

1.03

8.25

4.78

6.24

4.50

F06

9.45

3.68

12.47

1.00

111.24

4.72

100.66

12.30

96.89

38.54

F07

25.73

190.21

30.23

1.00

2.3E3

7.41

929.69

125.97

1.3E3

73.39

F08

5.39

13.06

7.76

1.00

40.64

2.65

37.05

10.68

31.79

13.34

F09

3.33

4.36

1.65

2.79

1.00

2.38

4.96

4.46

4.02

1.95

F10

66.92

1.00

2.7E3

3.31

1.3E7

2.15

1.0E7

1.3E5

7.3E6

9.6E4

F11 F12

1.1E4 5.5E12

1.00 1.3E14

3.3E5 1.6E7

293.84 1.00

4.4E7 4.9E8

63.80 1.3E14

1.8E7 1.3E9

2.1E5 1.3E9

2.2E7 1.3E14

2.7E5 3.0E10

F13

2.5E6

7.0E10

998.38

1.00

2.0E4

7.0E10

6.0E5

3.1E7

7.0E10

3.5E5

F14

3.06

34.64

10.38

1.60

56.49

1.00

33.91

8.48

29.06

8.30

Total

1

2

0

9

1

1

0

0

0

0

HS

KH

PBIL

PSO

Table 10 Mean normalized function values with initial population size NP = 200 ABC

ACO

DE

DEKH

ES

GA

F01

76.55

92.17

77.79

1.00

126.53

99.75

122.72

11.75

122.07

96.90

F02 F03

1.62 11.73

3.92 16.23

3.58 4.31

1.00 1.00

6.53 162.49

1.82 12.10

6.01 143.16

2.58 36.07

6.83 34.82

4.55 63.08 67.84

F04

35.82

252.42

18.25

1.00

1.0E3

20.82

702.28

46.78

713.12

F05

1.00

6.31

2.61

1.57

6.14

1.43

5.76

3.08

4.50

4.40

F06

15.22

3.78

12.16

1.00

112.73

7.30

112.98

12.31

103.72

37.21

F07

32.96

336.19

23.03

1.00

2.0E3

54.19

1.1E3

91.54

1.3E3

116.49

F08

5.59

14.82

7.72

1.00

47.52

5.02

33.80

10.13

39.79

16.14

F09

3.00

3.55

1.68

2.63

1.00

2.23

4.07

3.74

3.48

2.53

F10

1.9E3

2.7E6

4.2E3

1.00

5.5E6

21.41

3.2E6

6.0E4

3.0E6

2.0E5

F11

222.98

6.3E3

217.84

1.00

2.5E4

22.09

1.4E4

340.76

1.5E4

1.1E3

F12

2.3E9

1.9E9

2.9E4

1.00

1.5E6

1.9E9

6.5E6

3.8E8

1.9E9

3.8E6

F13

3.3E8

1.7E9

587.22

1.00

9.2E6

1.7E9

2.8E6

1.3E8

1.7E9

4.2E5

F14

2.23

19.33

6.93

1.00

32.29

1.43

18.84

6.58

17.12

5.52

Total

1

0

0

12

1

0

0

0

0

0

robustness of this hybrid method, it may be compared with other hybrid methods reviewed in Sect. 4.2.

References 1. Le´vano M, Nowak H (2011) New aspects of the elastic net algorithm for cluster analysis. Neural Comput Appl 20(6):835–850. doi:10.1007/s00521-010-0498-x 2. Talatahari S, Kheirollahi M, Farahmandpour C, Gandomi A (2012) A multi-stage particle swarm for optimum design of truss structures. Neural Comput Appl. doi:10.1007/s00521-012-1072-5

123

3. Wang G, Guo L, Duan H, Liu L, Wang H (2012) Path planning for UCAV using bat algorithm with mutation. Sci World J 2012:1–15. doi:10.1100/2012/418946 4. Li X, Yin M (2013) An opposition-based differential evolution algorithm for permutation flow shop scheduling based on diversity measure. Adv Eng Softw 55:10–31. doi:10.1016/j. advengsoft.2012.09.003 5. Zou D, Gao L, Li S, Wu J (2011) Solving 0-1 knapsack problem by a novel global harmony search algorithm. Appl Soft Comput 11(2):1556–1564. doi:10.1016/j.asoc.2010.07.019 6. Zou D, Liu H, Gao L, Li S (2011) An improved differential evolution algorithm for the task assignment problem. Eng Appl Artif Intell 24(4):616–624. doi:10.1016/j.engappai.2010.12.002

Neural Comput & Applic (2014) 25:297–308 7. Yang XS, Gandomi AH, Talatahari S, Alavi AH (2013) Metaheuristics in water. Geotechnical and Transport Engineering, Elsevier 8. Gandomi AH, Yang XS, Talatahari S, Alavi AH (2013) Metaheuristic applications in structures and infrastructures. Elsevier, Waltham 9. Goldberg DE (1998) Genetic algorithms in search. Optimization and machine learning. Addison-Wesley, New York 10. Zhao M, Ren J, Ji L, Fu C, Li J, Zhou M (2012) Parameter selection of support vector machines and genetic algorithm based on change area search. Neural Comput Appl 21(1):1–8. doi:10. 1007/s00521-011-0603-9 11. Loghmanian S, Jamaluddin H, Ahmad R, Yusof R, Khalid M (2012) Structure optimization of neural network for dynamic system modeling using multi-objective genetic algorithm. Neural Comput Appl 21(6):1281–1295. doi:10.1007/s00521-011-0560-3 12. Storn R, Price K (1997) Differential evolution-a simple and efficient heuristic for global optimization over continuous spaces. J Global Optim 11(4):341–359 13. Gandomi AH, Yang X-S, Talatahari S, Deb S (2012) Coupled eagle strategy and differential evolution for unconstrained and constrained global optimization. Comput Math Appl 63(1):191–200. doi:10.1016/j.camwa.2011.11.010 14. Khazraee S, Jahanmiri A, Ghorayshi S (2011) Model reduction and optimization of reactive batch distillation based on the adaptive neuro-fuzzy inference system and differential evolution. Neural Comput Appl 20(2):239–248. doi:10.1007/s00521-010-0364-x 15. Geem ZW, Kim JH, Loganathan GV (2001) A new heuristic optimization algorithm: harmony search. Simulation 76(2):60–68. doi:10.1177/003754970107600201 16. Wang H, Yuan X, Wang Y, Yang Y (2013) Harmony search algorithm-based fuzzy-PID controller for electronic throttle valve. Neural Comput Appl 22(2):329–336. doi:10.1007/s00521011-0678-3 17. Gholizadeh S, Barzegar A (2013) Shape optimization of structures for frequency constraints by sequential harmony search algorithm. Eng Optim 45(6):627–646. doi:10.1080/0305215x. 2012.704028 18. Karaboga D, Basturk B (2007) A powerful and efficient algorithm for numerical function optimization: artificial bee colony (ABC) algorithm. J Global Optim 39(3):459–471. doi:10.1007/s10898007-9149-x 19. Kennedy J, Eberhart R (1995) Particle swarm optimization. In: Paper presented at the proceeding of the IEEE international conference on neural networks, Perth, Australia, 27 Nov–1 Dec 20. Chen D, Zhao C, Zhang H (2011) An improved cooperative particle swarm optimization and its application. Neural Comput Appl 20(2):171–182. doi:10.1007/s00521-010-0503-4 21. Gandomi AH, Yun GJ, Yang XS, Talatahari S (2013) Chaosenhanced accelerated particle swarm algorithm. Commun Nonlinear Sci Numer Simul 18(2):327–340. doi:10.1016/j.cnsns.2012. 07.017 22. Back T (1996) Evolutionary algorithms in theory and practice. Oxford University Press, Oxford 23. Beyer H (2001) The theory of evolution strategies. Springer, New York 24. Dorigo M, Stutzle T (2004) Ant colony optimization. MIT Press, Cambridge 25. Gandomi AH, Yang X-S, Alavi AH (2011) Mixed variable structural optimization using Firefly Algorithm. Comput Struct 89(23–24):2325–2336. doi:10.1016/j.compstruc.2011.08.002 26. Cai X, Fan S, Tan Y (2012) Light responsive curve selection for photosynthesis operator of APOA. Int J Bio-Inspired Comput 4(6):373–379 27. Xie L, Zeng J, Formato RA (2012) Selection strategies for gravitational constant G in artificial physics optimisation based

307

28. 29.

30.

31.

32.

33.

34.

35.

36.

37.

38.

39.

40.

41.

42.

43.

44.

45.

46.

47.

on analysis of convergence properties. Int J Bio-Inspired Comput 4(6):380–391 Simon D (2008) Biogeography-based optimization. IEEE T Evol Comput 12(6):702–713. doi:10.1109/TEVC.2008.919004 Gandomi AH, Alavi AH (2011) Multi-stage genetic programming: a new strategy to nonlinear system modeling. Inf Sci 181(23):5227–5239. doi:10.1016/j.ins.2011.07.026 Yang XS, Deb S (2009) Cuckoo search via Le´vy flights. In: Proceeding of World congress on nature & biologically inspired computing (NaBIC 2009), Coimbatore, India, Dec 2009. IEEE Publications, USA, pp 210–214 Gandomi AH, Talatahari S, Yang XS, Deb S (2012) Design optimization of truss structures using cuckoo search algorithm. Struct Des Tall Spec. doi:10.1002/tal.1033 Gandomi AH, Yang X-S, Alavi AH (2013) Cuckoo search algorithm: a metaheuristic approach to solve structural optimization problems. Eng Comput 29(1):17–35. doi:10.1007/s00366011-0241-y Li X, Zhang J, Yin M (2013) Animal migration optimization: an optimization algorithm inspired by animal migration behavior. Neural Comput Appl 1–11. doi:10.1007/s00521-013-1433-8 Shumeet B (1994) Population-based incremental learning: a method for integrating genetic search based function optimization and competitive learning. Carnegie Mellon University, Pittsburgh Erol OK, Eksin I (2006) A new optimization method: Big BangBig Crunch. Adv Eng Softw 37(2):106–111. doi:10.1016/j. advengsoft.2005.04.005 Kaveh A, Talatahari S (2009) Size optimization of space trusses using Big Bang-Big Crunch algorithm. Comput Struct 87(17–18):1129–1140. doi:10.1016/j.compstruc.2009.04.011 Kaveh A, Talatahari S (2010) Optimal design of Schwedler and ribbed domes via hybrid Big Bang-Big Crunch algorithm. J Constr Steel Res 66(3):412–419. doi:10.1016/j.jcsr.2009.10.013 Kaveh A, Talatahari S (2010) A discrete big bang-big crunch algorithm for optimal design of skeletal structures. Asian J Civil Eng 11(1):103–122 Gandomi AH, Yang X-S, Alavi AH, Talatahari S (2013) Bat algorithm for constrained optimization tasks. Neural Comput Appl 22(6):1239–1255. doi:10.1007/s00521-012-1028-9 Yang XS, Gandomi AH (2012) Bat algorithm: a novel approach for global engineering optimization. Eng Comput 29(5):464–483. doi:10.1108/02644401211235834 Kaveh A, Talatahari S (2010) A novel heuristic optimization method: charged system search. Acta Mech 213(3–4):267–289. doi:10.1007/s00707-009-0270-4 Gandomi AH, Alavi AH (2012) Krill Herd: a new bio-inspired optimization algorithm. Commun Nonlinear Sci Numer Simulat 17(12):4831–4845. doi:10.1016/j.cnsns.2012.05.010 El-Abd M (2011) A hybrid ABC-SPSO algorithm for continuous function optimization. In: Swarm intelligence (SIS), 2011 IEEE symposium on, Paris, 11–15 Apr 2011. IEEE, pp 1–6. doi:10. 1109/SIS.2011.5952576 Wang G, Guo L (2013) A novel hybrid bat algorithm with harmony search for global numerical optimization. J Appl Math 2013:21. doi:10.1155/2013/696491 Duan H, Zhao W, Wang G, Feng X (2012) Test-sheet composition using analytic hierarchy process and hybrid metaheuristic algorithm TS/BBO. Math Probl Eng 2012:1–22. doi:10.1155/ 2012/712752 Gao X, Wang X, Jokinen T, Ovaska S, Arkkio A, Zenger K (2012) A hybrid PBIL-based harmony search method. Neural Comput Appl 21(5):1071–1083. doi:10.1007/s00521-011-0675-6 Geem ZW (2009) Particle-swarm harmony search for water network design. Eng Optim 41(4):297–311. doi:10.1080/ 03052150802449227

123

308 48. Gong W, Cai Z, Ling C (2010) DE/BBO: a hybrid differential evolution with biogeography-based optimization for global numerical optimization. Soft Comput 15(4):645–665. doi:10. 1007/s00500-010-0591-1 49. Kuo RJ, Syu YJ, Chen Z-Y, Tien FC (2012) Integration of particle swarm optimization and genetic algorithm for dynamic clustering. Inf Sci 195:124–140. doi:10.1016/j.ins.2012.01.021 50. Sun Y, Zhang L, Gu X (2012) A hybrid co-evolutionary cultural algorithm based on particle swarm optimization for solving global optimization problems. Neurocomputing 98:76–89. doi:10. 1016/j.neucom.2011.08.043 51. Sheikhan M, Mohammadi N (2012) Neural-based electricity load forecasting using hybrid of GA and ACO for feature selection. Neural Comput Appl 21(8):1961–1970. doi:10.1007/s00521-0110599-1 52. Marichelvam M (2012) An improved hybrid Cuckoo Search (IHCS) metaheuristics algorithm for permutation flow shop scheduling problems. Int J Bio-Inspired Comput 4(4):200–205. doi:10.1504/IJBIC.2012.048061 53. Wang G, Guo L, Wang H, Duan H, Liu L, Li J (2013) Incorporating mutation scheme into krill herd algorithm for global

123

Neural Comput & Applic (2014) 25:297–308

54. 55.

56.

57.

58.

59.

numerical optimization. Neural Comput Appl. doi:10.1007/ s00521-012-1304-8 Yao X, Liu Y, Lin G (1999) Evolutionary programming made faster. IEEE T Evol Comput 3(2):82–102 Yang X-S, Cui Z, Xiao R, Gandomi AH, Karamanoglu M (2013) Swarm intelligence and bio-inspired computation. Elsevier, Waltham Arslan M, C¸unkas¸ M, Sag˘ T (2012) Determination of induction motor parameters with differential evolution algorithm. Neural Comput Appl 21(8):1995–2004. doi:10.1007/s00521-011-0612-8 Li X, Yin M (2012) Application of differential evolution algorithm on self-potential data. PLoS ONE 7(12):e51199. doi:10. 1371/journal.pone.0051199 Jia L, Cheng D, Chiu M-S (2012) Pareto-optimal solutions based multi-objective particle swarm optimization control for batch processes. Neural Comput Appl 21(6):1107–1116. doi:10.1007/ s00521-011-0659-6 Zhang Y, Huang D, Ji M, Xie F (2011) Image segmentation using PSO and PCM with Mahalanobis distance. Expert Syst Appl 38(7):9036–9040. doi:10.1016/j.eswa.2011.01.041