Applied Soft Computing 28 (2015) 138–149 Contents lists available at ScienceDirect Applied Soft Computing journal homepage: www.elsevier.com/locate/asoc A new particle swarm optimization algorithm with adaptive inertia weight based on Bayesian techniques Limin Zhang a,b,c,∗ , Yinggan Tang a , Changchun Hua a , Xinping Guan a,b a b c Institute of Electrical Engineering, Yanshan University, Qinhuangdao 066004, China Department of Automation, Shanghai Jiao Tong University, Shanghai 200240, China Department of Mathematics and Computer Science, Hengshui University, Hengshui 053000, China a r t i c l e i n f o Article history: Received 1 April 2014 Received in revised form 15 October 2014 Accepted 20 November 2014 Available online 9 December 2014 Keywords: Particle swarm optimization Monte Carlo Gaussian distribution Bayesian techniques a b s t r a c t Particle swarm optimization is a stochastic population-based algorithm based on social interaction of bird flocking or fish schooling. In this paper, a new adaptive inertia weight adjusting approach is proposed based on Bayesian techniques in PSO, which is used to set up a sound tradeoff between the exploration and exploitation characteristics. It applies the Bayesian techniques to enhance the PSO’s searching ability in the exploitation of past particle positions and uses the cauchy mutation for exploring the better solution. A suite of benchmark functions are employed to test the performance of the proposed method. The results demonstrate that the new method exhibits higher accuracy and faster convergence rate than other inertia weight adjusting methods in multimodal and unimodal functions. Furthermore, to show the generalization ability of BPSO method, it is compared with other types of improved PSO algorithms, which also performs well. © 2014 Elsevier B.V. All rights reserved. 1. Introduction Particle swarm optimization (PSO) was firstly introduced by Kennedy and Eberhart in 1995 [1]. It belongs to evolutionary algorithm (EA), however differs from other evolutionary algorithms, which is inspired by the emergent motion of a flock of birds searching for food. PSO performs well in finding good solutions for optimization problems [2], and it has become another powerful tool besides other evolutionary algorithms such as genetic algorithms (GA) [3]. PSO is initialized with a population of particles randomly positioned in an n-dimensional search space. Every particle in the population has two vectors, i.e., velocity vector and position vector. The PSO algorithm is recursive, which motivates social search behavior among particles in the search space, where every particle represents one point. In comparison with other EAs such as GAs, the PSO has better search performance with faster and more stable convergence rates. Maintaining the balance between global and local search in the course of all runs is critical to the success of an optimization algorithm [4]. All of the evolutionary algorithms use various methods to achieve this goal. To bring about a balance between the two ∗ Corresponding author at: Institute of Electrical Engineering, Yanshan University, Qinhuangdao 066004, China. Tel.: +86 18230357919. E-mail address: limin [email protected] (L. Zhang). http://dx.doi.org/10.1016/j.asoc.2014.11.018 1568-4946/© 2014 Elsevier B.V. All rights reserved. searches, Shi and Eberhart proposed a PSO based on inertia weight in which the velocity of each particle is updated according to a fixed equation [5]. A higher value of the inertia weight implies larger incremental changes in velocity, which means the particles have more chances to explore new search areas. However, smaller inertia weight means less variation in velocity and slower updating for particle in local search areas. In this paper, the inertia weight strategies are categorized into three classes. The first class is simple that the value of the inertia weight is constant during the search or is selected randomly. In [6], the impact of the inertia weight is analyzed on the performance of the PSO. In [7], Eberhart and Shi use random value of inertia weight to enable the PSO to track the optima in a dynamic environment. In the second class, the inertia weight changes with time or iteration number. We name the strategy as time-varying inertia weight strategy. In [8,9], a linear decreasing inertia weight strategy is introduced, which performs well in improving the fine-tuning characteristic of the PSO. Lei et al. use the Sugeno function as inertia weight declined curve in [10]. Many other similar linear approaches and nonlinear methods are applied in inertia weight strategies such as in [11–13]. The last class is some methods that inertia weight is revised using a feedback parameter. In [4], a fuzzy system is proposed to dynamically adapt the inertia weight. In [14], the inertia weight is determined by the ratio of the global best fitness and the average of particles’ local best fitness in each iteration. In [15], A new strategy is presented that the inertia weight is dynamically L. Zhang et al. / Applied Soft Computing 28 (2015) 138–149 adjusted according to average absolute value of velocity, which follows a given nonlinear ideal velocity by feedback control, which can avoid the velocity closed to zero at the early stage. In [16], dynamic acceleration parameters (DAP) method is proposed, which owns a mechanism to self-tune the acceleration parameters by utilising the averaged velocity information of the particles. In [17], a new adaptive inertia weight strategy is proposed based on the success rate of the particles. In this strategy the success rate of the particles is used as a feedback parameter to realize the state of the particles in the search space and hence to adjust the value of inertia weight. A series of other studies based on mutation strategy has been done on the analysis and development of the PSO since it was introduced in 1995. These approaches aim to improve the PSO’s convergence velocity. In [18], a mutation operator is used that changes a particle dimension value using a random number drawn from a Gaussian distribution (GPSO). A particle is selected for mutation using a mutation rate that is linearly decreased during a run. In [19], a mutation strategy is proposed that a particle position is changed using a random number drawn from a Gaussian distribution. A mutation operator in [20] is similar to that of Ref. [18], but a Cauchy probability distribution is used instead (CPSO). The Cauchy distribution curve is similar to the Gaussian distribution curve, except it has more probability in its tails and thus making it more likely to return larger values. In [21], the HPSO with a wavelet mutation (HWPSO) is proposed, in which the mutation incorporates with a wavelet function. Hybrid PSOs (HPSOs) have been proposed to enhance the performance of the PSO, in which different mutation strategies are used. In [22], premature convergence is avoided by adding a mutation strategy, i.e., a perturbation to a randomly selected particle’s velocity vector. A Cauchy mutation (HPSO) [23] is proposed, which is used in best particle mutation, so that the best particle could lead the rest of the particles to better positions. The algorithm in [24] called IPSO + ACJ is tested on a suite of well known benchmark multimodel functions and the results. The main idea of our new jump strategy is that pbest and gbest are selected as mutated particles when they have not been improved in a predefined number of iterations. In [25], Ant colony optimization (ACO) and PSO work separately at each iteration and produce their solutions. In [26], The new mutation strategy makes it easier for particles in hybrid MRPSO (HMRPSO) to find the global optimum and seek a balance between the exploration of new regions and the exploitation of the old regions. Fuzzy approaches for PSO is a hot topic in these years. In [27], a fuzzy system is used to dynamically adjust the inertia weight and learning factors of PSO in each topology. In [28], a dynamic parameter adaptation is proposed to improve the convergence and diversity of the swarm in PSO using fuzzy logic. Valdezis et al. [29] introduced an improved FPSO + FGA hybrid method, which combines the advantages of PSO and GA. See Ref. [30] for a comprehensive review on the application of fuzzy logic in PSO, ACO and Gravitational Search Algorithm (GSA). There are some works that have used the Bayesian technique in smart computing. In [31], the Dynamic Bayesian Network (DBN) is used in PSO algorithm for particle motion. In swarm optimization, each particle tries to track the trajectory toward the place of better fitness. Martens et al. [32] explained scientifically the application of the Bayesian network in ACO. In our paper, Bayesian techniques is introduced into the PSO algorithm with a view to enhance its adaptive search ability. We call this algorithm as PSO with Bayesian techniques (BPSO). Different from other inertia weight strategies in the references, BPSO algorithm adjusts the inertia weight ω based on the past particle places automatically. This new development gives particles more opportunity to explore the solution 139 space than a standard PSO. Furthermore, the new algorithm accelerates search velocity for the particles in valuable search-space regions. This paper is organized as follows. In Section 2, generic PSO theory is reviewed and the change of particle position ε is analyzed. Section 3 introduces the advantage of PSO with Bayesian Techniques. Section 4 shows the experimental settings for the benchmarks, simulation results and parameters analysis in the BPSO. Finally, Section 5 presents conclusions resulting from the study. 2. Particle swarm optimization 2.1. Generic PSO theory PSO is also a population-based stochastic optimization algorithm and starts with an initial population of randomly generated solutions called particles. Each particle in PSO has a position and a velocity. PSO remembers both the best position found by all particles and the best positions found by each particle in the search process. For a search problem in an n-dimensional space, a potential solution is represented by a particle that adjusts its position and velocity according to Eqs. (1) and (2): vi,d (t + 1) = ωvi,d (t) + c1 r1 (Ppid − xi,d (t)) + c2 r2 (Pgd − xi,d (t)) (1) xi,d (t + 1) = xi,d (t) + vi,d (t + 1) (2) where c1 and c2 are two learning factors which control the influence of the social and cognitive components and ri = randi , (i = 1, 2) are numbers independently generated within the range of [0,1]. vi,d (t) is the velocity of individual i on dimension d. xi,d (t) is current particle position on dimension d. Ppid (pbest) is the best local position of individual i on dimension d, and Pgd (gbest) represents the best particle position among all the particles in the population on dimension d. ω is the inertia weight, which ensures the convergence of the PSO algorithm. According to Eq. (2), changes in the position of particles depend exclusively upon the position item of the PSO. Therefore, we only use the position item to investigate the search ability of particles during iterations. The implicit form of the position equation presented in Eq. (2) is used for a multi-particle PSO working in a multi-dimensional search space. Since the data of each dimension in the PSO are independent, the analysis below will be restricted to a single dimension. We simplify the Eqs. (1) and (2) as follow: V (t + 1) = ωV (t) + c1 r1 (Pp − X(t)) + c2 r2 (Pg − X(t)) (3) X(t + 1) = X(t) + V (t + 1) (4) The following formulas can be obtained from Eq. (4): V (t + 1) = X(t + 1) − X(t), V (t) = X(t) − X(t − 1) (5) By substituting Eq. (5) into Eq. (3), the following nonhomogeneous recurrence relation is obtained: X(t + 1) = (1 + ω − c1 r1 − c2 r2 )X(t) − ωX(t − 1) + (c1 r1 Pp − c2 r2 X(t)Pg ) (6) X(t + 1) = (1 + ω)X(t) − ωX(t − 1) + (c1 r1 Pp + c2 r2 Pg ) − (c1 r1 + c2 r2 )X(t) (7) Let ε = (c1 r1 Pp + c2 r2 Pg ) − (c1 r1 + c2 r2 )X(t), then X(t + 1) = (1 + ω)X(t) − ωX(t − 1) + ε (8) 140 L. Zhang et al. / Applied Soft Computing 28 (2015) 138–149 positions in s iteration, ω is used to accelerate up the movement of particles towards the optimum point or slow down the particles so that they converge to the optimum. Write Eq. (8) as follow X(t + 1) = W1 X(t) + W2 X(t − 1) + ε (9) ]T , Let W1 = (1 + ω), W2 = − ω, W = [W1 , W2 t = [X(t), X(t − 1)]T , T = [t , t+1 , · · · , t+s−1 ] , s is the constant. Y = W, H = [X(t + 1), X(t + 2), · · · , X(t + s)]T , Eq. (9) is written as H = Y + ε = W + ε -1 -7E-10 -7E-40 0 7E-40 7E-10 1 Fig. 1. Probability distribution histogram of ε using the Monte Carlo method. The first and second items on the right side of Eq. (8) both memorize the past values of position which have different weight. They also can be seen as a whole item to be called the memory item of position. Since the value of the third item on the right side of Eq. (8) is obtained from the previous experience of particles, we call it the learning item of position. According to ε, (c1 r1 + c2 r2 )X(t) in ε is the memory part and (c1 r1 Pp + c2 r2 Pg ) is the learning part. It is obvious that ε represents the change of particle position in every iteration. 2.2. Analysis of the change of particle position ε ε represents the gap between memory part and learning part. If the particle position converges to one position, ε is a Gaussian distribution. We can use Monte Carlo method to illustrate that. Monte Carlo is the art of approximating an expectation by the sample mean of a function of simulated random variables. It is apparently first used by Ulam and von Neumann as a Los Alamos code word for the stochastic simulations applied to building better atomic bombs. Monte Carlo methods are also online simulation methods that learn from experience based on randomly generated simulations. Given a random set of experiences, with the guarantee from the weak law of large numbers, the simulation results will eventually converge when each state is encountered for an infinite number of trials. It is evidenced in part by the voluminous literature of successful applications. In our paper, we use Monte Carlo method to analyze the properties of ε, respectively, using a large body of experimental data. Take the test function f6 as the example, the PSO is initialized as follow: In this paper, we propose a Bayesian PSO method that considers a probability density function over the weight space, and the optimal inertia weight vector is calculated by maximizing the posterior probability density function of the weight, where ε is an independent stochastic variable satisfying Gaussian distribution with zero mean value and variance 2 , i.e. (p(ε) ∼ N(ε|0, 2 )) Let D = {H, W}, ˇ = 1/ 2 , Then, the Bayesian method can be used to estimate the weight vector W, in which the probability density function over the weight vectors space is mainly considered. In general, the priori information on the weight vector W is difficult to be found, so we choose a fairly broad distribution as a priori. The expression of this distribution is Gaussian distribution, which is written as 3. PSO with Bayesian techniques 3.1. PSO with Bayesian techniques The inertia weight adjusting method proposed in our paper is inspired by the idea of self-adaption. Based on past particle 1 exp(−˛EW ) ZW (˛) P(W ) = (11) 1 1 2 Wi ||W ||2 = 2 2 M EW = (12) i=1 where ˛ controls the distribution of weight vector. ZW (˛) is a normalization factor. M is the total number of weights. If W is subject to Gaussian prior, the normalization factor ZW (˛) can be expressed by ZW (˛) = 2 M/2 (13) ˛ The likelihood function can be written as P(D|W ) = 1 exp(−ˇED ) ZD (ˇ) (14) where ˇ controls the distribution of Y, ZD (ˇ) is a normalization factor. ED is an error function. P(D|W) can also be written as P(D|W ) = P(H|(W, )) = 1 = exp ZD (ˇ) (1) The maximum number of iterations is 500. (2) The swarm size is 50. (3) c1 and c2 are set as 1.49445. In the experiment, We obtain the probability distribution of ε shown in Fig. 1. Under the condition of pbest = gbest, the value of gbest in the 50 particles is not changed after about 200 iterations. The mean value of ε is equal to zero and the variance is 2 , i.e., ε is an independent stochastic variable satisfying Gaussian distribution with zero mean value and variance 2 (p(ε) ∼ N(ε|0, 2 )). (10) ZD (ˇ) = e(˛ED ) = s P(X(t + i)|(W, )) i=1 ˇ 2 − {X(t + i) − i W } 2 s (15) i=1 2 2s ˇ (16) using formula P(W |D) = P(D|W )P(W ) P(D) (17) By substituting Eqs. (11) and (14) into Eq. (17), we obtain the posterior distribution as follows P(W |D) = 1 1 exp(−ˇED − ˛EW ) = exp(Q (W )) ZQ ZQ (18) where Q (W ) = ˇED + ˛EW (19) L. Zhang et al. / Applied Soft Computing 28 (2015) 138–149 ZQ (˛, ˇ) = exp(−ˇED − ˛EW )dW (20) To obtain the posterior distribution of W, we apply Gaussian approximation to W. This approximation is obtained through the second-order Taylor expansion of Q(W) around its most probable weights value WMP . Taylor expansion of Q(W) is written as Q (W ) = Q (WMP ) + 1 (W − WMP )T A(W − WMP ) 2 (21) where A is the Hessian matrix of the total (regularized) error function. The remainder terms of Q(W) have been ignored. 141 3.3. W estimation By maximizing the posterior distribution Eq. (28), we maximize its logarithm equivalently, which results in the following optimization problem: Q (W ) = ˇED + ˛EW = ˇ ˛ T (W − H) (W − H) + W T W 2 2 (29) By fixing the hyperparameters ˛ and ˇ, we maximize Q(W) with respect to W, and have 3.2. Parameters selection of ˛, ˇ According to the content we have discussed, parameters ˛, ˇ are unknown. Take parameters ˛, ˇ as integral variables. The posterior distribution of network weights is given by ∂(W − H)T ∂(W − H)T (W − H) + (W −H) ∂W ∂W + ˛W = P(W |D) = ∂Q (W ) ˇ = 2 ∂W P(W, ˛, ˇ|D)d˛dˇ ˇ {2T (W − H)} + ˛W = ˇT W − ˇT H 2 + ˛W = ˇT W − ˇT H + ˛W (30) = P(W |˛, ˇ, D)P(˛, ˇ|D)d˛ dˇ (22) taking ∂Q (W ) = 0, then W can be written as ∂W Since the posterior probability distribution P(W, ˛, ˇ|D) for the hyper-parameters is around their most probable values ˛MP and ˇMP [33]. Replace ˛, ˇ with ˛MP , ˇMP . Eq. (22) can be written as P(W |D) ≈ P(W |˛MP , ˇMP , D) 3.4. BPSO algorithm and its computational cost (24) by combining Eq. (21) with (24), P(D|˛, ˇ) can be written as 1 ZD (ˇ)ZW (˛) exp(−Q (W ))dW = ZM (˛, ˇ) ZD (ˇ)ZW (˛) (25) considering Eq. (25), the log of the evidence is then given as follow MP MP ln P(D|˛, ˇ) = −ˇED − ˛EW − , 2EW N N ln(ˇ) + ln(2) 2 2 ˇMP = n− 2ED (26) (27) where = 2 i i=1 i + ˛ , (i |i = 1, 2) The main process of BPSO algorithm is in the Algorithm 1. In the algorithm, inertia weight ω can change according to the past particle position. In every s iterations, ω is different in Eq. (3). However, the Bayesian approach suffers from local optimal problem. By introducing a mutation operation into BPSO, the ability to escape from local optimum is enhanced and the global search ability of the BPSO algorithm is strengthened. In [18], we see that Gaussian mutation tends to have a fine-tuning ability, however, it only produces short jump. To produce longer jumps, the cauchy mutation is selected as the local mutation operator, which is written as 1 L ln(det A) + ln(˛) 2 2 The optimal hyper-parameters values ˛MP , ˇMP are obtained by differentiating Eq. (26) [34]. The optimal values are as follow ˛MP = (31) (23) P(D|˛, ˇ)P(˛, ˇ) P(D) + ˇT H using the W, we can compute the inertia weight ω. using the formula as follow P(D|˛, ˇ) = −1 P(˛, ˇ|D)d˛dˇ = P(W |˛MP , ˇMP , D) P(˛, ˇ|D) = W = (ˇT + ˛) (28) i are the eigenvalues of the Hessian matrix of the un-regularized error. Pp,i = Pp,i + ˛ · cauchy(1) (32) Pp,i represents the best place of the ith particle in the iteration. cauchy(1) returns a big value, which is drawn from a cauchy distribution. Parameter ˛ controls the range of cauchy random number. The computation cost of the original PSO algorithm can be easily obtained, which is represent as O(maxgen * sizepop). sizepop is the size of the whole population, and maxgen is the maximum number of iterations. In BPSO algorithm, the total computational cost incurred by the computation of the W is denoted as O(maxgen * sizepop * s). s is regrouping period, which is less than 100. In our experiment, maxgen * sizepop is far greater than s, so the total computational cost of BPSO is the same as the original PSO algorithm. 142 L. Zhang et al. / Applied Soft Computing 28 (2015) 138–149 4. Simulations Algorithm 1 The details of BPSO. Initialize: maxgen: max iterations; i: the number of iterations; sizepop: size of the whole population; j: the number of particle; s: regrouping period; Pp,i : the best position of one particle in one iteration; Pg : the best position of all particles in one iteration; Initialize each particle’s position xi,d and constant c1 = c2 = 1.49445; regrouping period s = 30; cout(j): the best position of all particles in one iteration; q: the threshold; Iterate: 1: for i = 1 : maxgen do for j = 1 : sizepop do 2: Update each particle using local version PSO (1); 3: if mod(i,s)=0 and i ≥ 2then 4: 5: compute the weight W using (31); 6: end if end for 7: for j = 1 : sizepop do 8: if f(xj,d ) < f(Pp,j )then 9: f(Pp,j ) = f(xj,d ) 10: else 11: cout(j) = cout(j) + 1 12: end if 13: 14: if cout(j) > qthen mutate Pp,j using (32) 15: 16: cout(j) = 0 end if 17: if f(xj,d ) < f(Pg )then 18: f(Pg ) = f(xj,d ) 19: end if 20: 21: end for end for 22: Two comparisons have been conducted in this section. One is among unimodal functions and the other is among multimodal functions. 4.1. Experiment setup To test the performance of the proposed algorithm, we do experiments on 10 benchmark functions. These test functions, which are shown in Table 1, can be classified into two groups. The first five functions f1 –f5 are unimodal functions. For unimodal functions, the convergence rate of an algorithm is more interesting than the final result of optimization. Therefore, we not only give the final achievement but also show the convergence rate of each algorithm. The next five functions f6 –f10 are multimodal functions with many local optima. For multimodal functions, we only give the final result since it reflects an algorithm’s ability of escaping from poor local optima. The functions are used to test the global search ability of the algorithm in avoiding premature convergence. In this paper, all empirical experiments related to the PSO and its improvements are carried out with a same population size. Furthermore, in order to ensure that the initial values of particles in each algorithm are same, we use the MATLAB command rand(state, sum(i30)). The initial value of i of each run for all algorithms is the same. Parameter s is the interval of the adjacent two inertia weight change in all iterations. The parameters of the BPSO are used as follow. Table 1 Test function. Test function n 2 f1 = i=1 xi n 2 f2 = exp 0.5 x −1 i=1 i n 2 ([xi + 0.5]) f3 = i=1 n n f4 = |xi | + |xi | i=1 i=1 n 2 f5 = ixi i=1 n xi sin( |xi |) f6 = n ∗ 418.98291 − i=1 n 2 (xi − 10 cos(2xi ) + 10) f7 = i=1 n 1 f8 = −20 exp −0.2 xi2 n 1 n i=1 cos(2xi ) + 20 + e − exp n ) 2i=1 n−1 2 2 2 2 (xi − 1) ∗ [1 + sin (xi+1 )] + (xn − 1) [1 + sin (xn )] f9 = 0.1 sin (3xi ) + i=1 √ n n 1 2 f10 = 4000 i=1 xi − (a) i=1 cos(xi / i) + 1 Domain Global optimum Name [−100,100] 0 Sphere [−1,1] 0 Hartmann [−100,100] 0 Step [−100,100] 0 Schwefel’s p 2.22 [−100,100] 0 Sum Squares [−500,500] 0 Schwefel [−5.12,5.12] 0 Rastrigin [−32,32] 0 Ackley [−500,500] 0 Levy–Montalvo [−600,600] 0 Griewank (b) Fig. 2. Comparisons between different inertia weight strategies for f1 –f2 . L. Zhang et al. / Applied Soft Computing 28 (2015) 138–149 143 Fig. 3. Comparisons between different inertia weight strategies for f3 –f4 . From all strategies introduced in the previous sections, four strategies have been adopted for comparisons: decreasing inertia weight (Linear Decreasing) [5,6], random inertia weight (Random) [7],the Sugeno method (Sugeno) [10] and AIWPSO [17]. To further validate the effectiveness of the proposed algorithm in this paper, we compare BPSO with six existing PSO variants, including the PSO [5], PSO with Gaussian mutation (GPSO) [19], PSO with Gaussian mutation cauchy (CPSO)[19], hybrid particle swarm (HPSO) [23], improved PSO with a jump strategy(IPSO + AJS) [24], PSO with an MRS strategy (HMRPSO) [26] under the same maximum function evaluations (FEs). 4.2. Comparison results on unimodal functions Four inertia weight strategies are applied to the five test function f1 –f5 on dimensions 10 and 50. The results are listed in Tables 2 and 3, where “best” indicate the best function values found in the generation, “Mean” and “Std Dev” stand for mean value and standard deviation. Figs. 2–4 are comparison among BPSO and other strategies on dimension 10. The sphere function f1 is unimodal test function, which is often used. For this function, the performance of the BPSO is much more better than the decreasing inertia weight, random inertia weight, the Sugeno method and AIWPSO. Similar results are also obtained on f4 and f5 , which are shown in Tables 2 and 3. The differences between the inertia weight adjusting methods on f1 , f4 , and f5 suggest that the BPSO is better at a fine-gained search than both the other algorithms. f3 is called step function, which is characterized by plateaus and discontinuity. From Tables 2 and 3, using decreasing inertia weight, random inertia weight and the Sugeno method, optimal function value can be found. However, in Fig. 3(a), we can see that the convergence rate of the BPSO is faster than the other two PSO inertia weight adjusting methods. The performance of the BPSO is much more better than the other four methods. The main differences attribute to the change of the inertia weight ω, which affords the BPSO a greater chance to search in the local area. In other words, the feature of ω directly affects the particle’s movement, so that the particle preserve the fast converging feature in its searching area. The above mentioned PSO-based optimization algorithms (PSO, GPSO, CPSO, IPSO + AJS, HPSO, HMRPSO) are applied to the test problems. Table 4 shows the comparisons of six other PSO variants for function f1 –f5 . From Table 4 and Figs. 5– 7, BPSO performs better than the other three improved variants. Inf2 and f3 , globle optimum both can be found using BPSO and HMRPSO algorithm. But, in Figs. 5(b) and 6(a), we can see that the convergence rate of the BPSO is faster than HMRPSO. 4.3. Comparison results on multimodal functions On multimodal functions, the global optimum is more difficult to locate. Therefore, in the comparison, we study the accuracy, speed, and reliability of the PSO variants. Comparisons of solution accuracy on multimodal functions are given in Tables 2, 3 and 5. The Schwefel function f6 is a multimodal function with very deep sinusoidal indentations. It is one of the most difficult problems for optimization algorithms. It can be seen from the comparisons in Tables 2, 3 and 5 that, although none of the seven algorithms considered in this paper is efficient in solving this function, BPSO method is more efficient than the other four inertia weight adjusting methods and six PSO variant algorithms in producing the near optimal value. 50 10 0 10 Mean of average fitness (1) The number of runs is 500; the acceleration coefficients c1 = c2 =1.49445. (2) The population size is 50. (3) The parameter s of BPSO increase from 1 to 50, the step length is 2. −50 10 −100 10 −150 10 0 10 −200 10 −250 10 BPSO AIWPSO Decreasing Random Sugeno 200 205 210 −300 10 0 100 200 300 400 500 Iteration Fig. 4. Comparisons between different inertia weight strategies for f5 on 10 dimensions. 144 L. Zhang et al. / Applied Soft Computing 28 (2015) 138–149 Table 2 Comparisons between different inertia weight adjusting PSO methods (n = 10). Function n f1 10 f2 10 f3 10 f4 10 f5 10 f6 10 f7 10 f8 10 f9 10 f10 10 Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev BPSO AIW Decrease Random Sugeno 3.37E−271 1.60E−226 0.00E+00 0 0 0 0 0 0 1.93E−132 1.54E−122 1.47E−122 9.11E−277 9.43E−245 2.45E−241 0.0018 0.0020 0.0013 0 0 0 1.88E−16 8.88E−16 3.88E−14 4.28E−08 1.42E−06 1.61E−06 0 0 0 4.51E−01 7.61 7.64 7.56E−03 4.67E−03 0.004 2 2.5 0.71 16.28 3.21E+01 22.31 13.5 13.3 0.36 3.70E+03 3.73E+03 53.78 26.12 53.73 38.71 19.3 19.8 0.38 2.94 4.35 2.02 0.082 0.196 0.148 1.62E−09 5.55E−07 9.44E−07 6.20E−14 2.73E−10 3.86E−10 0 0 0 0.093 3.02 4.14 2.88E−07 2.67E−05 3.73E−05 3.73E+03 3.78E+03 75.16 11.94 18.24 6.47 19.96 20 0.0178 1.01E−05 3.70E−03 0.0063 0.062 9.68E−02 0.0317 1.39E−10 2.37E−02 4.11E−02 7.15E−04 6.48E−04 9.522E−05 0 0 0 0.012 3.13E−02 0.027 7.49E−09 4.84E−05 6.84E−05 3.69E+03 3.71E+03 34.04 42.8 68.12 21.95 1.65 7.95 10.58 2.07E−04 2.41 3.58 0.029 0.105 0.066 3.01E−09 1.87E−08 1.43E−08 7.39E−10 3.49E−09 3.89E−09 0 0 0 0.45 0.862 0.585 2.01E−05 6.39E−05 6.19E−05 3.73E+03 3.80E+03 63.65 12.93 25.54 18.49 2.58 14.2 10.06 2.63E−05 3.77E−03 0.0063 0.064 8.45E−02 0.029 The generalized Rastrigin’s function f7 can easily trap an optimization algorithm in a local optimum on its way to the global optimum. By adjusting step length s, particles in the BPSO are able to search more solution space. In Tables 2, 3 and 5, it can be clearly seen that the mean values of the BPSO are better than those of the other variants algorithms. The Ackley function f8 has several local minima. Here, the mean values and the standard deviations of the BPSO are also better than those of the inertia weight adjusting methods and six PSO variant algorithms. Therefore, it is evident that the BPSO can provide more stable and high accuracy results on this function. Table 3 Comparisons between different inertia weight adjusting PSO methods (n = 50). Function n f1 50 f2 50 f3 50 f4 50 f5 50 f6 50 f7 50 f8 50 f9 50 f10 50 Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev BPSO AIW Decrease Random Sugeno 2.46E−248 1.73E−218 0.00E+00 0 0 0 0 0 0 2.42E−123 6.36E−116 1.22E−120 6.75E−216 1.35E−216 2.33E−212 119 343 283 0 0 0 1.88E−16 8.88E−16 3.88E−14 4.33 4.61 0.27 0 0 0 1.14E+03 1.59E+03 4.00E+02 6.58E−01 9.01E−01 0.34 1149 1443 168 327 373 14 3.21E+04 5.38E+04 2.48E+04 1.95E+04 1.96E+04 71 2116.25 2421.16 269.13 20.6 19.6 1.5 283 343 67.72 1.21 1.32 0.09 5.86E+01 1.43E+02 6.62E+01 7.15E−02 8.83E−01 1.15 263 391 165 383 417 46 3.33E+03 4.16E+03 1.16E+03 1.92E+04 1.91E+04 251 859.94 1195.80 318 17.1 19.3 0.06 142 162 78.18 0.85 0.95 0.1 1.37 1.06E+02 1.81E+02 3.50E−01 4.64E−01 0.16 21 498 675 365 441 94 1.01E+03 8.38E+03 9218 1.95E+04 1.97E+04 190 1.07E+03 1387.65 425.37 17.2 18.8 1.77 138 181 86.05 1.09 1.45 0.08 3.55E+02 3.95E+02 3.95E+01 5.73E−01 1.15 0.21 162 185 32 343 373 42 3.83E+03 4.13E+03 414 1.90E+04 1.91E+04 75 833.94 1028.68 241.35 19.5 20 0.44 309 401 81.29 1.07 1.09 0.01 L. Zhang et al. / Applied Soft Computing 28 (2015) 138–149 145 Table 4 Comparisons between different PSO methods for unimodal test function. Function n f1 10 30 50 f2 10 30 50 f3 10 30 50 f4 10 30 50 f5 10 30 50 Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev BPSO PSO GPSO CPSO HPSO IPSO + ACJ HMRPSO 3.37E−271 1.60E−226 2.60E−224 1.33E−263 2.53E−165 148E−166 2.46E−248 1.73E−218 0.00E+00 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1.93E−132 1.54E−122 1.47E−122 4.24E−130 9.60E−104 1.662E−103 2.42E−123 6.36E−116 1.22E−120 9.11E−277 9.43E−245 2.45E−241 3.42E−264 9.15E−228 1.68E−228 6.75E−216 1.35E−216 2.33E−212 7.05E−10 1.04E−06 1.51E−06 1.17E+01 1.06E+01 1.96E+00 1.50E+02 1.56E+02 28.1 9.70E−09 1.17E−07 1.35E−07 5.64E−02 0.078 0.040 0.13 0.54 0.41 0 0 0 15 36.5 18.44 213 352.5 162.48 0.083 4.35E−01 0.335 2.46E+02 2.52E+02 8.618 418.19 447 33.77 4.59E−09 1.55E−06 2.51E−06 2.31E+02 4.12E+02 310.75 3.00E+03 4.36E+03 1426.56 0.01 3.40E−02 1.41E−02 23.22 2.45E+01 1.39E+00 131.43 182 35.8 0.023 6.24E−02 0.035 5.44E−07 9.61E−07 1.12E−06 7.92E−06 2.23E−06 1.28E−06 0 0 0 27 31 3.78 240 326.5 107.46 0.284 1.40E+01 18.55 2.07E+02 2.77E+02 73.67 1E+11 2.33E+16 4.6E+16 0.0878 4.83E−01 0.268 5.65E+02 6.44E+02 70.2 4299.33 5.47E+03 1205.36 8.03E−06 9.06E−05 1.22E−04 1.09E−03 3.35E−03 3.87E−03 2.09E−03 7.73E−03 4.06E−03 4.47E−08 5.7E−05 6.33E−05 1.63E−10 1.11E−07 1.61E−07 1.63E−8 1E−07 1.21E−07 0 0 0 0 0 0 0 0 0 0.0071 3.67E−02 0.043 0.0622 8.67E−02 0.0352 0.13 0.23 0.19 6.14E−09 1.84E−04 0.00025 6.54E−05 7.37E−03 0.0114 1.73E−03 2.00E−02 0.031 1.21E−13 4.99E−09 1.10E−08 1.05E+01 1.26E+01 3.42E+00 1.14E+02 1.35E+02 1.64E+01 8.55E−14 1.27E−10 2.04E−10 2.32E−02 0.098 0.14 0.59 1.13 0.57 0 0 0 50 56 7.93 218 328 119.25 0.184 8.29E−01 0.405 1.50E+02 2.37E+02 78.91 384.63 447 63.99 2.24E−07 3.19E−04 0.00046 4.81E+02 6.50E+02 147.08 2.57E+03 4.88E+03 1681.89 9.55E−08 2.85E−06 5.88E−06 9.20E+00 2.36E+01 1.31E+01 1.51E+02 2.41E+02 8.36E+01 9.11E−12 2.08E−10 2.36E−10 4.26E−02 0.054 0.043 0.17 1.61 1.71 0 0 0 38 91.25 80.78 38 250 101.28 0.157 6.06E+01 52.5 1.54E+02 2.05E+02 44.99 288.19 418 148.02 5.78E−07 2.53E−04 0.00047 4.13E+02 7.46E+02 406.54 5.45E+03 6.84E+03 987.69 1.57E−130 3.18E−126 7.10E−126 1.52E−126 5.84E−121 4.43E−121 1.53E−125 8.81E−121 6.43E−121 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1.82E−83 1.94E−80 3.82E−80 9.16E−83 1.75E−81 2.83E−81 9.45E−83 1.72E−81 3.63E−81 3.77E−175 2.69E−165 0 3.65E−170 8.93E−161 0 2.12E−166 5.32E−158 0 Fig. 5. Comparisons between different PSO methods for f1 –f2 . 146 L. Zhang et al. / Applied Soft Computing 28 (2015) 138–149 Fig. 6. Comparisons between different PSO methods for f3 –f4 . The Levy-Montalvo function f9 has approximately 15n local optima. In Table 5, when the number of dimensions is increased from 10 to 50, the PSO has been trapped in local optima. Meanwhile, the BPSO performs much better principally because the changed inertia weight ω affords the BPSO a greater opportunity to escape from poor local optima. f10 is the Griewank function. Because f10 is a difficult multimodal function that has an exponentially increasing number Table 5 Comparisons between different PSO methods for multimodal test function. Function n f6 10 30 50 f7 10 30 50 f8 10 30 50 f9 10 30 50 f10 10 30 50 Best Mean Dev Best Mean Dev Best Mean Dev est Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev Best Mean Dev BPSO PSO GPSO CPSO HPSO IPSO+ACJ HMRPSO 0.0018 0.0020 0.0013 0.011 0.132 0.0108 119 343 283 0 0 0 0 0 0 0 0 0 1.88E−16 8.88E−16 3.88E−14 8.88E−14 7.88E−13 1.21E−12 1.88E−16 8.88E−16 3.88E−14 4.28E−08 1.42E−06 1.61E−06 1.49E−06 3.52E−06 3.06E−06 4.33 4.61 0.27 0 0 0 0 0 0 0 0 0 1.81E+03 1919.25 242.73 8.87E+03 8932.18 89.12 1.43E+04 1.53E+04 826.41 5.96 10.14 3.17 5.53E+01 72.92 20.12 124.52 147.03 29.58 1.13E−05 2.01 0.93 4.62 6.22 1.41 7.88 8.74 0.61 0.12 5.27E+01 76.13 1.08E+03 1.78E+03 642.13 1.47E+03 3.18E+03 1284.62 0.099 5.05E+00 9.54 1.48 1.55 0.054 3.11 3.65 0.64 1.25E+03 2.03E+03 534.71 7.56E+03 8051.24 577.66 1.51E+04 1.62E+04 851.98 2.09 11.27 8.66 1.46E+01 54.22 37.63 90.62 173.61 81.32 0.51 1.33 0.49 5.86 6.25E+00 0.51 7.23 8.15E+00 0.63 0.11 9.18E+02 203.1 1.34E+03 2.95E+03 2500.06 4.82E+03 5.24E+03 422 0.031 0.26 0.87 1.11 1.18 0.072 1.72 2.51 0.75 0.022 502.55 599.31 2.24E−03 0.27 0.46 1.14E−03 2.05E−03 1.14E−03 4.88E−06 5.07E−05 5.16E−05 7.96E−06 0.00019 0.00029 8.33E−05 0.0049 0.007 9.68E−04 7.60E−03 0.011 0.0011 8.75E−04 0.00027 0.009 1.32E−02 0.01 0.0031 6.97E−03 0.0071 3.67E−04 1.31E−02 0.018 7.45E−03 1.11E−02 0.04 4.01E−04 8.06E−04 0.00091 2.50E−03 5.90E−03 0.0069 4.65E−03 7.09E−03 0.01 1.58E+03 2075.33 361.05 6.81E+03 7706.27 1049.47 1.58E+04 1.62E+04 356.62 6.96 14.12 6.02 4.40E+01 60.31 15.83 148.04 213.82 73.58 7.10E−05 1.12 1.06 3.34 4.41 0.41 6.69 8.44 1.17 0.36 0.41 0.31 6.18E+02 8.48E+02 293.36 3.09E+03 8.23E+02 644.95 0.097 2.21E+01 43.078 1.132 1.138 0.12 2.33 2.99 0.79 1.54E+03 1810.02 259.83 7.70E+03 7.43E+03 231.53 1.18E+04 1.34E+04 1412.65 9.94 17.51 7.88 7.08E+01 113.61 38.11 223.65 26.2 38.11 2.97E−05 7.94E−01 1.11 4.11 5.68 1.62 7.94 8.54 0.47 1.64E−02 6.64E−02 9.64E−02 1.34E+03 1.64E+03 457.81 1.32E+03 1.51E+03 478.22 0.038 1.77E+01 34.62 1.33 1.81 0.75 3.19 4.53 1.51 1.35E+03 1674.16 329.65 8.13E+03 8102.19 231.06 1.28E+04 1.48E+04 1358.96 0 0 0 0 0 0 0 0 0 8.88E−16 4.44E−15 2.9E−15 7.99E−15 8.89E−13 7.81E−13 7.99E−15 4.67E−13 5.33E−13 0.0078 1.52E−02 0.011 2.33 2.42 0.11 4.23 4.59 0.26 0.048 8.96E−02 0.11 0 0 0 0 0 0 L. Zhang et al. / Applied Soft Computing 28 (2015) 138–149 147 These results confirm that BPSO is also competitive on multimodal functions. 50 10 0 Mean of average fitness 10 −50 4.4. Analysis of parameters s and ω in the BPSO −100 The performance of BPSO is superior due to the adaptive nature of this algorithm. Fig. 8 shows the process of the inertia weight ω changing, in which BPSO is applied to the function f1 . We see that the changing method of ω in our paper is different from the other four inertia weight strategies and most value of ω is in [0.4, 0.9]. This is mainly because ω is mainly affected by the particle positions in s step length and adjusted automatically. In the BPSO, s is a parameter that influences the convergence rate. A smaller s means that the inertia weight ω has more opportunities to be changed but less past particle position informations are used. Larger s makes ω having less chance to be changed and more past particle position informations are referred. Thus, it is important to identify the appropriate value for s in the BPSO. We endeavor to get a better parameter in BPSO algorithm. In our paper, an efficient method is used that s varies from 2 to 50 in linearincreasing method. Experiment setup is same as above. We carry out experiments to get a better approximation to best s for the unimodal functions f2 and the multimodal functions f10 on dimensions 10 (see Figs. 9 and 10). 10 10 −150 10 BPSO 0 10 PSO GPSO CPSO HPSO IPSO+ACJ HMRPSO −200 10 −250 10 195 200 205 −300 10 0 100 200 300 400 500 Iteration Fig. 7. Comparisons between different PSO methods for f5 on 10 dimensions. of local minima as the dimension of the problem increases, it is difficult for PSO algorithms to find the optimal solution steadily. On the prospect of search speed, BPSO remains one of the fastest algorithms in terms of FEs and execution time. 0.9 0.8 0.7 Inertia weight 0.6 0.5 0.4 0.3 0.2 0.1 0 −0.1 0 10 20 30 40 50 The number of w change Fig. 8. Inertia weight adaptation in BPSO applied to f1 when s = 50. 0 0 10 s=2 s=4 s=6 s=8 s=10 −5 10 Mean of average fitness Mean of average fitness 10 −10 10 −15 10 −20 10 (a) 0 s=50 s=40 s=30 s=20 s=10 −5 10 −10 10 −15 10 −20 5 10 15 20 Iteration 25 30 35 40 10 0 5 10 (b) Fig. 9. Comparison different s in the BPSO for f2 on 10 dimension. 15 20 Iteration 25 30 35 148 L. Zhang et al. / Applied Soft Computing 28 (2015) 138–149 5 5 10 10 s=2 s=4 s=6 s=8 s=10 Mean of average fitness 10 10 −5 10 −10 10 −5 10 −10 10 −15 −15 10 10 −20 10 s=50 s=40 s=30 s=20 s=10 0 Mean of average fitness 0 0 (a) −20 10 20 30 40 50 Iteration 10 (b) 0 10 20 30 40 50 Iteration Fig. 10. Comparison different s in the BPSO for f10 on 10 dimension. In Fig. 9 and 10, the convergence rate with different values of s are shown for the test function. When s = 2, 4, 6, 8, 10, results on each test function become better with s increasing. However, when s = 10, 20, 30, 40, 50, results on each test function become worse with s increasing. From the figure, we can see that the value of s influences the convergence velocity, but not very significant. 5. Conclusion In this paper, we have proposed a new inertia weight adjusting strategy based on the Bayesian techniques. In this strategy, the Bayesian techniques are used to adjust the inertia weight on the basis of the past particle positions. Our objective is to use the Bayesian techniques and the cauchy mutation to seek a tradoff between the exploration of new positions and the exploitation of the past particle’s position. In the simulation, BPSO method has faster convergence rate than other methods in the five unimodal test functions and higher accuracy than other methods in the other five multimodal test functions. The results demonstrate that the BPSO method is useful to solve optimization problems. The application of BPSO method in steel industry is our further research in our future work. Acknowledgments This paper is partially supported by the Science Fund for Hundred Excellent Innovation Talents Support Program of Hebei Province, Doctoral Fund of Ministry of Education of China (20121333110008), Hebei Province Applied Basis Research Project (13961806D), Hebei Province Development of Social Science Research Project (201401315) and the National Natural Science Foundation of China (61273260, 61290322, 61273222, 61322303). References [1] J. Kennedy, R. Eberhart, Particle Swarm Optimization, in: IEEE International Conference on Neural Networks, Perth, Australia, 1995, pp. 1942–1948. [2] K.E. Parsopoulos, V.P. Plagianakos, G.D. Magoulas, M.N. Vrahatis, Objective function “stretching” to alleviate convergence to local minima, Nonlinear Anal. TMA 47 (5) (2001) 3419–3424. [3] R. Eberhart, Y. Shi, Comparison between genetic algorithms and particle swarm optimization, in: The 7th Annual Conference on Evolutionary Programming, San Diego, USA, 1998, pp. 611–616. [4] Y.H. Shi, R.C. Eberhart, Fuzzy adaptive particle swarm optimization, in: Proc. of the IEEE Congress on Evolutionary Computation, Seoul Korea, vol. 1, 2001, pp. 101–106. [5] Y.H. Shi, R.C. Eberhart, A modified particle swarm optimizer, in: Proc. of the IEEE Congress on Evolutionary Computation, USA, 1998, pp. 69–73. [6] Y.H. Shi, R.C. Eberhart, Parameter selection in particle swarm optimization, in: EP ’98: Proceedings of the 7th International Conference on Evolutionary Programming VII, Springer-Verlag, London, UK, 1998, pp. 591–600. [7] R.C. Eberhart, Y.H. Shi, Tracking and optimizing dynamic systems with particle swarms, in: Proc. of the IEEE Congress on Evolutionary Computation, Korea, 2001, pp. 94–100. [8] R.C. Eberhart, Y.H. Shi, Comparing inertia weights and constriction factors in particle swarm optimization, in: Proc. of the IEEE Congress on Evolutionary Computation, USA, vol. 1, 2000, pp. 84–88. [9] Y.H. Shi, R.C. Eberhart, Empirical study of particle swarm optimization, in: Proc. of the IEEE Congress on Evolutionary Computation, USA, vol. 3, 1999, pp. 1945–1950. [10] K. Lei, Y. Qiu, Y. He, A new adaptive well-chosen inertia weight strategy to automatically harmonize global and local search ability in particle swarm optimization, in: ISSCAA, 2006. [11] S. Fan, Y. Chiu, A decreasing inertia weight particle swarm optimizer, Eng. Optim. 39 (2) (2007) 203–228. [12] Y. Zheng, L. Ma, L. Zhang, J. Qian, Empirical study of particle swarm optimizer with an increasing inertia weight, in: Proc. of the IEEE Congress on Evolutionary Computation, vol. 1, 2003, pp. 221–226. [13] Y. Zheng, L. Ma, L. Zhang, J. Qian, On the convergence analysis and parameter selection in particle swarm optimization, in: Proceedings of the Second International Conference on Machine Learning and Cybernetics, vol. 3, 2003, pp. 1802–1807. [14] M.S. Arumugam, M.V.C. Rao, On the improved performances of the particle swarm optimization algorithms with adaptive parameters, crossover operators and root mean square (RMS) variants for computing optimal control of a class of hybrid systems, Appl. Soft Comput. 8 (1) (2008) 324–336. [15] G. Xu, An adaptive parameter tuning of particle swarm optimization algorithm, Appl. Math. Computat. 219 (9) (2013) 4560–4569. [16] M.K. Lim, C.H. Chan, D. Monekosso, P. Remagnino, Refined particle swarm intelligence method for abrupt motion tracking, Inform. Sci. 283 (2014) 267–287. [17] A. Nickabadi, M.M. Ebadzadeh, R. Safabakhsh, A novel particle swarm optimization algorithm with adaptive inertia weight, Appl. Soft Comput. 11 (2011) 3658–3670. [18] N. Higashi, H. Iba, Particle swarm optimization with gaussian mutation, in: Proceedings of the IEEE Swarm Intelligence Symphosium 2003, IEEE Press, 2003, pp. 72–79. [19] P.S. Andrews, An investigation into mutation operators for particle swarm optimization, in: Proc of the IEEE Congress on Evolutionary Computation, 2006, pp. 1044–1051. [20] A. Stacey, M. Jancic, I. Grundy, Particle swarm optimization with mutation, in: Proc. of the IEEE Congress on Evolutionary Computation, 2003, pp. 1425–1430. [21] S.H. Ling, H.H.C. Iu, K.Y. Chan, H.K. Lam, B.C.W. Yeung, F.H. Leung, Hybrid particle swarm optimization with wavelet mutation and its industrial applications, IEEE Trans. Syst. Man Cybern. B Cybern. 38 (3) (2008) 743–763. [22] A. Ratnaweera, S.K. Halgamuge, H.C. Watson, Self-organizing hierarchical particle swarm optimizer with time-varying acceleration coefficients, IEEE Trans. Evol. Comput. 8 (3) (2004) 240–255. [23] H. Wang, Y. Liu, A hybrid particle swarm algorithm with Cauchy mutation, in: Proc IEEE Conf. Swarm Intell. Symp., 2007, pp. 356–360. [24] H. Wang, Y. Liu, Z. Wu, H. Sun, S.Y. Zeng, L.S. Kang, An improved particle swarm optimization with adaptive jumps, in: Proc. of the IEEE Congress on Evolutionary Computation, 2008, pp. 392–397. [25] M.S. Kiran, M. Gunduz, O.K. Baykan, A novel hybrid algorithm based on particle swarm and ant colony optimization for finding the global minimum, Appl. Math. Computat. 219 (4) (2012) 1515–1521. L. Zhang et al. / Applied Soft Computing 28 (2015) 138–149 [26] H. Gao, W.B. Xu, A new particle swarm algorithm and its globally convergent modifications, IEEE Trans. Syst. Man Cybern. B, Cybern. 41 (5) (2011) 1334–1351. [27] J.C. Vzquez, F. Valdez, Fuzzy logic for dynamic adaptation in PSO with multiple topologies, IFSA/NAFIPS (2013) 1197–1202. [28] P. Melin, F. Olivas, O. Castillo, F. Valdez, Optimal design of fuzzy classification systems using PSO with dynamic parameter adaptation through fuzzy logic, Expert Syst. Appl. 40 (8) (2013) 3196–3206. [29] F. Valdez, P. Melin, O. Castillo, Modular Neural Networks architecture optimization with a new nature inspired method using a fuzzy combination of particle swarm optimization and genetic algorithms, Inform. Sci. 270 (2014) 143–153. 149 [30] F. Valdez, P. Melin, O. Castillo, A survey on nature-inspired optimization algorithms with fuzzy logic for dynamic parameter adaptation, Expert Syst. Appl. 1 (14) (2014) 6459–6466. [31] C.K. Monsun, K.D. Seppi, Bayesian optimization models for particle swarms, in: Proceedings of the 2005 Conference on Genetic and Evolutionary Computation, 2005. [32] D. Martens, B. Baesens, T. Fawcett, Editorial survey: swarm intelligence for data mining, Mach. Learn. 82 (1) (2011) 1–42. [33] M. Bishop, Neural Networks for Pattern Recognition, Oxford University Press, 1995. [34] H. Jaeger, H. Haas, Harnessing nonlinearity: predicting chaotic systems and saving energy in wireless communication, Science 304 (5667) (2004) 78–80.
© Copyright 2026 Paperzz