A new neural network based optimization algorithm is proposed. The presented model is a discrete-time, continuous-state Hopfield neural network and the states of the model are updated synchronously. The proposed algorithm combines the advantages of traditional PSO, chaos and Hopfield neural networks: particles learn from their own experience and the experiences of surrounding particles, their search behavior is ergodic, and convergence of the swarm is guaranteed. The effectiveness of the proposed approach is demonstrated using simulations and typical optimization problems.
1. Introduction
The discovery of chaos in astronomical, solar, fluid, and other systems sparked significant research in nonlinear dynamics exhibiting chaos. Chaos was found to be useful and have great potential in many disciplines such as mixing liquids with low power consumption, presenting outages in power systems, biomedical engineering applications involving signals from the brain and heart, to name just a few [1]. Chaotic systems exhibit three important properties. Firstly, a deterministic system is said to be chaotic whenever its evolution sensitively depends on the initial conditions. Secondly, there is an infinite number of unstable periodic orbits embedded in the underlying chaotic set. Thirdly, the dynamics of the chaotic attractor is ergodic, which implies that during its temporal evolution the system ergodically visits small neighborhoods around every point in each one of the unstable periodic orbits embedded within the chaotic attractor. Although it appears to be stochastic, it is generated by a deterministic nonlinear system. Lyapunov exponents characterize quantitatively stochastic properties of the dynamical systems. When the dynamical system is chaotic, there exists at least one lyapunov exponent λ>0. It is reported that chaotic behavior also exists in biological neurons and neural networks [2, 3]. Using chaos to develop novel optimization techniques gained much attention during the last decade. For a given energy or cost function, the chaotic ergodic orbits of a chaotic dynamic system used for optimization may eventually reach the global optimum or a point close to it with high probability [4, 5].
Since Hopfield and Tank [6] applied their neural network to the travelling salesman problem, neural networks have provided a powerful approach to a wide variety of optimization problems [7, 8]. However the Hopfield neural network (HNN) often gets trapped in a local minima. A number of modifications were made to Hopfield neural networks to escape from local minima. Some modifications, based on chaotic neural networks [9] and simulated annealing [10], were proposed to solve global optimization problems [11]. In [12–14] the guaranteed convergence of Hopfield neural networks is discussed.
Particle swarm optimization (PSO), developed by Clerc and Kennedy in 2002 [15], is a stochastic global optimization method which is based on simulation of social behavior. In a particle swarm optimizer, individuals “evolve" by cooperating with other individuals over several generations. Each particle adjusts its flying according to its own flying experiences and the flying experience of its companions. Each individual is named as a particle which, in fact, represents a potential solution to a problem. Each particle is treated as a point in a D-dimensional space. However, the PSO algorithm is likely to temporarily get stuck and may need a long period of time to escape from a local extremum [16]. It is difficult to guarantee the convergence of the swarm, especially when random parameters are used. In order to improve the dynamical behavior of PSO, one can combine chaos with PSO algorithms to enhance the performance of PSO. In [17–19] chaos were applied to the PSO to avoid the PSO getting trapped in local minima.
PSO is motivated by the behavior of organisms such as fish schooling and bird flocking [20]. During the process, future particle positions (determined by velocity) can be regarded as particle intelligence [21]. Using a chaotic intelligent swarm system to replace the original PSO might be convenient for analysis while maintaining stochastic search properties. Most importantly, the convergence of a particle swarm initialized with random weights is not guaranteed.
In this paper we propose a chaotic Hopfield neural network swarm optimization (CHNNSO) algorithm. The rest of the paper is organized as follows. In Section 2, the preliminaries of Hopfield neural networks and PSO are described. The chaotic Hopfield neural network model is developed in Section 3. In Section 4, the dynamics of the chaotic Hopfield neural network is analyzed. Section 5 provides simulation results and comparisons. The conclusion is given in Section 6.
2. Preliminaries2.1. Basic Hopfield Neural Network Theory [<xref ref-type="bibr" rid="B22">22</xref>]
A Hopfield net is a recurrent neural network having a synaptic connection pattern such that there is an underlying Lyapunov energy function for the activity dynamics. Started in any initial state, the state of the system evolves to a final state that is a (local) minimum of the Lyapunov energy function. The Lyapunov energy function decreases in a monotone fashion under the dynamics and is bounded below. Because of the existence of an elementary Lyapunov energy function for the dynamics, the only possible asymptotic result is a state on an attractor.
There are two popular forms of the model: binary neurons with discrete time which is updated one at a time and continuous time graded neurons. In this paper, the second kind of model is used. The dynamics of a n-neuron continuous Hopfield neural network is described by
(1)duidt=-uiτ+∑jTijxi(t)+Ii.
Here, ui∈(-∞,∞) is the input of neuron i, and the output of neuron i is
(2)xi(t)=gi[ui(t)],
where i=1,2,…n, τ is a positive constant, Ii is external inputs (e.g., sensory input or bias current) to neuron i and is sometimes called the “firing threshold" when replaced with -Ii. ui is the mean internal potential of the neuron which determines the output of neuron i. Tij is the strength of synaptic input from neuron i to neuron j. g is a monotone function that converts internal potential into firing rate input of the neuron. T is the matrix with elements Tij. When T is symmetric, the Lyapunov energy function is given by
(3)J=-12∑ijTijxixj-∑iIixi+1τ∑i∫0xjg-1(Z)dZ,
where g-1 is the inverse of the gain function g. There is a significant limiting case of this function when T has no diagonal elements and the input-output relation becomes a step, going from 0 to a maximum firing rate (for convenience, scaled to 1). The third term of this Lyapnouv function is then zero or infinite. With no diagonal elements in T, the minima of J are all located at corners of the hypercube 0≤xj≤1. In this limit, the states of the continuous variable system are stable.
Many optimization problems can be readily represented using Hopfield nets by transforming the problem into variables such that the desired optimization corresponds to the minimization of the respective Lyapunov energy function [6]. The dynamics of the HNN converges to a local Lyapunov energy minimum. If this local minimum is also the global minimum, the solution of the desired optimization task has been carried out by the convergence of the network state.
2.2. Basic PSO Theory
Many real optimization problems can be formulated as the following functional optimization problem:
(4)minf(Xi),Xi=[xi1,…,xin],s.t.xi∈[ai,bi],i=1,2,…,n.
Here f is the objective function, and Xi is the decision vector consisting of n variables.
The original particle swarm algorithm works by iteratively searching in a region and is concerned with the best previous success of each particle, the best previous success of the particle swarm and the current position and velocity of each particle [20]. Every candidate solution of f(Xi) is called a “particle.” The particle searches the domain of the problem according to
(5)Vi(t+1)=ωVi(t)+c1R1(Pi-Xi(t))+c2R2(Pg-Xi(t)),(6)Xi(t+1)=Xi(t)+Vi(t+1),
where Vi=[vi1,vi2,…,vin] is the velocity of particle i; Xi=[xi1,xi2,…,xin] represents the position of particle i; Pi represents the best previous position of particle i (indicating the best discoveries or previous experience of particle i); Pg represents the best previous position among all particles (indicating the best discovery or previous experience of the social swarm); ω is the inertia weight that controls the impact of the previous velocity of the particle on its current velocity and is sometimes adaptive [17]; R1 and R2 are two random weights whose components r1j and r2j (j=1,2,…,n) are chosen uniformly within the interval [0,1] which might not guarantee the convergence of the particle trajectory; c1 and c2 are the positive constant parameters. Generally the value of each component in Vi should be clamped to the range [-vmax,vmax] to control excessive roaming of particles outside the search space.
3. A Chaotic Hopfield Neural Network Model
From the introduction of basic PSO theory, every particle can be seen as the model of a single fish or a single bird. The position chosen by the particle can be regarded as a state of a neural network with a random synaptic connection. According to (5)-(6), the position components of particle i can be thought of as the output of a neural network as shown in Figure 1.
Particle structure.
In Figure 1, Rand1(·) and Rand2(·) are two independent and uniformly distributed random variables within the range [0,1], which refer to r1j and r2j, respectively. pij and pgj are the components of Pi and Pg, respectively. pgj is the previous best value amongst all particles, and pij, as an externally applied input, is the jth element of the best previous position Pi, and it is coupled with other components of Pi. The particles migrate toward a new position according to (5)-(6). This process is repeated until a defined stopping criterion is met (e.g., maximum number of iterations or a sufficiently good fitness value).
As pointed out by Clerc and Kennedy [15], the powerful optimization ability of the PSO comes from the interaction amongst the particles. The analysis of complex interaction amongst the particles in the swarm is beyond the scope of this paper which focuses on the construction of a simple particle using a neural network perspective and convergence issues. Artificial neural networks are composed of simple artificial neurons mimicking biological neurons. The HNN has the property that as each neuron in a HNN updates, an energy function is monotonically reduced until the network stabilizes [23]. One can therefore map an optimization problem to a HNN such that the cost function of the problem corresponds to the energy function of the HNN and the result of the HNN thus suggests a low cost solution to the optimization problem. The HNN might therefore be a good choice to model particle behavior.
In order to approach Pg and Pi, the HNN model should include at least two neurons. For simplicity, the HNN model of each particle position component has two neurons whose outputs are xij(t) and xipj(t). In order to transform the problem into variables such that the desired optimization corresponds to the minimization of the energy function, the objective function should be determined firstly. As xij(t) and xipj(t) should approach pgj and pij, respectively, (xij(t)-pgj)2 and (xipj(t)-pij)2 can be chosen as two parts of the energy function. The third part of energy function (xij(t)-xipj(t))2 is added to accompany (xipj(t)-pij)2 to cause xij(t) to tend towards pij. Therefore the HNN Lyapunov energy function for each particle is proposed:
(7)Jij(t)=A(xij(t)-pgj)2+B(xipj(t)-pij)2=+C(xij(t)-xipj(t))2,
where A, B, and C are positive constants.
Here the neuron input-output function is chosen as a sigmoid function, given by (9) and (11). Equations (8) and (10) are the Euler approximation of (1) of the continuous Hopfield neural network [14]. The dynamics of component j of particle i is described by
(8)uij(t+1)=kuij(t)+α∂Jij(t)∂xij(t),(9)xij(t+1)=gi(uij(t+1))=11-eξuij(t+1),(10)uipj(t+1)=kuipj(t)+α∂Jij(t)∂xipj(t),(11)xipj(t+1)=gi(uipj(t+1))=11-eξuipj(t+1).
According to (5)-(6) and Figure 1, the PSO uses random weights to simulate birds flocking or fish searching for food. When birds flock or fish search for food, they exhibit chaos like behavior, yet (8)–(11) do not generate chaos. Aihara et al. [9] proposed a kind of chaotic neuron, which includes relative refractoriness in the model to simulate chaos in a biological brain. To use this result zi(k)(xij(k)-I0) and zi(k)(xipj(k)-I0) are added to (8) and (10) to cause chaos. Equations (8) and (10) then become
(12)uij(t+1)=kuij(t)+α∂Jij(t)∂xij(t)-zi(t)(xij(t)-I0),(13)uipj(t+1)=kuipj(t)+α∂Jij(t)∂xipj(t)-zi(t)(xipj(t)-I0).
In order to escape from chaos as time evolves, we set
(14)zi(t+1)=(1-β)zi(t).
In (8)–(14): ξ, α, and k are positive parameters; zi(t) is self-feedback connection weight (the refractory strength); β is the damping factor of the time-dependent zi(t), (0≤β≤1); I0 is a positive parameter.
All the parameters are fixed except z(t) which is varied.
The combination of (9), (11)–(14) is called chaotic Hopfield neural network swarm optimization(CHNNSO) proposed by us. According to (8)–(14), the following procedure can be used for implementing the proposed CHNNSO algorithm.
Initialize the swarm, assign a random position in the problem hyperspace to each particle, and calculate the fitness function which is given by the optimization problem whose variables are corresponding to the elements of particle position coordinates.
Synchronously update the positions of all the particles using (9), (11)–(14) and change the two states every iteration.
Evaluate the fitness function for each particle.
For each individual particle, compare the particle’s fitness value with its previous best fitness value. If the current value is better than the previous best value, then set this value as the pij and the current particle’s position, xij, as pij, else if the pi is updated, then reset z=z0.
Identify the particle that has the best fitness value. When iterations are less than a certain value, and if the particle with the best fitness value is changed then reset z=z0 to keep the particles chaotic to prevent premature convergence.
Repeat steps (2)–(5) until a stopping criterion is met (e.g., maximum number of iterations or a sufficiently good fitness value).
As can be seen from (9) and (11), the particle position component xij is located in the interval [-1,1]. The optimization problem variable interval must therefore be mapped to [-1,1] and vice versa using
(15)xj=-1+2(xj-aj)bj-aj,j=1,2,…,n,xj=aj+12(xj+1)(bj-aj),j=1,2,…,n.
Here, aj and bj are the lower boundary and the upper boundary of xj(t), respectively, and only one particle is analyzed for simplicity.
4. Dynamics of Chaotic Hopfield Network Swarm Optimization
In this section, the dynamics of the chaotic Hopfield network swarm optimization (CHNNSO) is analyzed. The first subsection discusses the convergence of the chaotic particle swarm. The second subsection discusses the dynamics of the simplest CHNNSO with different parameter values.
4.1. Convergence of the Particle SwarmTheorem 1 (Wang and Smith [<xref ref-type="bibr" rid="B14">14</xref>]).
If one has a network of neurons with arbitrarily increasing I/O functions, there exists a sufficient stability condition for a synchronous TCNN (transiently chaotic neural network) equation (12), namely,
(16)k>1,2kβmax>-Tmin′.
Here 1/βmax=min(1/gi′) (gi′ denotes the derivative with respect to time t of the neural I/O function for neuron i, in this paper is the sigmoid function (9). Tmin′ is the minimum eigenvalue of the connected weight matrix of the dynamics of a n-neuron continuous Hopfield neural network).
Theorem 2.
A sufficient stability condition for the CHNNSO model is k>1.
Proof.
When t→∞,
(17)z(t+1)=0.
It then follows that the equilibria of (12) and (13) can be evaluated by
(18)Δuij(t+1)=(k-1)uij(t)+α∂Jij(t)∂xij(t),Δuipj(t+1)=(k-1)uipj(t)+α∂Jij(t)∂xipj(t).
According to (7), we get
(19)Δuij(t+1)=(k-1)uij(t)+2Aα(xij(t)-pgj)=+2Cα(xij(t)-xipj(t))=0,(20)Δuipj(t+1)=(k-1)uipj(t)+2Bα(xipj(t)-pij)=-2Cα(xij(t)-xipj(t))=0,(21)Tmin′=min(eigenvalue(2Aα+2Cα-2Cα-2Cα2Bα+2Cα))=(A+B+2C)α-α(A-B)2+4C2=α(A+B+2C-(A-B)2+4C2).
In this paper,
(22)βmax=1min(1/g′)=ξ4.
It is clear that Tmin′>0 and the stability condition (16) is satisfied when ξ>0. The above analysis verifies Theorem 2.
Theorem 3.
The particles converge to the sphere with center point pgj and radius R=max∥xiej-pij∥ (xiej is the final convergence equilibria, if the optimization problem is in two-dimensional plane, the particles are finally in a circle).
It is easy to show that the particle model given by (7) and (8)–(14) has only one equilibrium as t→∞, that is, z(t)=0. Hence, as t→∞, Xi belongs to the hypersphere whose origin is Pg and the radius is R. Solving (9), (11), (19), and (20) simultaneously, we get
(23)(k-1)ln(1-(1/xij)(t))ξ+2Aα(xij(t)-Pgj)+2Cα(xij(t)-xipj(t))=0,(24)(k-1)ln(1-(1/xipj(t)))ξ+2Bα(xipj(t)-Pij)-2Cα(xij(t)-xipj(t))=0.
With (23) and (24) satisfied, there must exist the final convergence equilibria xiej and xipej. So the best place the particle swarm can find is ∥pgj-xiej(t+1)∥ and radius is R=max∥xiej-pij∥.
The above analysis therefore verifies Theorem 3.
4.2. Dynamics of the Simplest Chaotic Hopfield Neural Network Swarm
In this section, the dynamics of the simplest particle swarm model is analyzed. Equations (7) and (8)–(13) are the dynamic model of a single particle with subscript i ignored. According to (7) and Theorem 3, the parameters A, B, and C control the final convergent radius. According to trial and error, the parameters A, B, and C can be chosen in the range from 0.005 to 0.05. According to (16) and (22), k>1 and ξ>0. In the simulation, the results are better when k is in the neighborhood of 1.1 and ξ is in the neighborhood of 150. The parameters β and Z(0) control the time of the chaotic period. If β is too big and/or Z(0) is too small, the system will quickly escape from chaos and performance will be poor. The parameter I0=0.2 is standard in the literature on chaotic neural networks. The simulation showed that the model is not sensitive to the values of parameters ξ and k, for example, 100<ξ<200 and 1<k<1.5 are feasible.
Then the values of the parameters in (7)–(14) are set to
(25)A=0.02,B=C=0.01,ξ=150,α=1,β=0.001,Z(0)=0.3,k=1.1,I0=0.2,pi=0.5,pg=0.2.
Figure 2 shows the time evolution of xi(t), z(t) and the Lyapunov exponent λ of xi(t). The Lyapunov exponent λ characterizes the rate of separation of infinitesimally close trajectories. A positive Lyapunov exponent is usually taken as an indication that the system is chaotic [1]. Here, λ is defined as
(26)λ=limn→+∞1m∑t=0m-1ln|dxi(t+1)dxi(t)|.
At about 200 steps, z(t) decays to a small value and xi(t) departs from chaos which corresponds with the change of λ from positive to negative.
Time evolutions of xi(t),z(t), and the Lyapunov exponent λ.
According to Figure 2, the convergence process of a simple particle position follows the nonlinear bifurcation making the particle converge to a stable fixed point from a strange attractor. In the following section, it is shown that the fixed point is determined by the best previous position Pg among all particles and the best position Pi of the individual particle.
Remark 4.
The proposed CHNNSO model is a deterministic Chaos-Hopfield neural network swarm which is different from existing PSOs with stochastic parameters. Its search orbits exhibit an evolutionary process of inverse period bifurcation from chaos to periodic orbits then to sink. As chaos is ergodic and the particle is always in a chaotic state at the beginning (e.g., in Figure 2), the particle can escape when trapped in local extrema. This proposed CHNNSO model will therefore in general not suffer from being easily trapped in a the local optimum and will continue to search for a global optimum.
5. Numerical Simulation
To test the performance of the proposed algorithms, two famous benchmark optimization problems and an engineering optimization problem with linear and nonlinear constraints are used. The solutions to the two benchmark problems can be represented in the plane and therefore the convergence of the CHNNSO can be clearly observed. The results of the third optimization problem when compared with other algorithms are displayed in Table 1. We will compare the CHNNSO with the original PSO [20].
The parameters of the Hartmann function (when n=3 and q=4).
i
ai1
ai2
ai3
ci
pi1
pi2
pi3
1
3
10
30
1
0.3689
0.1170
0.2673
2
0.1
10
35
1.2
0.4699
0.4387
0.7470
3
3
10
30
3
0.1091
0.8732
0.5547
4
0.1
10
35
3.2
0.03815
0.5473
0.8828
5.1. The Rastrigin Function
To demonstrate the efficiency of the proposed technique, the famous Rastrigin function is chosen as a test problem. This function with two variables is given by
(27)f(X)=x12+x22-cos(18x1)-cos(18x2),-1<xi<1,i=1,2.
The global minimum is −2 and the minimum point is (0, 0). There are about 50 local minima arranged in a lattice configuration.
The proposed technique is applied with a population size of 20 and the maximum number of iterations is 20000. The chaotic particle swarm parameters are chosen as A=0.02, B=C=0.01, β=0.001, z(0)=0.3, α=1, ξ=150, k=1.1, and I0=0.2.
The position of every particle is initialized with a random value. The time evolution of the cost of the Rastrigin function is shown in Figure 3. The global minimum at −2 is obtained by the best particle with (x1,x2)=(0,0).
Time evolutions of Rastrigin function.
From Figure 3, it can be seen that the proposed method gives good optimization results. Since there are two variables in the Rastrigin function, it is easy to show the final convergent particle states in the plane.
In Figure 4, the “*”s are the best experiences of each particle. The “+”s are the final states of the particles. The global minimum (0,0) is also included in the “*”s and the “+”s. Most “*”s and “+”s are overlapped at the global minimum (0,0). According to Theorem 3, the particles will finally converge to a circle finally. For this Rastrigin problem, the particles’ final states converge to the circle as shown in Figure 4, and hence the global convergence of the particles is guaranteed.
The final states and the best experiences of the particles achieved from the proposed CHNNSO for Rastrigin function.
Figure 5 displays the results when the original PSO [20] was used to optimize the Rastrigin function. In the numerical simulation, the particle swarm population size is also 20 and parameters c1 and c2 are set to 2 and ω set to 1. vmax is set equal to the dynamic range of each dimension. The “*”s in Figure 5 are the final states of all the particles corresponding to the “+”s in Figure 4. It is easy to see that the final states of the particles are ruleless even though the global minimum of −2 is obtained by the best experience of the particle swarm, that is, (x1,x2)=(0,0) as shown in Figure 5.
The final states of the particles achieved from original PSO for Rastrigin function.
By comparing the results obtained by the proposed CHNNSO in Figure 4 with the results of the original PSO in Figure 5, it can be seen that the final states of the particles of the proposed CHNNSO are attracted to the best experience of all the particles and that convergence is superior. The final states of CHNNSO particles are guaranteed to converge which is not the case for original PSO implications.
When their parameters c1 and c2 are both set to 1.494 and ω to a value of 0.729, vmax is set equal to the dynamic range on each dimension. The constriction factors c1, c2, ω are applied to improve the convergence of the particle over time by damping the oscillations once the particle is focused on the best point in an optimal region. The main disadvantage of this method is that the particles may follow wider cycles and may not converge when the individual best performance is far from the neighborhoods best performance (two different regions) [24].
5.2. The Schaffer’s F6 Function
To further investigate the performance of the CHNNSO,
(28)f(X)=0.5+sin(x12+x22)2-0.51.0+0.001(x12+x22)2,
the Schaffer’s F6 function [25] is chosen. This function has a single global optimum at (0,0) and fmin(x,y)=0, and a large number of local optima. The global optimum is difficult to find because the value at the best local optimum differs with only about 10-3 from the global minimum. The local optima crowd around the global optimum. The proposed technique is applied with a population size of 30, the iterations are 100000, and the parameters of CHNNSO are chosen as
(29)A=B=C=0.01,β=0.0003,z(0)=0.3,ξ=155,k=1.1,I0=0.2.
The position of each particle is initialized with a random value. In Figure 6, the “*”s are the best experiences of each particle. The global minimum at 2.674×10-4 is obtained by the best particle with (x1,x2)=(-0.00999,0.01294) which is included in the “*”s. The “+”s are the final states of the particles. According to Theorem 3 the particles’ final states converge in a circle as shown in Figure 6 which proves global convergence. From Figure 6 it is clearly seen that the particles’ final states are attracted to the neighborhoods of the best experiences of all the particles and the convergence is good.
The final states and the best experiences of the particles achieved from the proposed CHNNSO for Schaffer’ F6 function.
Figure 7 shows the final particle states when the original PSO [20] was used to optimize the Schaffer’s F6 function. In this numerical simulation of the original PSO, the particle swarm population size is also 20 and parameters c1 and c2 are both set to 2 and set ω is set to a value of 1. vmax is set equal to the dynamic range of each dimension. The “*”s in Figure 7 are the final states of all the particles corresponding to the “+”s in Figure 6. It is easy to see that the final states of the particles are ruleless in Figure 7. The global minimum 6.3299×10-4 is obtained by the best particle with (x1,x2)=(0.01889×10-3, -0.4665×10-3). The best experience from the original PSO is not as good as the best of the proposed CHNNSO.
The final states of the particles achieved from original PSO for Schaffer’s F6 function.
Comparing the results obtained from the proposed CHNNSO in Figure 6 and the original PSO in Figure 7, it is clearly seen that the particles’ final states of the proposed CHNNSO are finally attracted to the best experience of all the particles and the convergence is better than that of the original PSO. The CHNNSO can guarantee the convergence of the particle swarm, but the final states of the original PSO are ruleless.
5.3. The Hartmann Function
The Hartmann function when n=3; q=4 is given by
(30)f(x)=-∑i=1qciexp(-∑j=1naij(xj-pij)2),
with xj belonging to
(31)S={x∈Rn∣0≤xj≤1,1≤j≤n},
Table 1 shows the parameter values for the Hartmann function when n=3,q=4.
When n=3, Xmin=(0.114,0.556,0.882), f(xmin)=-3.86.
The time evolution of the cost of the Hartmann function is 15000. In Figure 8, only subdimensions are pictured. In Figure 8, the “+”s are the final states of the particles and the “*”s denote the best experiences of all particles. From Figure 8, it can be easily seen that the final states of the particles converge to the circle. The center point is (0.0831,0.5567,0.8522) and the radius is 0.1942. The final particle states confirm Theorem 3, and the final convergency is guaranteed.
The final states of the particles achieved from original PSO for Hartmann function.
5.4. Design of a Pressure Vessel
The pressure vessel problem described in [26, 27] is an example which has linear and nonlinear constraints and has been solved by a variety of techniques. The objective of the problem is to minimize the total cost of the material needed for forming and welding a cylindrical vessel. There are four design variables: x1 (Ts, thickness of the shell), x2 (Th, thickness of the head), x3 (R, inner radius), and x4 (L, length of the cylindrical section of the vessel). Ts and Th are integer multiples of 0.0625 inch, which are the available thickness of rolled steel plates, and R and L are continuous. The problem can be specified as follows:
(32)Minimizef(X)=0.6224x1x3x4+1.7781x2x32f(X)=+3.1661x12x4+19.84x12x3,subjecttog1(X)=-x1+0.0193x3≤0,g2(X)=-x2+0.00954x3≤0,g3(X)=-πx32x4-43πx3+1296000≤0,g4(X)=x4-240≤0.
The following range of the variables were used [27]:
(33)0<x1≤99,0<x2≤99,10≤x3≤200,10≤x4≤200.
de Freitas Vas and de Graça Pinto Fernandes [26] proposed an algorithm to deal with the constrained optimization problems. Here this algorithm [26] is combined with CHNNSO to search for the global optimum. The proposed technique is applied with a population size of 20 and the maximum number of iterations is 20000. Then the values of the parameters in (7)–(14) are set to
(34)A=0.02,B=C=0.01,ξ=150,α=1,β=0.001,Z(0)=0.3,k=1.1,I0=0.2,pi=0.5,pg=0.5.
From Table 2, the best solution obtained by the CHNNSO is better than the other two solutions previously reported.
Comparison of the results for pressure Vessel design problem.
Design variables
Best solutions found
This paper
Hu et al. [27]
Coello [28]
x1(Ts)
0.8125
0.8125
0.8125
x2(Th)
0.4375
0.4375
0.4375
x3(R)
42.09845
42.09845
40.3239
x4(L)
176.6366
176.6366
200.0
g1(X)
0
0
0.03424
g2(X)
-0.03588
-0.03588
-0.052847
g3(X)
-1.1814*10-11
-5.8208*10-11
-27.105845
g4(X)
-63.3634
-63.3634
-40.0
f(X)
6059.1313
6059.1313
6288.7445
6. Conclusion
This paper proposed a chaotic neural networks swarm optimization algorithm. It incorporates the particle swarm searching structure having global optimization capability into Hopfield neural networks which guarantee convergence. In addition, by adding chaos generator terms into Hopfield neural networks, the ergodic searching capability is greatly improved in the proposed algorithm. The decay factor introduced in the chaos terms ensures that the searching evolves to convergence to global optimum after globally chaotic optimization. The experiment results of three classic benchmark functions showed that the proposed algorithm can guarantee the convergence of the particle swarm searching and can escape from local extremum. Therefore, the proposed algorithm improves the practicality of particle swarm optimization. As this is a general particle model, some techniques such as the local best version algorithm proposed in [29] can be used together with the new model. This will be explored in future work.
Acknowledgments
This work was supported by China/South Africa Research Cooperation Programme (nos. 78673 and CS06-L02), South African National Research Foundation Incentive Grant (no. 81705), SDUST Research Fund (no. 2010KYTD101), and Key scientific support program of Qingdao City (no. 11-2-3-51-nsh).
SprottJ. C.AiharaK.MatsumotoG.SkardaC. A.FreemanW. J.How brains make chaos in order to make sense of the worldWangL.ZhengD. Z.LinQ. S.Survey on chaotic optimization methodsLiB.JiangW. S.Optimizing complex functions by chaos searchHopfieldJ. J.TankD. W.‘Neural’ computation of decisons in optimization problemsWangZ.LiuY.FraserK.LiuX.Stochastic stability of uncertain Hopfield neural networks with discrete and distributed delaysTanakaT.HiuraE.Computational abilities of a chaotic neural networkAiharaK.TakabeT.ToyodaM.Chaotic neural networksKirkpatrickS.GelattC. D.Jr.VecchiM. P.Optimization by simulated annealingChenL.AiharaK.Chaotic simulated annealing by a neural network model with transient chaosChenL.AiharaK.Chaos and asymptotical stability in discrete-time neural networksWangL.On competitive learningWangL.SmithK.On chaotic simulated annealingClercM.KennedyJ.The particle swarm-explosion, stability, and convergence in a multidimensional complex spaceKeJ.QianJ. X.QiaoY. Z.A modified particle swarm optimization algorithm,LiuB.WangL.JinY. H.TangF.HuangD. X.Improved particle swarm optimization combined with chaosXiangT.LiaoX.WongK. W.An improved particle swarm optimization algorithm combined with piecewise linear chaotic mapFanC.JiangG.A simple particle swarm optimization combined with chaotic searchProceedings of the 7th World Congress on Intelligent Control and Automation (WCICA '08)June 2008Chongqing, China5935982-s2.0-5214911910210.1109/WCICA.2008.4592989KennedyJ.EberhartR.Particle swarm optimizationProceedings of the IEEE International Conference on Neural NetworksDecember 1995Perth, Australia194219482-s2.0-0029535737ChenS. H.JakemanA. J.NortonJ. P.Artificial Intelligence techniques: an introduction to their use for modelling environmental systemsHopfieldJ. J.Hopfield networkHopfieldJ. J.Neurons with graded response have collective computational properties like those of two-state neuronsdel ValleY.VenayagamoorthyG. K.MohagheghiS.HernandezJ. C.HarleyR. G.Particle swarm optimization: basic concepts, variants and applications in power systemsSchafferJ. D.CaruanaR. A.EshelmanL. J.DasR.A study of control parameters affectiong online performance of genetic algorithms for function optimizationProceedings of the 3rd International Conference on Genetic Algorithms19895160de Freitas VazA. I.da Graça Pinto FernandesE. M.Optimization of nonlinear constrained particle swarmHuX.EberhartR. C.ShiY.Engineering optimization with particle swarmProceedings of the IEEE Swarm Intelligence Symposium20035357CoelloC. A. C.Theoretical and numerical constraint-handling techniques used with evolutionary algorithms: a survey of the state of the artKennedyJ.Small worlds and mega-minds: effects of neighborhood topology on particle swarm performance