Glowworm Swarm Optimization and Its Application to Blind Signal Separation

Traditional optimization algorithms for blind signal separation (BSS) aremainly based on the gradient, which requires the objective function to be continuous and differentiable, so the applications of these algorithms are very limited. Moreover, these algorithms have problems with the convergence speed and accuracy. To overcome these drawbacks, this paper presents a modified glowworm swarm optimization (MGSO) algorithm based on a novel step adjustment rule and then applies MGSO to BSS. Taking kurtosis of the mixed signals as the objective function of BSS, MGSO-BSS succeeds in separating the mixed signals in Matlab environment. The simulation results prove that MGSO is more effective in capturing the global optimum of the objective function of the BSS algorithm and has faster convergence speed and higher accuracy, compared with particle swarm optimization (PSO) and GSO.


Introduction
"Cocktail party" problem can be seen as a classic example of blind signal separation: imagine being at a friend's party where, during your conversation with your friend, even though the sounds that reach your ears are a complicated mix of music, other people talking, wine glasses tinkling, and so on, you are able to understand your friend and enjoy the music at the same time.It is the task of blind signal separation (BSS) to recover unknown independent source signals obtained from sensors.The BSS technology has received considerable attention in recent years because of its significant potential applications such as sonar and radar signal processing [1,2], wireless communication [3], geophysical exploration [4,5], biomedical signal processing [6,7], speech and image processing [8,9], and machine fault diagnosis [10,11].BSS has two important components: the objective function and the optimization algorithm.The objective function is responsible for determining the statistical independence of separation signals, and the optimization algorithm is to ensure that the objective function value reaches its peak in subsequent updates.The convergence speed and accuracy of BSS mainly rely on the latter.Therefore, how to choose an appropriate algorithm is the crucial challenge of BSS.
Conventional optimization algorithms for BSS are based on gradient techniques, yet these methods would get a "poor" solution unless suitable initial parameters are given.However, it is very difficult to select these parameters because of the blind hypothesis.In particular, these algorithms cannot be used when the objective function is discontinuous and nondifferentiable.To solve the above problems, swarm-based algorithms have been gradually applied to BSS in the past few years, such as genetic algorithms (GA) [12,13], particle swarm optimization (PSO) [12,14], and artificial bee colony (ABC) [15].Swarm-based algorithms belong to a family of natureinspired, population-based optimizations and the behavior of their agents is inspired by biological swarms like ants, fish, bees, frogs, and bacteria, which interact in accordance with certain behavioral law to cooperatively achieve some necessary tasks.Compared with conventional gradient-based approaches, these techniques for BSS are characterized by higher accuracy, efficiency, and robustness.However, there is 2 Mathematical Problems in Engineering room for improvement of the performance of these optimization algorithms in terms of their tendency to fall into local optimum, convergence rate, and computational accuracy.
The GSO acronym can stand for two different swarmbased optimizations: Genetical Swarm Optimization [16,17] and Glowworm Swarm Optimization.Genetical Swarm Optimization is a hybrid evolutionary algorithm that combines the well-known PSO and Genetic Algorithm (GA).Glowworm Swarm Optimization [18][19][20][21][22][23] proposed by Krishnanand and Ghose imitates the behavior that a glowworm carries a luminescence quantity called luciferin along with itself to exchange information with companions.GSO in this paper is only used for Glowworm Swarm Optimization.
GSO can effectively avoid missing the optimal solution because of intelligent changes of the decision radius and is very competent in capturing the global optimum of the objective function in finite-dimensional vector space.At present, GSO has been successfully applied in various fields, such as vehicle routing problem [24], dock scheduling problem [25], and wireless sensor networks [26].Despite the above-mentioned advantages, the standard GSO has a tradeoff between convergence speed and accuracy because of the fixed step-size (the suggested step-size is 0.03).
In this paper, we present a modified GSO (MGSO) algorithm to conquer the above defects and then apply MGSO to BSS; finally, the experiment proves the effectiveness of MGSO-BSS.The remainder of this paper is organized as follows: the next section gives a complete presentation of the basic GSO and describes the proposed methods; Section 3 introduces the working principle of BSS; in Section 4, seeking mode of new BSS algorithm based on MGSO is described; in Section 5, we carry out experiments to evaluate MGSO-BSS and analyze the simulation results; the last section contains the concluding remarks on this work.

Basic Glowworm Swarm Optimization
2.1.Algorithm Representation.In GSO, a swarm of glowworms are initially deployed randomly in the solution space.Each glowworm represents a solution of objective function in the search space and carries a certain quantity of luciferin along with it.The luciferin level is associated with the fitness of the agent's current position.The brighter individual means a better position (is a better solution).Using a probabilistic mechanism, each agent can only be attracted by a neighbor whose luciferin intensity is higher than its own within the local-decision domain and then moves towards it.The density of a glowworm's neighbors affects its decision radius and determines the size of its local-decision domain: when the neighbor-density is low, the local-decision domain will enlarge in order to find more neighbors; otherwise, it will reduce to allow the swarm split into smaller groups.
The above process is repeated until the algorithm satisfies the termination condition.At this point, the majority of individuals gather around brighter glowworms.Briefly, the GSO involves five main phases: luciferin-update phase, neighborhood-select phase, moving probability-computer phase, movement phase, and the decision radius update phase.

Luciferin-Update Phase.
The luciferin update depends on the fitness value and previous luciferin value, and its rule [18][19][20][21][22][23] is given by Here,   () denotes the luciferin value of glowworm  at time ,  is the luciferin decay constant,  is the luciferin enhancement constant;   ( + 1) ∈   is the location of glowworm  at time  + 1, and Fitness(  ( + 1)) represents the value of the fitness at glowworm 's location at time  + 1.

Neighborhood-Select Phase.
Neighbors   () [18][19][20][21][22][23] of glowworm  at  time consist of the brighter ones and can be written as Here,   () represents the Euclidean distance between glowworms  and  at time , and    () represents the decision radius of glowworms  at time .

Moving Probability-Computer Phase.
A glowworm uses a probability rule to move towards other glowworms having higher luciferin level.The probability   () [18][19][20][21][22][23] of glowworm  moving towards its neighbor  can be stated as follows: Here, ‖ ⋅ ‖ represents the Euclidean norm operator, and  is the step-size.

Decision Radius Update Phase.
In each update, decision radius of glowworm  is given as follows: Here,  is a constant,   denotes the sensory radius of glowworm , and   is a parameter to control the neighbor number.Figure 1 shows the sensory radius and decision radius of glowworm .

Modified Glowworm Swarm Optimization (MGSO).
For the standard GSO, if the fixed step-size is large, each glowworm covers a large jump (equal to step-size).Therefore, these glowworms may move so fast as to miss the optimum solution in the updates.When the distance between a glowworm and its best neighbor is less than  in (4), the glowworm would oscillate.However, if the step-size decreases, the convergence rate becomes slow.Consequently, it is difficult to decide the most appropriate step-size.In this paper, the step-size is not fixed and varies for each glowworm in each iteration.Here, let the step-size of glowworm  be the function of update number , a dynamic step-size strategy is proposed to accelerate the convergence speed in the early stage of search and improve the calculation accuracy in the later stage of search.The step-size function (showed in Figure 2) is expressed as follows: Here,  and  are positive factors, and  is a step minimum-threshold.
If the initial step-size in ( 6) is equal to or slightly bigger than the fixed step-size in (4), the real-time step-size in (6) would be smaller than that in (4) at the end.Therefore, the probability and amplitude of oscillation of the agent of MGSO near the optimal solution would be much smaller than those of the standard GSO.Particularly, when the parameter values in (6) are properly chosen, the oscillation can be ignored.

Blind Signal Separation
3.1.Mathematical Model.This section gives the basic formulations [27] of BSS and describes its major steps.The general system model for BSS is shown in Figure 3.In the figure, s() = [ 1 (),  2 (), . . .,   ()]  is an -dimensional vector of unknown source signals and is instantaneously and linearly mixed by a random full-rank matrix A ∈ R × .Here, there is an assumption that each component of s() is statistically independent.
The observed signals (also called mixed signals) x() = [ 1 (),  2 (), . . .,   ()]  can be expressed as a linear transformation of the source vector as follows: Here, A ∈ R × is also unknown, and n() is the dimensional vector of additive noise and is usually ignored.
The task of BSS is to obtain the separation matrix W ∈ R × based only on the observed signals by some algorithm.The recovery signals y() = [ 1 (),  2 (), . . .,   ()]  of source signals can be written as:

Amplitude of Separation Signal.
To clearly illustrate this problem, (7) can be written in the form Here,   is the (, ) element of mixing matrix A, and  > 0 is a constant.The reason for this amplitude is that any scalar multiplier in the source   could always be cancelled by dividing the corresponding row   of A by the same scalar.Fortunately, this ambiguity of separation signal is insignificant in most applications.

Order of Separation
Signal.Formula (7) can also be written in the form x = AP −1 Ps. (10) Here, P is a permutation matrix.From (7), it can be concluded that P and its inverse P −1 can be substituted in the model.The elements of Ps are a set of the original independent variables, but in a different order.Then, the matrix AP −1 is just another unknown mixing matrix.

Measuring Algorithm
For most (but not quite all) non-Gaussian random variables, kurtosis is nonzero.

Signal Preprocessing.
In general, kurtosis is difficult to compute from measured data as the data points required for reasonable accuracy are very large.To overcome this problem, the preprocessing steps need to be performed before the mixed signals are separated.This method can simplify BSS to the estimation of rotation angle of the joint probability density function (PDF).By this way, the computation cost can be reduced by approximately half.The preprocessing includes centering and whitening (sphering) of the data [27].

3.4.1.
Centering.Subtracting its mean vector [  ] from the signal   , we can center the signal   into zero-mean variable.However, this does not mean that the mean could not be estimated.After estimating the mixing matrix A with centered data, we can complete the estimation by adding the mean vector of s back to the centered estimates of s.Centering is described as follows: 3.4.2.Whitening.Whitening is a critical procedure that reduces the number of parameters to be calculated.The observed signal x is converted to a whitened vector k with the help of a whitening matrix V to ensure [kk  ] = I (identity matrix).The process can be mathematically expressed as follows: There are many ways to whiten signals, such as principal components analysis (PCA) and singular value decomposition (SVD).The whitening matrix determined by the PCA approach is given as Here, D is the diagonal matrix of eigenvalues of the covariance matrix C = {x()x()  }, and E is the orthogonal matrix containing eigenvectors of C.After the mixed signals are whitened, instead of estimating the  2 coefficients of matrix A, we only need to estimate an orthogonal mixing matrix U = VA which only has ( − 1)/2 parameters, about half the number of parameters in matrix A.

Seeking Mode of New BSS Algorithm Based on MGSO
It can be described as follows.
Step 1. Read the observed signals x, and then center and whiten them.
Step 2. Initialize parameters of , , ,   ,  0 , and , and then generate a certain number of separation matrixes as glowworm individuals and initialize the position () and decision radius   () of these glowworms, and then calculate the initial fitness value of each glowworm in the search space.
Step 3. Calculate the optimal location  opt and the optimal fitness Fitness( opt ) of these particles.
Step 6. Calculate the optimal fitness value of all glowworms.If this value is superior to Fitness( opt ),  opt and Fitness( opt ) will be updated.
Step 7. Determine whether to satisfy the termination condition.If satisfied, jump out of the loop; otherwise, jump to Step 4.

Simulation Experiments
without delay as in formula (7) (n is ignored); the mixed signals are shown in Figure 5.The choice of parameters plays an important role in the performance of MGSO algorithm.The values of main parameters of MGSO are kept fixed in this experiment, and listed in Table 1. and approximately plots the track of glowworms seeking optimal solution.The fitness curves in Figure 8 mean that the standard PSO easily falls into a local optimum; the standard GSO has a good ability to search for the optimal value, but its convergence is not fast enough against others; MPSO has the best behavior on accuracy and speed.Figure 9 shows scatter plots of the source signal against its estimate using the MPSO algorithm.As can be seen from Figure 9, the source signals are well restored, but their orders are changed:  1 is the recovery signal of  3 ,  2 is the recovery signal of  1 , and  3 is the recovery signal of  2 .

Conclusions
To design an excellent optimization algorithm for BSS, a novel step adjustment strategy for the basic GSO is proposed in this paper.Based on independent component analysis (ICA), the modified GSO (MGSO) succeeds in separating the mixed signals using computer simulation.The experiment results show that the new MGSO-BSS has stronger global search ability and faster convergence rate than PSO-BSS and GSO-BSS, along with much higher accuracy.Therefore, it is concluded that the MGSO algorithm is more suitable for BSS.

Figure 1 :
Figure 1: Sensory and decision radius of glowworm .

Figure 2 :
Figure 2: Curve of proposed step-size function.

5. 1 .
Experimental Environment and Parameter Setting.Three different digitized signals presented in Figure 4 are used in this simulation; these signals are mixed by the matrix Figure 6  is the restored signals obtained by GMSO after 15 updates; Figure7is two-dimensional map of W  (1, 1) and W (2,1)

Figure 8 :
Figure 8: Fitness curves obtained by different optimization algorithms.

Figure 9 :
Figure 9: Scatter plots of s and y.
Performance.The Central Limit Theorem, a classical result in probability theory, tells us that the distribution of a sum of several independent random

Table 1 :
The values of the parameters of BSS-MGSO in this experiment.