1 Introduction

Particle swarm optimization (PSO) is a population-based heuristic algorithm for numerical optimization. It was first introduced by Eberhart and Kennedy (1995) and Kennedy and Eberhart (1995) as an alternative to evolutionary algorithms (EAs) (Bäck et al. 1997) that were dominant at that time. PSO has many in common with EAs (Bäck et al. 1997). For instance, it uses a population of search points, new potential solutions are stochastically generated based on the information drawn from the best ones, and it requires only function values disregarding the existence of gradient information. These properties allow PSO to efficiently work on problems where only limited information on the objective function is available or accessible. Such problems are met in black-box optimization as well as in cases where computation is characterized by uncertainties, inaccurate data, noisy and time-dependent environments (Parsopoulos 2002, 2010).

Since its introduction, PSO has gained increasing popularity. This can be ascribed to its experimentally verified efficiency in challenging optimization problems as well as its easy implementation. The minor effort required to implement PSO in modern programming languages has attracted researchers from different scientific disciplines in search of a simple yet efficient optimization algorithm that can tackle complicated problems without strong mathematical prerequisites. As a result, there is a remarkable number of PSO-based applications in diverse scientific fields (Poli 2007), although several modifications in the original PSO model were necessary to enhance its performance on some occasions.

Clerc and Kennedy (2002) derived the stability analysis of PSO, which was later extended by Trelea (2003). In their analysis, several PSO models were considered and theoretically analyzed to distinguish the most promising one. These works revealed that a model with a constriction coefficient on the velocities, along with proper parameter configuration, could alleviate the deficiencies of the first PSO model. The parameter setup of this model was theoretically justified, resulting in a PSO variant that proved to be popular especially among non-expert practitioners with limited knowledge of such algorithms, their operation and manipulation.

The theoretical investigation of the constriction coefficient model was mostly based on loose assumptions. However, due to the necessity for particular mathematical manipulations, the memory of the algorithm was unavoidably considered to remain fixed during its execution. In other words, the information that is used to attract the particles towards specific (promising) regions of the search space was not updated throughout the algorithm’s run. Yet, in practice this memory is frequently updated. Hence, every change in memory can be considered as a restarting of the algorithm’s dynamics. This observation has not attracted much attention although it can be highly related to PSO’s efficiency. Indeed, it is easily verified that in a typical PSO run, even on a problem of moderate difficulty, the information stored in PSO’s memory may continually change. Thus, the question naturally comes up: how crucially is the performance of PSO affected by the continuously restarted dynamics? Newer theoretical studies such as (Blackwell 2006) shed some light on this issue. Nevertheless, it still remains a rather neglected aspect of the PSO algorithm.

The present paper aims at experimentally probing this attribute of PSO. More specifically, we propose a constriction coefficient PSO model, called Particle Swarm Optimization with Deliberate Loss of Information (PSO-DLI), equipped with a stochastic decision-making scheme that determines whether a particle shall be evaluated and update its memory at each iteration. This scheme can be considered as a hybrid combination of the standard (synchronous) and the asynchronous PSO model, which is mostly used in parallel implementations. However, the proposed approach has a unique characteristic: every particle’s motion is undisrupted even if it is not selected for evaluation and memory update. Thus, the algorithm can deploy its dynamic without continuous disruptions imposed by the dynamic’s restarting described above.

On the other hand, it is easily understood that such a model suffers from loss of information at each iteration, namely unevaluated new particle positions. Thus, the current study has the mission to experimentally investigate the impact of such a loss of information on the algorithm’s performance. For this purpose, experiments are conducted on a number of widely used benchmark problems as well as on problems drawn from real-life applications. More specifically, the standard test suite that consists of the Sphere, Rosenbrock, Rastrigin, Griewank and Ackley function, both in their original, generalized forms as well as in their rotated and shifted forms that were used for the CEC2008 competition on large-scale global optimization (Tang et al. 2007), were considered. Additionally, a set of nonlinear systems stemming from neurophysiology, chemical equilibrium, kinematic, combustion and economic modeling applications (Grosan 2008), was considered and solved with the proposed PSO-DLI approach. Finally, we considered a set of 19 benchmark problems that served as test suite for the recently published special issue on “scalability of evolutionary algorithms and other metaheuristics for large-scale continuous optimization problems” of the present journal (Lozano et al. 2011). The special issue was devoted to empirical evaluations of several algorithms on large-scale instances of the specific problems.Footnote 1

In our experiments, PSO-DLI was compared against the asynchronous PSO model on the standard test suite and, additionally, against genetic algorithms, differential evolution and the more sophisticated MONS approach (Grosan 2008) on the nonlinear systems. On the CEC’2008 test suite, it was compared also to more specialized and complicated algorithms such as EPUS-PSO and DMS-PSO (Hsieh et al. 2008; Zhao et al. 2008). The results revealed a remarkable potential of PSO-DLI to compete with all these algorithms on these problems. Finally, on the special issue’s test suite, PSO-DLI was compared to the 16 algorithms that participated in the competition, exhibiting reasonably good performance. Considering that its implementation practically requires only marginal alteration of the existing PSO models and codes, we can suggest PSO-DLI as a very promising PSO variant that can be very useful especially to non-expert practitioners that wish to use a simple PSO model such as the constriction coefficient (or the algebraically equivalent inertia weight) one.

The rest of the paper is organized as follows: Sect. 2 offers the necessary background on PSO, both in its synchronous and asynchronous version. The proposed PSO-DLI algorithm is presented in Sect. 3 and experimental results are reported and discussed in Sect. 4. Finally, the paper concludes in Sect. 5.

2 Background information

For completeness purpose, a brief presentation of the required background information is given in the following sections. This includes the standard (synchronous) PSO algorithm as well as its asynchronous counterpart. These approaches will constitute the ground for the presentation of the proposed PSO-DLI scheme in a later section.

2.1 Particle swarm optimization

Consider the n-dimensional global optimization problem:

$$ \min_{x \in X \subset {\mathbb{R}}^n} f(x). $$

PSO employs a set of potential solutions,

$$ S = \{x_1, x_2, \ldots, x_N \}, \quad x_i \in X, \quad i \in I=\{1,2,\ldots,N\}, $$

to probe the search space. This set is called a swarm, while each vector in S corresponds to a particle:

$$ x_i = (x_{i1}, x_{i2}, \ldots, x_{in})^\top \in X, \qquad i \in I. $$

Each particle is randomly initialized in X and allowed to iteratively move within it. Its motion is determined by an adaptable velocity, while it retains in memory the best position it has ever visited. These quantities are denoted as:

$$ v_i = (v_{i1}, v_{i2}, \ldots, v_{in})^\top, \quad p_i = (p_{i1}, p_{i2}, \ldots, p_{in})^\top, \quad i \in I, $$

respectively. Thus, if t denotes the current iteration of the algorithm, it holds that:

$$ p_i(t) = \arg \min_{q \in \{0,1,\ldots,t\}} \{f(x_i(q))\}, \quad i \in I. $$

Best positions constitute the memory of the particles and it is used to guide them towards the promising regions of the search space, i.e., regions that posses lower function values.

In order to avoid the premature convergence of the particles on local minimizers due to the sole use of own gathered information, the concept of neighborhood was introduced (Kennedy 1999; Suganthan 1999). Specifically, each particle assumes a set of neighboring particles with which it exchanges information, i.e., the best position among all the particles that constitute a neighborhood is communicated only among them and it is used for their velocity update. To put it formally, a neighborhood of the ith particle is defined as a set:

$$ {\text{NB}}_{i,s} = \{ j_1, j_2, \ldots, j_s \}, \quad j_k \in I, \quad k=1,2,\ldots,s, \quad i \in {\text{NB}}_{i,s}, $$

which consists of the indices of all the particles with which it can exchange information. Then, the neighborhood’s best position:

$$ p_{g_i} = \arg \min_{j \in {\text{NB}}_{i,s}} \{f(p_j)\}, $$
(1)

is used along with p i to update the ith particle’s velocity at each iteration. The parameter s defines the number of particles that constitute the neighborhood and it is often called the neighborhood size. Obviously, it must hold that 1 ≤ s ≤ N. By definition, the ith particle is included in its own neighborhood. In the special case where s = N, the whole swarm constitutes the neighborhood. The latter case defines the so-called global PSO model (denoted as gbest), while strictly smaller neighborhoods correspond to the local PSO model (denoted as lbest).

The schemes that are used for determining the particles that constitute each neighborhood are called neighborhood topologies and they have a crucial impact on PSO’s performance. Probably, the most trivial choice is the random selection of s neighbors per particle. However, this scheme is accompanied with an undesirable lack of control of the information flow among the particles, which may prohibit the effective exploration of the search space within a reasonable number of iterations. Therefore, in practice it is rarely favored against alternative topologies, such as the ring topology. According to it, each particle assumes as neighbors the particles with its adjacent indices. The number of neighbors is determined by a parameter, r, called the neighborhood radius. Thus, a ring neighborhood of radius r of the ith particle, is defined as the set of indices:

$$ {\text{NB}}_{i,r} = \{ i-r, i-r+1,\ldots,i,\ldots, i+r-1, i+r \}, $$

where the indices recycle after the value N, i.e., the index 1 is assumed to follow immediately after N. The recycling gives the sense of moving on a clockwise ordered ring, justifying the name of this topology.

Based on the definitions above, the iterative scheme of PSO is defined as follows (Clerc 2002):

$$ v_{ij}(t+1)=\chi [v_{ij}(t) + c_1 {\mathcal{R}}_1 (p_{ij}(t) - x_{ij}(t)) + c_2 {\mathcal{R}}_2 (p_{g_i,j}(t) - x_{ij}(t))], $$
(2)
$$ x_{ij}(t+1)=x_{ij}(t) + v_{ij}(t+1), $$
(3)

where, \(i = 1,2,\ldots,N; \, j = 1,2,\ldots,n; \) the parameter χ is the constriction coefficient; acceleration constants c 1 and c 2 are called the cognitive and social parameter, respectively; and \(\mathcal{R}_1, \mathcal{R}_2, \) are random variables uniformly distributed in the range [0, 1]. It shall be noted that a different value of \(\mathcal{R}_1\) and \(\mathcal{R}_2\) is sampled for each i and j in Eq. (2). Also, the best position of each particle is updated at each iteration, as follows:

$$ p_{i}(t+1) = \left\{\begin{array}{ll} x_{i}(t+1), & \hbox{if}\,f\left(x_{i}(t+1)\right) < f\left(p_{i}(t)\right),\\ p_{i}(t), &\hbox{otherwise}, \end{array} \right. \quad i \in I. $$
(4)

The PSO variant described above was introduced by Clerc and Kennedy (2002) and it has gained increased popularity, especially in interdisciplinary applications. This comes on account of its close relevance to the original PSO model, which implies simplicity and efficiency, as well as on its theoretical support. Based on the stability analysis (Clerc 2002), the parameter set:

$$ \chi = 0.729, \quad c_1 = c_2 = 2.05, $$
(5)

was determined as a satisfactory setting that produces a balanced convergence speed of the algorithm. Nevertheless, alternative settings have been introduced in relevant works (Trelea 2003).

The initialization of swarm and velocities is usually performed randomly and uniformly in the search space, although more sophisticated techniques can enhance the overall performance of PSO as reported in (Parsopoulos 2002a, b). Pseudocode of PSO is reported in Algorithm 1, where its essentially synchronous nature becomes apparent. By this, we mean that all particles update their velocities and current positions (lines 8–14 of Algorithm 1) prior to the update of their memory (lines 15–18). Alternative models that work rather asynchronously have also been proposed. Their basic elements are briefly presented in the next section.

2.2 Asynchronous particle swarm optimization

Asynchronous PSO models (henceforth denoted as PSO-ASY) have been developed as alternatives to the synchronous model described in the previous section. The main difference from the synchronous PSO is that, in a given iteration, each particle updates and communicates its memory to its neighbors immediately after its move to a new position. Thus, the particles that remain to be updated in the same iteration can exploit the new information immediately, instead of waiting for the next iteration as in the synchronous model.

The differences from the synchronous PSO model become apparent in the pseudocode of Algorithm 2, where we can see the existence of a single for-loop (lines 8–18) on the number of particles. Thus, for the determination of \(p_{g_i}\) in line 10 of Algorithm 2, the latest findings of all particles in the neighborhood are taken into consideration even if they were achieved at the same iteration. This is the reason for dropping the iteration index of \(p_{g_i}\) in line 12, simply denoting \(p_{g_i}\) instead of \(p_{g_i}(t).\) Contrary to this, the corresponding velocity update in line 11 of Algorithm 1 is solely based on the information coming from the previous iteration.

The impact of the asynchronous update in PSO’s performance heavily depends on the given optimization problem and the considered implementation. In general, the asynchronous model is characterized by faster disclosure of new best positions than in the standard (synchronous) PSO model, thereby increasing convergence speed. This comes at the cost of getting trapped by rapidly attracting all particles to a deceitful solution. Thus, the preference between synchronous or asynchronous PSO shall be dictated by the specific problem at hand.

However, in practice it may also be dictated by the available hardware. For example, in parallel environments it is desirable to have the smallest possible idle time per processor. In such environments, the time required per function evaluation plays a significant role. If the evaluations require almost identical time for any point in the search space, the synchronous PSO is satisfactory since all particles will be evaluated and updated almost concurrently. On the other hand, if there is significant time deviation for the evaluation of different points in the search space, the asynchronous model is expected to outperform the synchronous one with respect to the required running time.

Actually, Algorithm 2 introduces a parametric PSO-ASY model. The parameter ρasy is used to simulate both the serial (single-processor) and parallel implementation of PSO-ASY by properly tuning its value in line 9, as follows:

$$ \rho_{\rm asy}=0 \hbox{\,(serial PSO-ASY)}, \quad 0 < \rho_{\rm asy} < 1 \hbox{\,(parallel PSO-ASY)}. $$
(6)

Several asynchronous PSO variants have been proposed in the literature (Akat 2008; Desell et al. 2009; Gaz 2007; Hernane et al. 2010; Koh et al. 2006). Most of these approaches refer to parallel implementations of PSO that were shown to be competitive or even superior to the synchronous ones. Nevertheless, even in serial implementations the asynchronous models have exhibited promising results.

In the next section, the proposed PSO-DLI algorithm is described.

3 Particle swarm optimization with deliberate loss of information

The inspiration behind the development of the proposed Particle Swarm Optimization with Deliberate Loss of Information (PSO-DLI) algorithm, sprang from the analysis of Clerc and Kennedy (2002) on the dynamics of the PSO variant defined by Eqs. (2) and (3) in Sect. 2.1. Based on this analysis, the constriction coefficient model of PSO was distinguished among others on the basis of its good convergence properties. The analysis was based on the assumption that the best positions of the particles remain fixed throughout the algorithm’s execution, aiming at the determination of proper model and parameter configurations such that the swarm consistently converges on sub-optimal solutions. In fact, Clerc and Kennedy proved convergence towards a state where the velocities vanish, while the particles’ positions approximate a linear combination of the two (fixed) best positions involved in Eq. (2).

However, in realistic cases the best positions of the particles do not remain fixed. Even in problems of moderate difficulty, it is experimentally verified that the best positions are frequently changing. This implies that the dynamics of the model are restarted after each best position update, prohibiting the particles from taking full advantage of the model’s convergence properties. Moreover, it has been experimentally observed that, especially at the later stages of the optimization procedure, only some of the best positions’ updates offer critical new information to the particles, while the rest lead only to marginal improvements.

These critical observations suggest that a particle may not be able to deploy its exploration/exploitation capability due to frequent changes of its best position, which may be rather unnecessary with respect to the gained improvement. In addition, each best position update implies that a function evaluation preceded. It is widely accepted that the total number of function evaluations constitutes the main computational cost of optimization algorithms. This is based on the reasonable assumption that, in difficult problems, the time required for the evaluation of the underlying objective function dominates the rest of the algorithm’s operations. Thus, the regular evaluation of each particle and its best position update at every single iteration of the algorithm, may not only disrupt its convergence dynamics but also aggravates it with possibly futile additional cost whenever its discoveries are of minor importance.

The main idea behind PSO-DLI is the alleviation of consistent evaluation and best position update of each particle at each iteration, without disrupting the particles’ motion. This can be achieved by allowing each particle to move to a new position using Eqs. (2) and (3) but probabilistically deciding whether it shall be evaluated and update its best position or not. If the particle is eventually evaluated then its best position update follows the standard procedure of Eq. (4). On the other hand, if it is not selected for evaluation then it simply remains to the new current position, ignoring its function value and retaining its previous best position.

By accepting this loss of information, i.e., the neglected function value at the new position, the available computational budget (maximum number of function evaluations) of the algorithm is retained, without interrupting the particle’s convergent motion towards the already detected best positions. Naturally, this fosters the danger of possibly disregarding significant information. Nevertheless, the proposed scheme is expected to be advantageous under the (experimentally verified) assumption that a large number of best position updates do not offer radical information, i.e., they do not enhance the particle’s exploration capability. In fact, they suppress its exploitation activity stemming from its theoretically implied tendency for convergence towards an aggregation of previously detected best positions.

A simple scheme for probabilistically selecting the particles that will loose information, can be based on a fixed user-defined probability value, \(\rho_{\rm loss}\in [0,1]. \) Then, for the ith particle at iteration t, a uniformly distributed random number, \(\mathcal{R} \in [0,1], \) is sampled and the particle is evaluated and updates its best position only if:

$$ {\mathcal{R}} > \rho_{\rm loss}. $$
(7)

Otherwise, it retains its best position and suffers the loss of its current position’s function value. At a first glance, the proposed scheme seems to simply resemble the procedure of the asynchronous PSO model described in Sect. 2.2. This is partially true, since the evaluation procedure of PSO-DLI is asynchronous in nature. However, it has the unique characteristic that the particles that suffer loss of information are not held at their previous positions but they are allowed to follow the dynamics of the constriction coefficient model by moving to a new (unevaluated) current position. This way, it is expected that the gain from the particle’s exploitation capabilities, along with its convergent behavior, will offer significantly better information when it will be eventually evaluated at a later iteration. To the best of the authors’ knowledge, this property has not been considered in the established asynchronous models up-to-date.

The PSO-DLI scheme is clarified in the pseudocode of Algorithm 3. The main difference from Algorithms 1 and 2, lies beneath line 15 of Algorithm 3 where the randomized evaluation and update decision takes place. Notice that the particle moves without intervention to the new position (lines 11–12) regardless of the decision taken for its evaluation and best position update. Obviously, since PSO-DLI is expected to perform less function evaluations per iteration than the original PSO, the total number of iterations that can be conducted for a specific budget of F max function evaluations is higher for PSO-DLI than the standard synchronous PSO.

Also, the implementation of PSO-DLI requires only minor additional effort than PSO, since there are no complicated procedures incorporated into the algorithm. Thus, it is easily understood that it can be very useful in applications where the original constriction coefficient (or its equivalent inertia weight) PSO model is frequently used. Moreover, PSO-DLI retains all the advantages of the PSO scheme regarding its potential for parallelization and combination with other global or local optimization algorithms. Additionally, this is accompanied by a remarkable increase in efficiency as it is shown in the next section.

4 Experimental results

PSO-DLI constitutes a minor modification of the standard constriction coefficient PSO model, although with a great performance impact. It aims at offering an efficient alternative to the standard PSO model in relative applications. Thus, the experimental setting in the present work primarily targeted at a first validation of PSO-DLI in widely used test problems and challenging applications, against the specific PSO model as well as standard versions of different popular EAs for tackling such problems.

For this purpose, four rounds of experiments were conducted. In the first round, PSO-DLI was validated against the standard synchronous and asynchronous PSO model on five widely used benchmark problems, providing the first evidence on its efficiency. In the second round of experiments, the same algorithms were applied on six real-world application problems modeled as systems of nonlinear equations. Additionally to the aforementioned comparisons, PSO-DLI was also compared against different evolutionary approaches on these problems. For this purpose, a Genetic Algorithm (GA), the basic Differential Evolution (DE) variant as well as the MONS approach (Grosan 2008) was considered.

In the third round of experiments, PSO-DLI was further assessed against established EAs on harder instances of the five benchmark problems used in the first round. More specifically, rotated and shifted versions of these problems, which have been used as benchmarks for the IEEE CEC’08 large-scale optimization competition (Tang et al. 2007), were considered. In this framework, we probed PSO-DLI’s competitiveness against two specialized PSO-based approaches for large-scale optimization, namely the Efficient Population Utilization Strategy for Particle Swarm Optimizer (EPUS-PSO) (Hsieh et al. 2008, 2009) and the Dynamic Multi-Swarm Particle Swarm Optimizer with local search (DMS-PSO) (Zhao et al. 2008).

In the last round of experiments and despite the fact that PSO-DLI was neither designed nor tuned for large-scale problems, we took a further step and applied it on the test suite used in a recently published special issue of Soft Computing (Lozano et al. 2011). Our main consideration in this case was the evaluation of its ability to perform closely to the 16 algorithms presented in the special issue.

In all experiments, the employed PSO-DLI was based on the lbest variant of the constriction coefficient PSO model with a ring topology of radius 1. This choice was driven mostly by its popularity and good exploration properties. Regarding its parameters, the default values, χ = 0.729, c 1 = c 2 = 2.05, defined in Eq. (5) were used. All the considered test problems are described in Appendix with the exception of the 19 problems of the fourth round of experiments that were omitted due to space limitations. However, the interested reader can have full access in all relevant information and sources through the web site http://sci2s.ugr.es/EAMHCO. In all experiments, the swarm was randomly and uniformly initialized within the corresponding search space.

4.1 First round of experiments: standard test suite

In the first round of experiments, the performance of PSO-DLI was assessed on the five widely used benchmark problems defined in "Standard test suite" of Appendix. These problems constitute the standard test suite and will be henceforth denoted as TP0–TP4. Their popularity is attributed to their different features, including unimodality/multimodality, separability/non-separability, generalization in arbitrary dimension as well as strongly/loosely correlated variables. The dimensions and range considered for each problem of the standard test suite in our experiments are reported in Table 1.

Table 1 Dimension and range considered for each problem of the standard test suite

A number of comparative experiments were conducted on the standard test suite. More specifically, we employed the synchronous lbest PSO model, the corresponding asynchronous PSO-ASY model as well as the proposed PSO-DLI algorithm. The PSO-ASY model was considered with probabilities:

$$ \rho_{\rm asy} \in \{ 0.0, 0.3, 0.6, 0.9\}, $$

which correspond to both its serial and parallel variant. The corresponding variants will be henceforth denoted as ASY0.0, ASY0.3, ASY0.6 and ASY0.9, respectively. PSO-DLI was considered with the same loss probabilities:

$$ \rho_{\rm loss}\in \{ 0.0, 0.3, 0.6, 0.9 \}. $$

Notice that ρloss = 0.0 corresponds to the standard PSO algorithm (no information loss). The rest will be henceforth denoted as DLI0.3, DLI0.6 and DLI0.9, respectively.

For each problem instance and algorithm variant, 100 independent experiments were conducted. At each experiment, the algorithms were allowed to perform a total number of F max = 1000 × n function evaluations using a swarm of N = 10 × n particles, where n is the dimension of the corresponding problem instance. The employed parameter values are summarized in Table 2.

Table 2 Parameter values for the considered PSO, PSO-ASY and PSO-DLI algorithms

At the end of each experiment, the following quantities were recorded:

  1. (a)

    the best detected solution and its function value;

  2. (b)

    the number of iterations performed until F max was reached;

  3. (c)

    the total number of best positions updates, i.e., the total number of times where the particles discovered better best positions than the ones they had possessed.

Obviously, for the standard synchronous PSO model the number of iterations referred in (b) was always equal to F max /N. This is a consequence of the fixed number of function evaluations performed per iteration in standard PSO. However, this is not always the case for PSO-ASY and PSO-DLI since some particles may skip evaluation. Thus, these approaches are expected to perform more iterations than PSO for a given number of function evaluations. Also, the recorded quantity in (c) was considered as a measure of the swarm’s exploitation activity.

The results are numerically reported in Table 3. More specifically, for each problem instance, the mean and the standard error (sample’s standard deviation) of the best attained solution values are reported, along with the corresponding numbers of iterations and best position updates (all averaged over 100 experiments) for the standard PSO as well as for the best-performing PSO-ASY and PSO-DLI variants. The smallest mean function value per problem instance is boldfaced along with its standard error.

Table 3 Results for the standard test suite

The results are also graphically illustrated in Fig. 1. Specifically, the distributions of the attained best solution values are depicted in boxplots per problem instance and algorithm. On each box, the central mark is the median, the edges of the box are the 25th and 75th percentiles, the whiskers extend to the most extreme data points not considered outliers, and outliers are plotted individually (red crosses). Moreover, the notches define 95 % confidence intervals of median equality for box-to-box comparisons.

Fig. 1
figure 1

Graphically illustrated results for the standard test suite

Each row of boxplots in Fig. 1 corresponds to a specific problem dimension (10, 50 or 100) and it is followed by a corresponding row of bargraphs that demonstrate the percentage (%) of performance improvement achieved by PSO-ASY and PSO-DLI against PSO, with respect to the mean (black bars) and standard error (white bars) of the best attained solution values. Obviously, negative bars indicate percentage of performance decline. These graphs aim at providing visual comparisons between the algorithms, using the standard PSO as reference point. A more compact graphical representation of the relative performance of PSO-DLI and PSO-ASY with respect to PSO, is offered in the left part of Fig. 3.

The derivation of sound conclusions regarding the performance differences between the algorithms was further facilitated by statistical significance tests. For this purpose, the nonparametric Wilcoxon rank-sum test was used for pairwise comparisons of PSO-DLI with PSO-ASY and PSO under the null hypothesis that “the best solution values achieved by the compared algorithms are drawn from identical continuous distributions with equal medians” at a 95 % level of significance. Then, the total number of cases (out of 15) where PSO-DLI had statistically significant better performance than the other approaches was recorded and it is reported in Fig. 2i. Finally, the frequency of appearance of each variant of PSO-ASY and PSO-DLI as the best-performing among all variants of the same algorithm per problem dimension is reported in Fig. 2ii and iii, respectively.

Fig. 2
figure 2

Frequencies of domination among different algorithm variants

A first inspection of the results presented in Table 3 and Fig. 1 offers some interesting conclusions. As we can see, both PSO-ASY and PSO-DLI clearly outperformed the standard PSO in all cases. PSO-DLI achieved the overall best performance in all but two cases, namely the 100-dimensional instances of TP1 and TP4, where it was highly competitive to the dominant PSO-ASY variant. This is clearly derived also from Figs. 2i and 3. Moreover, DLI0.9 was the most promising PSO-DLI variant in almost all problem instances, with a single exception in TP1. This is depicted also in Fig. 2iii. The superiority of DLI0.9 with respect to the mean attained best solution value was habitually accompanied by the smallest standard errors, which is indicative of its robustness. On the other hand, ASY0.0 was the dominant among the PSO-ASY variants especially for higher-dimensional cases as we can see in Fig. 2ii. However, in the 10-dimensional cases the other PSO-ASY variants were also distinguished. This indicates sensitivity of PSO-ASY on the problem’s dimension and structure.

Fig. 3
figure 3

Left relative performance of PSO-DLI and PSO-ASY with respect to the standard PSO. Right domination scaling between PSO-DLI and PSO for different swarm sizes

A closer inspection of the results enriches our knowledge on PSO-DLI’s performance. The last two columns of Table 3 show that PSO-DLI always performs the largest number of best position updates. This is a sound indication of PSO-DLI’s better exploitation activity that is evidently promoted by the undisrupted motion of the particles. This evidence, combined with the dominance of DLI0.9 against the other PSO-DLI variants, suggests that higher values of the loss probability can be associated with better performance. As expected, the higher number of best position updates is associated with a higher mean number of iterations as it is displayed in the corresponding columns of Table 3. This is a natural consequence of PSO-DLI’s probabilistic particle-update scheme as it was explained in Sect. 3.

Regarding the effect of problem’s dimension on performance, Fig. 1 draws a clear picture. Observing the provided bargraphs in columns (same problem, increasing dimension), we can see an undisputed decline of the attained improvement (recall that bargraphs denote percentage of improvement/decline). This can be attributed to the total number of function evaluations, F max, which was linearly increased with the dimension (recall that F max = 1000 × n), while there are experimental evidence that the degree of difficulty of a given problem increases exponentially with its dimension.

Additionally, the verified influence of the swarm size on PSO’s dynamic (Bartz-Beielstein et al. 2004) impelled us to consider the case of swarm size scaling in order to verify the potential superiority of PSO-DLI against PSO. For this purpose, the experiments for PSO-DLI and PSO were repeated for all problems and dimensions, with different swarm sizes, \(N=1 \times n, 2 \times n,\ldots,10 \times n. \) Statistical significance tests were conducted between the algorithms to reveal the dominant approach. The results are reported in the right part of Fig. 3. More specifically, for each problem and dimension, a bar illustrates the number of different (out of 10) swarm sizes where the one algorithm outperformed (with statistical significance) the other as well as the number of cases with statistically indistinguishable performance. As we can see, the results are in good agreement with those previously presented for higher swarm size, with PSO being competitive only on a few cases, mostly for lower dimensions.

In conclusion, the experiments on the standard test suite suggest that PSO-DLI can be very competitive to fundamental synchronous and asynchronous PSO schemes. In order to attain further intuition on PSO-DLI’s performance on less typical benchmark problems, further experiments were conducted on a set of challenging applications that involve the solution of nonlinear systems. The obtained results are reported in the next section.

4.2 Second round of experiments: nonlinear systems

In the second round of experiments, the real-world applications defined in "Nonlinear systems" of Appendix were considered. The corresponding problems are modeled as systems of nonlinear equations and they are formulated as global optimization problems through a transformation described in "Nonlinear systems". These problems will be henceforth denoted as TP5–TP10 and their dimensions and ranges are reported in Table 4. The same variants of the algorithms as in the previous section were considered, along with the parameter values reported in Table 2 (excluding problem dimensions). The presentation of the results follows closely the one in the previous section. Thus, data are numerically reported in Table 5 and graphically illustrated in Figs. 4 and 5.

Table 4 Dimension and range for the nonlinear systems
Table 5 Results for the nonlinear systems

The general picture of the results is aligned to that of the previous section. Namely, PSO-DLI outperformed PSO-ASY and PSO in all test problems, offering significant solution improvements that range from almost 30 % up to more than 90 % as it is depicted in the bargraphs of Fig. 4. This superiority was also supported by the statistical tests and it is reflected in PSO-DLI’s domination depicted in Fig. 5i. Moreover, the relative improvements achieved by PSO-DLI and PSO-ASY against PSO were always in favor of the proposed approach as illustrated in Fig. 5iii. Indeed, as we can see PSO-ASY’s improvements were limited in a range between 10 % and slightly over 20 %, while at the same time PSO-DLI’s improvements soared over 80 %. The two distinguishable clusters in the scatter plot of Fig. 5iii may be the results of possible structural resemblance of the corresponding problems, although this speculation requires further investigation that lies out of the scope of this paper.

Fig. 4
figure 4

Graphically illustrated results for the nonlinear systems

Fig. 5
figure 5

Frequencies of domination among different algorithm variants and relative performance of PSO-DLI and PSO-ASY with respect to the standard PSO for the nonlinear systems

An interesting evidence from this round of experiments was the designation of DLI0.6 as the best among PSO-DLI variants in more than half of the problems. This is in agreement with the expectation of better PSO-DLI performance under higher loss-probability values, which was identified in the standard test suite. However, it can be conceived also as an evidence of a plausible dependence of the most efficient variant on the corresponding problem at hand. Nevertheless, either PSO-DLI variants retained the highest number of best positions updates in all cases.

The second round of experiments was enriched with further comparisons of PSO-DLI with other established algorithms. More specifically, a steady state Genetic Algorithm (GA) (Goldber 1989) as well as the Differential Evolution (DE) (Storn 1997) algorithm were tested against different PSO-DLI variants under the configuration reported in Table 6. The employed GA approach was implemented using the GAlibFootnote 2 software. In order to facilitate comparisons also with the results provided in Grosan (2008) for the multi-objective MONS approach, the experimental setup reported in Table 7 was adopted for all algorithms.

Table 6 Configuration of the employed GA and DE algorithm
Table 7 Swarm size and maximum number of function evaluations for the nonlinear systems in Grosan (2008)

In accordance to the previous experiments, we performed 100 independent runs per algorithm using the provided maximum number of function evaluations and recording the best solution value attained per experiment. The obtained results are reported in Table 8 with respect to the mean and standard error of the best function values. The results of the MONS approach were adopted from the original source (Grosan 2008) where no standard errors are reported.

Table 8 Comparative results of PSO-DLI with the MONS, GA and DE algorithm. Results for MONS are reproduced from (Grosan 2008) were no standard errors are reported

Some interesting observations can be made in Table 8. Firstly, we can observe that all PSO-DLI variants were among the best-performing algorithms in all problems. Particularly in TP5, TP6 and TP10, they outperformed the rest of the algorithms. However, DE was shown to outperform PSO-DLI in the rest of the test problems. Also, contrary to our previous findings, DLI0.3 was shown to be superior than DLI0.6 and DLI0.9.

A reasonable explanation for this outcome lies in the specific experimental setup. More specifically, the number of function evaluations adopted from (Grosan 2008) was significantly higher by several orders of magnitude than the one used in our previous experiments. This was beneficial both for the DE approach as well as for DLI0.3, since any possible gain produced by the proposed loss of information scheme was absorbed by the competency of the algorithm due the excessive computational budget. Thus, it can be inferred that increasing the available computational budget produces a tendency of PSO-DLI to approximate the behavior of the standard PSO.

Although the reported results for the rest of the algorithms are rather indicative, PSO-DLI has shown convincing evidence that it can reach high performance standards. This was further verified on harder instances of the standard test suite as reported in the following section.

4.3 Third round of experiments: modified standard test suite

In the third round of experiments, the performance of PSO-DLI was assessed on shifted and rotated versions of the problems of the standard test suite, henceforth called the shifted test suite. These problems, denoted as SH-TP0–SH-TP4, were proposed in the IEEE Congress on Evolutionary Computation 2008 (IEEE CEC’08) for large-scale optimization competition (Tang et al. 2007) and they are defined in "Shifted test suite" of Appendix. The PSO-DLI variants used in our previous experiments were applied on the shifted test suite and their performance was compared with that of two PSO-based approaches that participated in the aforementioned competition at IEEE CEC’08, namely the Efficient Population Utilization Strategy for Particle Swarm Optimizer (EPUS-PSO) (Hsieh et al. 2008, 2009) and the Dynamic Multi-Swarm Particle Swarm Optimizer with local search (DMS-PSO) (Zhao et al. 2008).

The test problems were all considered in their 100-dimensional instances, while their ranges were the ones reported in Table 1 for n = 100. The parameter configuration of the PSO-DLI variants was identical to that of Table 2 for n = 100. This setting is in accordance to the one used for EPUS-PSO and DMS-PSO in Tang et al. (2007) with respect to the available computational budget, i.e., the total number of function evaluations.

The obtained results are reported in Table 9. As required for the specific test problems, the algorithms were compared with respect to the absolute error of the obtained best function value to the true global minimum, averaged over all experiments. The results for EPUS-PSO and DMS-PSO were adopted from their original references (Hsieh et al. 2008; Zhao et al. 2008). In the case of DMS-PSO, the minimum resolution close to zero was not explicitly reported. These cases are marked in Table 9.

Table 9 Results for the 100-dimensional instance of the shifted test suite

As we can see, the PSO-DLI variants produced competitive results. In SH-TP0 and SH-TP3, DLI0.9 strikingly achieved the best performance (excluding the marked cases). This is surprising if we take into consideration that PSO-DLI constitutes a very simple approach that does not employ local search or any other sophisticated mechanism specifically suited to this kind of problems, in contrast to the other two approaches. Even for the rest of the problems its performance was highly competitive to the specialized algorithms. This is a very promising evidence that offers motivation for the development of hybrid PSO-DLI variants that can incorporate mechanisms for enhancing their performance in large-scale optimization problems.

4.4 Fourth round of experiments: large-scale competition problems in special issue

Recently, a set of 19 benchmark functions was proposed as a common test suite for the assessment of scalability of metaheuristics on large-scale problems. The test suite constituted the experimental basis for the “special issue on scalability of evolutionary algorithms and other metaheuristics for large-scale continuous optimization problems” of the Soft Computing journal. General information can be found in the editorial of the special issue (Lozano et al. 2011), while their definitions are freely available on the internet,Footnote 3 hence they are omitted due to space limitations. Sixteen algorithms appeared in the special issue and competed against each other on the specific problems for dimensions ranging from 50 up to 1000. The average performance for each algorithm is summarized in a table posted on the previously mentioned web site.Footnote 4 The size of the table is prohibitive for replication here, thus we refer the reader directly to the web source.

We considered the specific test suite as our last experimentation framework in the present study, with the 19 test problems being henceforth denoted as SC-TP1–SC-TP19. Since the tuning of PSO-DLI for tackling large-scale problems of very high dimensionality (e.g., 1000) was out of scope of the present study, we restricted our experiments to dimensions 50, 100 and 200, taking a further step from the dimensions considered in the previous section. The configurations of PSO-DLI that were used in the previous sections, were also adopted in this set of experiments.

PSO-DLI was applied on all test problems according to the experimental setup suggested in (Lozano et al. 2011). Its average error was recorded and reported in Table 10. The main purpose was to investigate its ability to approximate the error achieved by the rest of the algorithms, within a prescribed accuracy range. Thus, we compared PSO-DLI’s average errors with the corresponding ones reported in the web sources above, for the 16 algorithms that competed in the special issue. The comparisons were made for the different levels of accuracy defined in the following set:

$$ \{ 10^{-5}, 10^{-4}, 10^{-3}, 10^{-2}, 10^{-1}, 10^{0}, 10^{1}, 10^{2}, 10^{3}, 10^{4}, 10^{5} \}. $$

If PSO-DLI was capable of achieving the same (or better) performance with another algorithm with the selected accuracy level on a specific problem-dimension pair, then this was considered as a successful instance. We recorded the successes of PSO-DLI for all test problems, dimensions, algorithms and accuracy levels, which counts a total of 19 × 3 × 16 = 912 instances per accuracy level. Obviously, the remarkably high accuracy values included above, were considered for revealing the scaling in the successful instances.

Table 10 Average error of PSO-DLI in test problems SC-TP1–SC-TP19

The results are graphically illustrated in Fig. 6. Specifically, for each accuracy level there is a bar that shows the number of successful instances (out of 912). Also, each bar is separated in three parts of different colors, denoting the corresponding successful instances per problem dimension. The figure clearly suggests that for demanding accuracies PSO-DLI was capable of achieving the same (or better) performance in more than 1/3 of the cases. This can be consider as a promising sign since PSO-DLI was neither designed nor tuned for such high-dimensional cases as the rest of the algorithms. Nevertheless, it constitutes a strong motivation for the future development and improvement of PSO-DLI schemes in order to render it fully competitive to the state-of-the-art in large-scale problems.

Fig. 6
figure 6

Cumulative distribution of the number of successful instances per accuracy level

As a final observation, we shall mention the consistency of PSO-DLI under dimension and accuracy scaling, as it is implied by the smooth transition of the bars when accuracy increases as well as by the retained proportion of each color in the bars.

5 Conclusions

We introduced PSO-DLI, an algorithm based on the constriction coefficient model of PSO, which promotes asynchronous update and undisrupted movement of the particles accompanied by loss of information. Representative variants of the algorithm were extensively tested both on benchmark problems and challenging applications. Also, they were compared with other widely used algorithms, some of which incorporate sophisticated techniques to tackle specific types of problems.

The obtained results were very promising, offering intuition on PSO-DLI’s performance under different experimental settings. The proposed approach was able to outperform the rest of the algorithms in many problem instances. It was also revealed that in cases where the available computational budget was rather limited, the proposed scheme with higher loss probabilities exhibited very promising performance. On the other hand, in cases where excessive computational budget was available, the gain became higher for smaller loss probabilities indicating that PSO-DLI tends to resemble the standard PSO model.

In addition, PSO-DLI is based on a simple idea that takes full advantage of the dynamics of the constriction coefficient PSO model. Its implementation requires only minor modifications of the original PSO model and alleviates expensive bookkeeping operations. Taking into consideration its encouraging results as well as its asynchronous nature, PSO-DLI could be considered as an effective basis for the development of hybrid PSO-based algorithms for parallel systems and large-scale problems. In fact, this can be a very interesting direction for future research on the proposed algorithm.

Nevertheless, it must be emphasized that PSO-DLI is designed to take full advantage of the exploitation properties that accompany the specific constriction coefficient PSO model. Thus, the generalization of the derived analysis and conclusions in radically different environments such as in integer, noisy or dynamic problems, where the average behavior of the constriction coefficient PSO model is altered, is not straightforward and can constitute a challenging subject for further investigation, along with the demanding large-scale cases.