Improved and generalized learning strategies for dynamically fast and statistically robust evolutionary algorithms

https://doi.org/10.1016/j.engappai.2007.06.005Get rights and content

Abstract

This paper characterizes general optimization problems into four categories based on the solution representation schemes, as they have been the key to the design of various evolutionary algorithms (EAs). Four EAs have been designed for different formulations with the aim of utilizing similar and generalized strategies for all of them. Several modifications to the existing EAs have been proposed and studied. First, a new tradeoff function-based mutation has been proposed that takes advantages of Cauchy, Gaussian, random as well as chaotic mutations. In addition, a generalized learning rule has also been proposed to ensure more thorough and explorative search. A theoretical analysis has been performed to establish the convergence of the learning rule. A theoretical study has also been performed in order to investigate the various aspects of the search strategy employed by the new tradeoff-based mutations. A more logical parameter tuning has been done by introducing the concept of orthogonal arrays in the EA experimentation. The use of noise-based tuning ensures the robust parameter tuning that enables the EAs to perform remarkably well in the further experimentations. The performance of the proposed EAs has been analyzed for different problems of varying complexities. The results prove the supremacy of the proposed EAs over other well-established strategies given in the literature.

Introduction

Since its advent, evolutionary optimization has gone through various stages of evolution that have made it more robust and fast; robust enough to be applied to all spheres of optimization problems and fast enough to take decisions in micro-seconds (Dimopoulos and Zalzala, 2000; Nissen and Propach, 1998; Yao et al., 1999). In the present scenario, evolutionary algorithms (EAs) are not limited to the applications in artificial intelligence, and have been increasingly used in various dimensions to solve real world problems (Dimopoulos and Zalzala, 2000; Sinha et al., 2003). Although, a plethora of literature pertaining to the search strategies adopted by various EAs is available (Nissen and Propach, 1998; Rana et al., 1996; Kazarlis et al., 2001; Yao et al., 1999; Salomon, 1998; Choi and Oh, 2000; Yoon and Moon, 2002; Kim and Myung, 1997; Storn, 1999), most of them restrict themselves to solving a particular class of optimization problem and thus, fail to provide generalized strategies that can be robustly used for wide spectrum of optimization problems in science, business and engineering applications. In general, optimization problems can be classified into two groups—numerical optimization and combinatorial optimization (Tsai et al., 2004; Gen and Cheng, 1999). Yet another classification exists that is based on the representation schemes—binary string, floating point string and integer bit string representation (Rana et al., 1996; Kazarlis et al., 2001; Yao et al., 1999; Salomon, 1998; Choi and Oh, 2000; Yoon and Moon, 2002; Kim and Myung, 1997; Storn, 1999; Zhang and Leung, 1999; Burke and Newall, 1999). In general, the first two representations are more suited and easy to implement on numerical optimization problems; whereas, the third representation enjoys solving the combinatorial optimization problems with a greater ease and efficiency (Nijssen and Bäck, 2003; Goldberg, 1989). Thus, to propose effective algorithms that would efficiently work in all the representations is a convoluted task. The schematic representation of various optimization problems along with their most suited representation classes are shown in Fig. 1.

All of the above-mentioned factors motivated the authors to propose robust and fast EAs with improved learning strategies to achieve the global optimization in all of these cases. Before introducing the proposed algorithms, the general formulation of a numerical optimization problem can be mathematically described asminimizeg(t),t=(t1,t2,,tn)Λn,where tΓX,Γ represents the ‘feasible region’ defined asΓ={tΛn|cj(t)0j{1,2,,J}},where cj(t),j{1,2,,J} are the ‘constraints’ on the problem; the n-dimensional ‘search space’ is defined by XΛn and is constrained by the ‘parametric constraints’ given below:ti̲tit¯i,i{1,2,,n}where ti̲ and t¯i represent the lower and upper limits of the variable ti. Here, it is evident that the solution is represented as floating point string of type A (Fig. 1); however, it can also be represented in the binary string (type B of Fig. 1) by considering the representation of floating numbers into binary numbers.

For the second class of optimization problems, i.e. combinatorial optimization problems, the general formulation can be given asminimizeh(s)s=(s1,s2sn)S,where S is the set of integers; sJ,J denotes the ‘feasible range’ of variables and is given byJ={sS|cj(s)0j{1,2,,J}},where cj(s) represents the ‘constraints’ of the problem; the ‘search range’ is defined by S, which is constrained by the ‘individual range’ defined byc̲icic¯i,i{1,2,,n},where c̲i and c¯i are the lower and upper limits of the variable cj. Generally, this formulation is utilized for the type D representation. However, the representation type C is used in the combinatorial problems class mapped by traveling salesman problem (TSP). Its general formulation is described asminimizefTSP=i=1Nj=1NdistijΩiji,j{1,2,,N},andij,whereΩij={1ifnodeiandjareselectedinthetour,0otherwise,subject to the constraints cj″, j={1, 2, …, n}, where N is the number of nodes; distij is the distance between nodes i and j; Here, only minimization cases are considered as the problems pertaining to maximization of objective can also be mapped into minimization problem by making suitable changes in the formulation.

Owing to the existence of all these formulations, authors intend to present some improved learning rules suited to each formulation in order to eliminate the difficulty and dilemma of the practitioners and users of EAs while searching the best strategy for their problem. In addition to this, various testing phases based on design of experiments are introduced in order to achieve robust parameter control and tuning.

In the earlier studies, various modifications to the initial versions of EAs have been proposed to suit the varying complexity of the optimization problems. Some of the vital factors that affect the search strategy of the EAs are: (1) population representation, (2) information sharing among members of the population, (3) mutation of the current population, (4) generation scheme for new population, (5) learning from previous generations (Kazarlis et al., 2001; Yao et al., 1999; Salomon, 1998; Choi and Oh, 2000; Yoon and Moon, 2002; Kim and Myung, 1997; Storn, 1999; Zhang and Leung, 1999; Burke and Newall, 1999; Eiben et al., 1999; Harik et al., 1999; Michalewicz et al., 2000; Runarsson and Yao, 2000; François and Lavergne, 2001; Kazarlis et al., 2001; Ong and Keane, 2004). Even a small variation in any of these factors has a considerable effect on the EA performance.

The following discussion provides an insight to some of the recent developments in the field of evolutionary computation. In order to do away the slow convergence of the classical evolutionary programming (CEP), Yao et al. (1999) introduced fast evolutionary programming (FEP) that utilized the mutation based on Cauchy function because of its higher probability of making longer jumps Runarsson and Yao (2000) introduced a novel approach to balance objective and penalty functions stochastically, i.e. ‘stochastic ranking’ and presented a fresh view on penalty function methods in terms of dominance of penalty and objective functions. Salomon (1998) presented a new hybrid approach, the evolutionary-gradient-search method for the problems pertaining to numerical optimization. A much broader collection of developments in evolutionary computation for manufacturing optimization can be found in Dimopoulos and Zalzala (2000). Zhang and Leung (1999) proposed an orthogonal genetic algorithm for multimedia cast routing that incorporated orthogonal design to crossover operation. Francois (1998) presented a mutation or selection evolutionary strategy (MOSES) to solve complex discrete optimization problems. It theoretically studied the relationship between convergence, parameters of the method, and the geometry of the optimization problem. An extensive empirical study on the synergy among multiple crossover operators has been provided in Yoon and Moon (2002).

Choi and Oh (2000) presented a new mutation rule for evolutionary programming (EP) that has been motivated from the back-propagation learning rule of the neural networks. A method of decomposing larger problems into smaller components, each of which is of a size that the EA can effectively handle, was proposed by Burke and Newall (1999) to solve timetabling problem. Eiben et al. (1999) provided a broader classification for parameter control mechanisms of EAs. They also provided a survey of various types of control that have been used by evolutionary computation community. A statistical method was described by François and Lavergne (2001), which was able to find good parameter settings for EAs. The method utilized a functional relationship between the algorithms performance and its parameter values. Yet another statistical approach can be found in Czarn et al. (2004).

Ahn and Ramakrishna (2003) described two elitism-based compact genetic algorithms in order to design efficient compact type GAs by treating them as the estimation of distributed algorithms (EDAs) to efficiently solve complex optimization problems. The compact GAs utilized the concept of successive learning in generations utilizing the probability strings for the solution representation. Various other learning approaches also persist in the literature. In general, learning algorithms are termed as learning automata approaches (Najim and Poznyak, 1994). In Agache and Oommen (2002), two new generalized pursuit algorithms were presented that falls in the category of fastest learning automata algorithms. Howell et al. (2002) proposed a hybrid genetic and learning automata approach that enjoys the merits of both the strategies. Various other relevant work like Papadimitriou (1994), Najim and Poznyak (1994) in the context of learning algorithms need to be mentioned here.

Many works have been reported in literature in order to map the system chaos and noise within the working dimensions of EAs (Nissen and Propach, 1998; Rana et al., 1996; Caponetto et al., 2003). Caponetto et al. (2003) introduced chaotic sequences instead of random sequences during all the phases of random evolution process. The approach was based on the spread spectrum characteristic of chaotic sequences. In Nissen and Propach (1998), noise function was introduced to the deterministic objective function values in order to map the practical scenario of the existence of noise while experimentation, stochastic simulation, sampling and even interaction with users.

As it can easily be manifested from the above literature review that though a wide range of modifications have been carried out to enhance the performance of EAs, a robust strategy is mandatory that can work in general for various formulations of objective functions found in the real world scenario. Also, an all purpose strategy for robust parameter control and tuning in the presence of noise is essential to help the practitioners of EAs to come out of the dilemma of choosing the control and tuning methods in changing situations.

The aims of this paper are manifold. First, it introduces the concept of developing general-purpose fast EAs that equivalently achieve better results in all the previously mentioned formulations (i.e. A, B, C, and D). Second, it utilizes the enhanced learning mechanisms for all these formulations in order to achieve better tradeoff between exploration and exploitation of the search space; and to achieve time gain over prevailing strategies. This has also been theoretically validated. Third, it introduces a new tradeoff function to decide the mutation step size. This function makes a tradeoff between the fast convergence of Cauchy function and the explorative search of Gauss function to achieve better solutions. Fourth, the chaotic sequences to map the random dynamics have also been utilized to mutate the solutions. Thus, the evolutionary strategies utilized in the paper, the mutation at each step is performed both by the tradeoff function and the chaotic sequence generator and the better offspring is evolved further. Fifth, the paper proposes a design of experiments model to conduct various experiments sequentially utilizing the orthogonal array that symbolizes the variation in various parameter values. This approach reduces the tedious task of parameter control and tuning that were earlier carried out by exhaustive experiments. Comparisons with the existing EAs are also carried out that utilize the well-designed experiments, and thus the efficacy of the strategies has been established. Finally, the noise factor has been incorporated in the deterministic objective function value, and designed experiments are conducted to establish the robustness of the algorithm.

The rest of the paper is organized in the following sequence: The next section introduces the main intricacies in defining the general-purpose EAs. Section 3 introduces new generalized learning strategies and presents the new guided mutation rule. It also provides the general-purpose algorithms for all the four problem types. Section 4 deals with the mathematical aspects of the proposed algorithms and establishes their convergence. Section 5 provides the classification of the test functions utilized to evaluate the effectiveness of the proposed algorithm for numerical optimization problems on the basis of their dimensional complexity and degree of entrapment. It also presents the test problems of combinatorial optimization utilized for the comparative study. Section 6 formulates the design of experiments for the parameter tuning and the analytical study of the proposed algorithm. The results of the extensive computational experiments, inferences drawn from it and related discussions are presented in Section 7. Section 8 concludes the paper.

Section snippets

An insight to intricacies in defining general-purpose algorithms

As discussed in the previous section, a general optimization problem can be conveniently framed into one of the formulations depicted in Fig. 1, though, with slight changes. The major obstacle in the presentation of robust algorithms that can equivalently work in any of the varying environment or formulation is the peculiar data type handling method used in it.

Due to varying representation schemes, the general rules for mutation cannot be implemented in all the cases. Hence, the following

The generalized dynamically fast EAs

This paper utilizes the following adaptations and modifications in the existing strategies

  • Learning strategy utilizing the neighborhood information and improved updating rule.

  • Guided adaptive mutation rule based on the better of the step obtained from tradeoff function between Cauchy and Gaussian distributions, and from the random chaotic distributions.

Based on the aforementioned aspects, the general rules for all the formulations are given in the following discussion.

Can the learning rule converge?

To show the convergence, the learning strategies are split into two parts—first, the learning strategies which utilize Eq. (13) and second, those utilizing Eq. (17). Let the case of first type be taken. The convergence of the strategy can be proved in two steps. First, the following theorem is established.

Theorem 1

For some ℸ >0 and И<, ∃ς*>0, υ*>0 and ginitial<∞ such that ς(0,ς*)υ(0,υ*) under Ls, probability ρ (all actions are chosen at least И times before g) >1−ℸ, ∀gginitial.

Proof

This theorem is

Test functions pertaining to type A and type B

To evaluate the performance of the proposed EAs and carry out the designed experiments for the numerical optimization problems (A and B), various benchmark test functions available in the literature have been utilized (Yao et al., 1999). Broadly the following classification of test functions has been utilized:

  • (1)

    Unimodal functions, f1f3 (i.e. with no local minima).

  • (2)

    Multimodal functions with many local minima, f4f5.

  • (3)

    Multimodal functions with few local minima, f6f7.

The functions and their

Experimental design for parameter tuning

It has been an issue of challenge in the EA optimization field to determine the parameter setting that yield efficient performance on the problems at hand. Recently, no free lunch theorem (Wolpert and Macready, 1997) has revealed that the average performance of about all pairs of algorithms is approximately same. This means that if any algorithm performs better in some particular set of problems, it is bound to perform worse in other problems. Also, if an algorithm works well on a particular

Computational results and inferences

In order to start the computational analysis, first the best performing set of parameters has been obtained according to the previously mentioned procedure. The following subsection analyzes various results obtained for parameter tuning.

Concluding remarks and future scope

This paper targets the real-time optimization from a distinctive perspective. The generalized optimization problems have been first characterized as four basic type problems on the basis of varying solution representations utilized. Further, some generalized and enhanced learning rules have been defined for the EAs. In addition, a new tradeoff function-based mutation has been introduced that ensures more explorative and thorough search. Based on these modifications in the general EAs, four new

References (45)

  • M. Gen et al.

    Algorithm for solving large scale 0–1 goal programming and its applications to reliability optimization problem

    International Journal of Computers and Industrial Engineering

    (1989)
  • M. Agache et al.

    Generalized pursuit learning schemes: new families of continuous and discretized learning automata

    IEEE Transactions on Systems, Man and Cybernetics—Part B

    (2002)
  • Chang Wook Ahn et al.

    Elitism based compact genetic algorithms

    IEEE Transactions on Evolutionary Computation

    (2003)
  • E.K. Burke et al.

    A multistage evolutionary algorithm for the timetable problem

    IEEE Transactions on Evolutionary Computation

    (1999)
  • M.V. Butz et al.

    Toward a theory of generalization and learning in XCS

    IEEE Transactions on Evolutionary Computation

    (2004)
  • R. Caponetto et al.

    Chaotic sequences to improve the performance of evolutionary algorithms

    IEEE Transactions on Evolutionary Computation

    (2003)
  • D.H. Choi et al.

    A new mutation rule for evolutionary programming motivated from backpropogation learning

    IEEE Transactions on Evolutionary Computation

    (2000)
  • M. Clerc et al.

    The particle swarm—explosion, stability, and convergence in a multidimensional complex space

    IEEE Transactions on Evolutionary Computation

    (2002)
  • A. Czarn et al.

    Statistical exploratory analysis of genetic algorithms

    IEEE Transactions on Evolutionary Computation

    (2004)
  • C. Dimopoulos et al.

    Recent developments in evolutionary computation for manufacturing optimization: problems, solutions and comparisons

    IEEE Transactions on Evolutionary Computation

    (2000)
  • A.E. Eiben et al.

    Parameter control in evolutionary algorithms

    IEEE Transactions on Evolutionary Computation

    (1999)
  • O. Francois

    An evolutionary strategy for global minimization and its Markov chain analysis

    IEEE Transactions on Evolutionary Computation

    (1998)
  • Olivier François et al.

    Design of evolutionary algorithms—a statistical perspective

    IEEE Transactions on Evolutionary Computation

    (2001)
  • M. Gen

    Reliability optimization by 0–1 programming for a system with several failure modes

    IEEE Transactions on Reliability

    (1975)
  • M. Gen et al.

    Genetic Algorithms

    (1999)
  • D. Goldberg

    Genetic Algorithms in Search, Optimization and Machine Learning

    (1989)
  • U. Hammel et al.

    Evolution strategies on noisy functions. How to improve convergence properties,

  • G.R. Harik et al.

    The compact genetic algorithm

    IEEE Transactions on Evolutionary Computation

    (1999)
  • M.N. Howell et al.

    Genetic learning automata for function optimization

    IEEE Transactions on System, Man, and Cybernatics—Part B

    (2002)
  • S.L. Hung et al.

    A parallel genetic/neural network learning algorithm for MIMD shared memory machines

    IEEE Transactions on Neural Networks

    (1994)
  • R.A. Hunt

    Calculus with Analytic Geometry

    (1986)
  • Cited by (12)

    • Automatic clustering algorithm based on multi-objective Immunized PSO to classify actions of 3D human models

      2013, Engineering Applications of Artificial Intelligence
      Citation Excerpt :

      Inspired by MOCK, Saha and Bandyopadhyay (2009, 2010) have developed simulated annealing based multi-objective clustering algorithm that uses symmetry distance. The improved and generalized learning strategies for dynamically fast and statistically robust evolutionary algorithms are developed by Dashora et al. (2008). Recently Ma et al. (2009) have proposed an immunodominance and clonal selection inspired multi-objective clustering.

    • An investigation on evolutionary reconstruction of continuous chaotic systems

      2013, Mathematical and Computer Modelling
      Citation Excerpt :

      In [23,24], EAs have been successfully used for real-time chaos control and in [25,26] EAs were used for optimization of Chaos Control. Another examples of EA’s use can be found in [27] which developed statistically robust EAs, and on the opposite side [28] used EAs for fuzzy power system stabilizer, which has been applied on single-machine infinite bus system and multi-machine power system. Other research was done by [29].

    • Real-time deterministic chaos control by means of selected evolutionary techniques

      2009, Engineering Applications of Artificial Intelligence
    • Evolutionary identification of chaotic system

      2009, IFAC Proceedings Volumes (IFAC-PapersOnline)
    View all citing articles on Scopus
    View full text