Improved and generalized learning strategies for dynamically fast and statistically robust evolutionary algorithms
Introduction
Since its advent, evolutionary optimization has gone through various stages of evolution that have made it more robust and fast; robust enough to be applied to all spheres of optimization problems and fast enough to take decisions in micro-seconds (Dimopoulos and Zalzala, 2000; Nissen and Propach, 1998; Yao et al., 1999). In the present scenario, evolutionary algorithms (EAs) are not limited to the applications in artificial intelligence, and have been increasingly used in various dimensions to solve real world problems (Dimopoulos and Zalzala, 2000; Sinha et al., 2003). Although, a plethora of literature pertaining to the search strategies adopted by various EAs is available (Nissen and Propach, 1998; Rana et al., 1996; Kazarlis et al., 2001; Yao et al., 1999; Salomon, 1998; Choi and Oh, 2000; Yoon and Moon, 2002; Kim and Myung, 1997; Storn, 1999), most of them restrict themselves to solving a particular class of optimization problem and thus, fail to provide generalized strategies that can be robustly used for wide spectrum of optimization problems in science, business and engineering applications. In general, optimization problems can be classified into two groups—numerical optimization and combinatorial optimization (Tsai et al., 2004; Gen and Cheng, 1999). Yet another classification exists that is based on the representation schemes—binary string, floating point string and integer bit string representation (Rana et al., 1996; Kazarlis et al., 2001; Yao et al., 1999; Salomon, 1998; Choi and Oh, 2000; Yoon and Moon, 2002; Kim and Myung, 1997; Storn, 1999; Zhang and Leung, 1999; Burke and Newall, 1999). In general, the first two representations are more suited and easy to implement on numerical optimization problems; whereas, the third representation enjoys solving the combinatorial optimization problems with a greater ease and efficiency (Nijssen and Bäck, 2003; Goldberg, 1989). Thus, to propose effective algorithms that would efficiently work in all the representations is a convoluted task. The schematic representation of various optimization problems along with their most suited representation classes are shown in Fig. 1.
All of the above-mentioned factors motivated the authors to propose robust and fast EAs with improved learning strategies to achieve the global optimization in all of these cases. Before introducing the proposed algorithms, the general formulation of a numerical optimization problem can be mathematically described aswhere t∈Γ∩X,Γ represents the ‘feasible region’ defined aswhere are the ‘constraints’ on the problem; the n-dimensional ‘search space’ is defined by X⊆Λn and is constrained by the ‘parametric constraints’ given below:where and represent the lower and upper limits of the variable ti. Here, it is evident that the solution is represented as floating point string of type A (Fig. 1); however, it can also be represented in the binary string (type B of Fig. 1) by considering the representation of floating numbers into binary numbers.
For the second class of optimization problems, i.e. combinatorial optimization problems, the general formulation can be given aswhere S is the set of integers; denotes the ‘feasible range’ of variables and is given bywhere c′j(s) represents the ‘constraints’ of the problem; the ‘search range’ is defined by , which is constrained by the ‘individual range’ defined bywhere and are the lower and upper limits of the variable c′j. Generally, this formulation is utilized for the type D representation. However, the representation type C is used in the combinatorial problems class mapped by traveling salesman problem (TSP). Its general formulation is described aswheresubject to the constraints cj″, j={1, 2, …, n}, where N is the number of nodes; distij is the distance between nodes i and j; Here, only minimization cases are considered as the problems pertaining to maximization of objective can also be mapped into minimization problem by making suitable changes in the formulation.
Owing to the existence of all these formulations, authors intend to present some improved learning rules suited to each formulation in order to eliminate the difficulty and dilemma of the practitioners and users of EAs while searching the best strategy for their problem. In addition to this, various testing phases based on design of experiments are introduced in order to achieve robust parameter control and tuning.
In the earlier studies, various modifications to the initial versions of EAs have been proposed to suit the varying complexity of the optimization problems. Some of the vital factors that affect the search strategy of the EAs are: (1) population representation, (2) information sharing among members of the population, (3) mutation of the current population, (4) generation scheme for new population, (5) learning from previous generations (Kazarlis et al., 2001; Yao et al., 1999; Salomon, 1998; Choi and Oh, 2000; Yoon and Moon, 2002; Kim and Myung, 1997; Storn, 1999; Zhang and Leung, 1999; Burke and Newall, 1999; Eiben et al., 1999; Harik et al., 1999; Michalewicz et al., 2000; Runarsson and Yao, 2000; François and Lavergne, 2001; Kazarlis et al., 2001; Ong and Keane, 2004). Even a small variation in any of these factors has a considerable effect on the EA performance.
The following discussion provides an insight to some of the recent developments in the field of evolutionary computation. In order to do away the slow convergence of the classical evolutionary programming (CEP), Yao et al. (1999) introduced fast evolutionary programming (FEP) that utilized the mutation based on Cauchy function because of its higher probability of making longer jumps Runarsson and Yao (2000) introduced a novel approach to balance objective and penalty functions stochastically, i.e. ‘stochastic ranking’ and presented a fresh view on penalty function methods in terms of dominance of penalty and objective functions. Salomon (1998) presented a new hybrid approach, the evolutionary-gradient-search method for the problems pertaining to numerical optimization. A much broader collection of developments in evolutionary computation for manufacturing optimization can be found in Dimopoulos and Zalzala (2000). Zhang and Leung (1999) proposed an orthogonal genetic algorithm for multimedia cast routing that incorporated orthogonal design to crossover operation. Francois (1998) presented a mutation or selection evolutionary strategy (MOSES) to solve complex discrete optimization problems. It theoretically studied the relationship between convergence, parameters of the method, and the geometry of the optimization problem. An extensive empirical study on the synergy among multiple crossover operators has been provided in Yoon and Moon (2002).
Choi and Oh (2000) presented a new mutation rule for evolutionary programming (EP) that has been motivated from the back-propagation learning rule of the neural networks. A method of decomposing larger problems into smaller components, each of which is of a size that the EA can effectively handle, was proposed by Burke and Newall (1999) to solve timetabling problem. Eiben et al. (1999) provided a broader classification for parameter control mechanisms of EAs. They also provided a survey of various types of control that have been used by evolutionary computation community. A statistical method was described by François and Lavergne (2001), which was able to find good parameter settings for EAs. The method utilized a functional relationship between the algorithms performance and its parameter values. Yet another statistical approach can be found in Czarn et al. (2004).
Ahn and Ramakrishna (2003) described two elitism-based compact genetic algorithms in order to design efficient compact type GAs by treating them as the estimation of distributed algorithms (EDAs) to efficiently solve complex optimization problems. The compact GAs utilized the concept of successive learning in generations utilizing the probability strings for the solution representation. Various other learning approaches also persist in the literature. In general, learning algorithms are termed as learning automata approaches (Najim and Poznyak, 1994). In Agache and Oommen (2002), two new generalized pursuit algorithms were presented that falls in the category of fastest learning automata algorithms. Howell et al. (2002) proposed a hybrid genetic and learning automata approach that enjoys the merits of both the strategies. Various other relevant work like Papadimitriou (1994), Najim and Poznyak (1994) in the context of learning algorithms need to be mentioned here.
Many works have been reported in literature in order to map the system chaos and noise within the working dimensions of EAs (Nissen and Propach, 1998; Rana et al., 1996; Caponetto et al., 2003). Caponetto et al. (2003) introduced chaotic sequences instead of random sequences during all the phases of random evolution process. The approach was based on the spread spectrum characteristic of chaotic sequences. In Nissen and Propach (1998), noise function was introduced to the deterministic objective function values in order to map the practical scenario of the existence of noise while experimentation, stochastic simulation, sampling and even interaction with users.
As it can easily be manifested from the above literature review that though a wide range of modifications have been carried out to enhance the performance of EAs, a robust strategy is mandatory that can work in general for various formulations of objective functions found in the real world scenario. Also, an all purpose strategy for robust parameter control and tuning in the presence of noise is essential to help the practitioners of EAs to come out of the dilemma of choosing the control and tuning methods in changing situations.
The aims of this paper are manifold. First, it introduces the concept of developing general-purpose fast EAs that equivalently achieve better results in all the previously mentioned formulations (i.e. A, B, C, and D). Second, it utilizes the enhanced learning mechanisms for all these formulations in order to achieve better tradeoff between exploration and exploitation of the search space; and to achieve time gain over prevailing strategies. This has also been theoretically validated. Third, it introduces a new tradeoff function to decide the mutation step size. This function makes a tradeoff between the fast convergence of Cauchy function and the explorative search of Gauss function to achieve better solutions. Fourth, the chaotic sequences to map the random dynamics have also been utilized to mutate the solutions. Thus, the evolutionary strategies utilized in the paper, the mutation at each step is performed both by the tradeoff function and the chaotic sequence generator and the better offspring is evolved further. Fifth, the paper proposes a design of experiments model to conduct various experiments sequentially utilizing the orthogonal array that symbolizes the variation in various parameter values. This approach reduces the tedious task of parameter control and tuning that were earlier carried out by exhaustive experiments. Comparisons with the existing EAs are also carried out that utilize the well-designed experiments, and thus the efficacy of the strategies has been established. Finally, the noise factor has been incorporated in the deterministic objective function value, and designed experiments are conducted to establish the robustness of the algorithm.
The rest of the paper is organized in the following sequence: The next section introduces the main intricacies in defining the general-purpose EAs. Section 3 introduces new generalized learning strategies and presents the new guided mutation rule. It also provides the general-purpose algorithms for all the four problem types. Section 4 deals with the mathematical aspects of the proposed algorithms and establishes their convergence. Section 5 provides the classification of the test functions utilized to evaluate the effectiveness of the proposed algorithm for numerical optimization problems on the basis of their dimensional complexity and degree of entrapment. It also presents the test problems of combinatorial optimization utilized for the comparative study. Section 6 formulates the design of experiments for the parameter tuning and the analytical study of the proposed algorithm. The results of the extensive computational experiments, inferences drawn from it and related discussions are presented in Section 7. Section 8 concludes the paper.
Section snippets
An insight to intricacies in defining general-purpose algorithms
As discussed in the previous section, a general optimization problem can be conveniently framed into one of the formulations depicted in Fig. 1, though, with slight changes. The major obstacle in the presentation of robust algorithms that can equivalently work in any of the varying environment or formulation is the peculiar data type handling method used in it.
Due to varying representation schemes, the general rules for mutation cannot be implemented in all the cases. Hence, the following
The generalized dynamically fast EAs
This paper utilizes the following adaptations and modifications in the existing strategies
- •
Learning strategy utilizing the neighborhood information and improved updating rule.
- •
Guided adaptive mutation rule based on the better of the step obtained from tradeoff function between Cauchy and Gaussian distributions, and from the random chaotic distributions.
Based on the aforementioned aspects, the general rules for all the formulations are given in the following discussion.
Can the learning rule converge?
To show the convergence, the learning strategies are split into two parts—first, the learning strategies which utilize Eq. (13) and second, those utilizing Eq. (17). Let the case of first type be taken. The convergence of the strategy can be proved in two steps. First, the following theorem is established. Theorem 1 For some ℸ >0 and И<∞, ∃ς*>0, υ*>0 and ginitial<∞ such that under Ls, probability ρ (all actions are chosen at least И times before g) >1−ℸ, ∀g⩾ginitial. Proof This theorem is
Test functions pertaining to type A and type B
To evaluate the performance of the proposed EAs and carry out the designed experiments for the numerical optimization problems (A and B), various benchmark test functions available in the literature have been utilized (Yao et al., 1999). Broadly the following classification of test functions has been utilized:
- (1)
Unimodal functions, f1–f3 (i.e. with no local minima).
- (2)
Multimodal functions with many local minima, f4–f5.
- (3)
Multimodal functions with few local minima, f6–f7.
The functions and their
Experimental design for parameter tuning
It has been an issue of challenge in the EA optimization field to determine the parameter setting that yield efficient performance on the problems at hand. Recently, no free lunch theorem (Wolpert and Macready, 1997) has revealed that the average performance of about all pairs of algorithms is approximately same. This means that if any algorithm performs better in some particular set of problems, it is bound to perform worse in other problems. Also, if an algorithm works well on a particular
Computational results and inferences
In order to start the computational analysis, first the best performing set of parameters has been obtained according to the previously mentioned procedure. The following subsection analyzes various results obtained for parameter tuning.
Concluding remarks and future scope
This paper targets the real-time optimization from a distinctive perspective. The generalized optimization problems have been first characterized as four basic type problems on the basis of varying solution representations utilized. Further, some generalized and enhanced learning rules have been defined for the EAs. In addition, a new tradeoff function-based mutation has been introduced that ensures more explorative and thorough search. Based on these modifications in the general EAs, four new
References (45)
- et al.
Algorithm for solving large scale 0–1 goal programming and its applications to reliability optimization problem
International Journal of Computers and Industrial Engineering
(1989) - et al.
Generalized pursuit learning schemes: new families of continuous and discretized learning automata
IEEE Transactions on Systems, Man and Cybernetics—Part B
(2002) - et al.
Elitism based compact genetic algorithms
IEEE Transactions on Evolutionary Computation
(2003) - et al.
A multistage evolutionary algorithm for the timetable problem
IEEE Transactions on Evolutionary Computation
(1999) - et al.
Toward a theory of generalization and learning in XCS
IEEE Transactions on Evolutionary Computation
(2004) - et al.
Chaotic sequences to improve the performance of evolutionary algorithms
IEEE Transactions on Evolutionary Computation
(2003) - et al.
A new mutation rule for evolutionary programming motivated from backpropogation learning
IEEE Transactions on Evolutionary Computation
(2000) - et al.
The particle swarm—explosion, stability, and convergence in a multidimensional complex space
IEEE Transactions on Evolutionary Computation
(2002) - et al.
Statistical exploratory analysis of genetic algorithms
IEEE Transactions on Evolutionary Computation
(2004)
Recent developments in evolutionary computation for manufacturing optimization: problems, solutions and comparisons
IEEE Transactions on Evolutionary Computation
Parameter control in evolutionary algorithms
IEEE Transactions on Evolutionary Computation
An evolutionary strategy for global minimization and its Markov chain analysis
IEEE Transactions on Evolutionary Computation
Design of evolutionary algorithms—a statistical perspective
IEEE Transactions on Evolutionary Computation
Reliability optimization by 0–1 programming for a system with several failure modes
IEEE Transactions on Reliability
Genetic Algorithms
Genetic Algorithms in Search, Optimization and Machine Learning
Evolution strategies on noisy functions. How to improve convergence properties,
The compact genetic algorithm
IEEE Transactions on Evolutionary Computation
Genetic learning automata for function optimization
IEEE Transactions on System, Man, and Cybernatics—Part B
A parallel genetic/neural network learning algorithm for MIMD shared memory machines
IEEE Transactions on Neural Networks
Calculus with Analytic Geometry
Cited by (12)
Automatic clustering algorithm based on multi-objective Immunized PSO to classify actions of 3D human models
2013, Engineering Applications of Artificial IntelligenceCitation Excerpt :Inspired by MOCK, Saha and Bandyopadhyay (2009, 2010) have developed simulated annealing based multi-objective clustering algorithm that uses symmetry distance. The improved and generalized learning strategies for dynamically fast and statistically robust evolutionary algorithms are developed by Dashora et al. (2008). Recently Ma et al. (2009) have proposed an immunodominance and clonal selection inspired multi-objective clustering.
An investigation on evolutionary reconstruction of continuous chaotic systems
2013, Mathematical and Computer ModellingCitation Excerpt :In [23,24], EAs have been successfully used for real-time chaos control and in [25,26] EAs were used for optimization of Chaos Control. Another examples of EA’s use can be found in [27] which developed statistically robust EAs, and on the opposite side [28] used EAs for fuzzy power system stabilizer, which has been applied on single-machine infinite bus system and multi-machine power system. Other research was done by [29].
Real-time deterministic chaos control by means of selected evolutionary techniques
2009, Engineering Applications of Artificial IntelligenceEvolutionary identification of chaotic system
2009, IFAC Proceedings Volumes (IFAC-PapersOnline)On mutual relations amongst evolutionary algorithm dynamics and its hidden complex network structures: An overview and recent advances
2016, Nature-Inspired Computing: Concepts, Methodologies, Tools, and ApplicationsA utility-driven approach to supplier evaluation and selection: Empirical validation of an integrated solution framework
2016, International Journal of Production Research