Abstract
Analyzing the runtime of a Randomized Search Heuristic (RSH) by theoretical means often turns out to be rather tricky even for simple optimization problems. The main reason lies in the randomized nature of these algorithms. Both the optimization routines and the initialization of RSHs are typically based on random samples. It has often been observed, though, that the expected runtime of RSHs does not deviate much from the expected runtime when starting in an initial solution of average fitness. Having this information a priori could greatly simplify runtime analyses, as it reduces the necessity of analyzing the influence of the random initialization. Most runtime bounds in the literature, however, do not profit from this observation and are either too pessimistic or require a rather complex proof. In this work we show for the two search heuristics Randomized Local Search (RLS) and the (1+1) Evolutionary Algorithm on the two well-studied problems OneMax and LeadingOnes that their expected runtimes indeed deviate by at most a small additive constant from the expected runtime when started in a solution of average fitness. For RLS on OneMax, this additive discrepancy is \(-1/2 \pm o(1)\), leading to the first runtime statement for this problem that is precise apart from additive o(1) terms: The expected number of iterations until an optimal search point is found, is \(n H_{n/2} - 1/2 \pm o(1)\), where \(H_{n/2}\) denotes the (n / 2)th harmonic number when n is even, and \(H_{n/2}:= (H_{\lfloor n/2 \rfloor } + H_{\lceil n/2 \rceil })/2\) when n is odd. For this analysis and the one of the (1+1) EA optimizing OneMax we use a coupling technique, which we believe to be interesting also for the study of more complex optimization problems. For the analyses of the LeadingOnes test function, we show that one can express the discrepancy solely via the waiting times for leaving fitness level 0 and 1, which then easily yields the discrepancy, and in particular, without having to deal with so-called free-riders.
Similar content being viewed by others
Notes
Since there is some ambiguity with respect to the term “average fitness”, we provide the following formal definition. Let \(f:S\rightarrow {\mathbb {R}}\) for some finite, non-empty search space S. We call \(a:=\sum _{s\in S}{f(s)}/|S|\) the average fitness of f. A search point s with \(f(s)=a\) is thus said to be a search point of average fitness. When we say that we regard a random search point of average fitness, we implicitly mean to take a uniformly random search point from the set \(A:=\{s \mid f(s)=a\}\) of all search points with average fitness. In many cases, A is empty, e.g., if the function values are all integer, but the average fitness is not. We therefore need to regard, in our formal definition, the two fitness levels \(a_{-}:=\min \{f(s) \mid s \in S, f(s)\le a\}\) and \(a_{+}:=\min \{f(s) \mid s \in S, f(s) \ge a\}\) of the largest fitness value smaller or equal to a and the smallest fitness value larger or equal to a, respectively. Let \(A_{-}\) and \(A_{+}\) be the respective sets of search points of fitness \(a_{-}\) and \(a_{+}\). A random search point of average fitness is a search point taken uniformly at random from the union \(A_{-} \cup A_{+}\). Note that for OneMax and even n, by Remark 2 and Lemma 1, our claims remain true if, instead of taking a random search point of average fitness, we take an arbitrary search point of fitness n / 2 as initial solution.
Note that Eq. (1) also holds for our extension of the definition of the harmonic numbers to half-integral numbers. By our definition, \(H_{n + \frac{1}{2}}:=(H_n+H_{n+1})/2 = H_n + \frac{1}{2} \cdot \frac{1}{n+1}\) and a routine calculation shows that this is \(\ln (n+\frac{1}{2}) + \gamma + \frac{1}{2(n+\frac{1}{2})} \pm \varTheta ((n+\frac{1}{2})^{-2})\).
In the proof of Theorem 1 we had shown that for RLS the corresponding difference is roughly \(4k^2/n\). The bound for the \((1 + 1)\) EA is thus slightly weaker than that for RLS in that we do not compute the constant terms.
References
Borisovsky, P.A., Eremeev, A.V.: Comparing evolutionary algorithms to the (1+1)-EA. Theor. Comput. Sci. 403, 33–41 (2008)
Böttcher, S., Doerr, B., Neumann, F.: Optimal fixed and adaptive mutation rates for the LeadingOnes problem. In: Proceedings of the 11th International Conference on Parallel Problem Solving from Nature (PPSN’10), pp. 1–10. Springer (2010)
Doerr, B.: Analyzing randomized search heuristics: tools from probability theory. In: Auger, A., Doerr, B. (eds.) Theory of Randomized Search Heuristics, pp. 1–20. World Scientific Publishing, New York (2011)
Doerr, B., Doerr, C.: The impact of random initialization on the runtime of randomized search heuristics. In: Proceedings of the 16th Annual Genetic and Evolutionary Computation Conference (GECCO’14), pp. 1375–1382. ACM (2014)
Doerr, B., Doerr, C., Ebel, F.: Lessons from the black-box: Fast crossover-based genetic algorithms. In: Proceedings of the 15th Annual Genetic and Evolutionary Computation Conference (GECCO’13), pp. 781–788. ACM (2013)
Doerr, B., Fouz, M., Witt, C.: Quasirandom evolutionary algorithms. In: Proceedings of the 12th Annual Genetic and Evolutionary Computation Conference (GECCO’10), pp. 1457–1464. ACM (2010)
Doerr, B., Johannsen, D., Winzen, C.: Multiplicative drift analysis. Algorithmica 64, 673–697 (2012)
Doerr, B., Künnemann, M.: How the (1+\(\lambda \)) evolutionary algorithm optimizes linear functions. In: Proceedings of the 15th Annual Genetic and Evolutionary Computation Conference (GECCO’13), pp. 1589–1596. ACM (2013)
Droste, S., Jansen, T., Wegener, I.: A rigorous complexity analysis of the (1 + 1) evolutionary algorithm for separable functions with Boolean inputs. Evol. Comput. 6, 185–196 (1998)
Goldman, B.W., Punch, W.F.: Parameter-less population pyramid. In: Proceedings of the 2014 Conference on Genetic and Evolutionary Computation (GECCO’14), pp. 785–792. ACM (2014)
He, J., Yao, X.: Drift analysis and average time complexity of evolutionary algorithms. Artifi. Intell. 127, 57–85 (2001)
Hwang, H., Panholzer, A., Rolin, N., Tsai, T., Chen, W.: Probabilistic analysis of the (1+1)-evolutionary algorithm. CoRR, abs/1409.4955 (2014)
Ladret, V.: Asymptotic hitting time for a simple evolutionary model of protein folding. J. Appl. Prob. 42, 39–51 (2005)
Lehre, P.K., Witt, C.: Black-box search by unbiased variation. Algorithmica 64, 623–642 (2012)
Lindvall, T.: Lectures on the Coupling Method. John Wiley, New York (1992)
Mitzenmacher, M., Upfal, E.: Probability and Computing: Randomized Algorithms and Probabilistic Analysis. Cambridge University Press, Cambridge (2005)
Motwani, R., Raghavan, P.: Randomized Algorithms. Cambridge University Press, Cambridge (1995)
Oliveto, P.S., Yao, X.: Runtime analysis of evolutionary algorithms for discrete optimization. In: Auger, A., Doerr, B. (eds.) Theory of Randomized Search Heuristics, pp. 21–52. World Scientific Publishing, New York (2011)
Qian, C., Yu, Y., Zhou, Z.-H.: On algorithm-dependent boundary case identification for problem classes. In: Proceedings of the 12th International Conference on Parallel Problem Solving from Nature (PPSN’12), vol. 7491 of LNCS, pp. 62–71. Springer (2012)
Sedgewick, R.: The analysis of quicksort programs. Acta Inform. 7, 327–355 (1977)
Spivey, M.Z.: Combinatorial sums and finite differences. Discret. Math. 307, 3130–3146 (2007)
Sudholt, D.: A new method for lower bounds on the running time of evolutionary algorithms. IEEE Trans. Evol. Comput. 17, 418–435 (2013)
Thorisson, H.: Coupling, Stationarity, and Regeneration. Springer, Berlin (2000)
Wegener, I.: Methods for the analysis of evolutionary algorithms on pseudo-Boolean functions. In: Sarker, R., Yao, X., Mohammadian, M. (eds.) Evolutionary Optimization, pp. 349–369. Kluwer, Alphen aan den Rijn (2002)
Witt, C.: Tight bounds on the optimization time of a randomized search heuristic on linear functions. Comb. Probab. Comput. 22, 294–318 (2013)
Witt, C.: Fitness levels with tail bounds for the analysis of randomized search heuristics. Inf. Process. Lett. 114, 38–41 (2014)
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Doerr, B., Doerr, C. The Impact of Random Initialization on the Runtime of Randomized Search Heuristics. Algorithmica 75, 529–553 (2016). https://doi.org/10.1007/s00453-015-0019-5
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00453-015-0019-5