Evolution strategy

From testwiki
Revision as of 04:35, 9 February 2025 by imported>Citation bot (Alter: title, template type, issue. Add: pmid, chapter, bibcode. Removed URL that duplicated identifier. Removed parameters. Formatted dashes. | Use this bot. Report bugs. | Suggested by Headbomb | Linked from Wikipedia:WikiProject_Academic_Journals/Journals_cited_by_Wikipedia/Sandbox | #UCB_webform_linked 256/1156)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

Template:Evolutionary algorithms Template:For Evolution strategy (ES) from computer science is a subclass of evolutionary algorithms, which serves as an optimization technique.[1] It uses the major genetic operators mutation, recombination and selection of parents.[2]

History

The 'evolution strategy' optimization technique was created in the early 1960s and developed further in the 1970s and later by Ingo Rechenberg, Hans-Paul Schwefel and their co-workers.[1]

Timeline of ES - selected algorithms[1]
Year Description Reference
1973 ES introduced with mutation and selection [3]
1994 Derandomized self-adaptation ES - Derandomized scheme of mutative step size control is used [4]
1994 CSA-ES - usage information from the old generations [5]
2001 CMA-ES [6]
2006 Weighted multi-recombination ES - usage of weighted recombination [7]
2007 Meta-ES - incremental aggregation of partial semantic structures [8]
2008 Natural ES - usage of natural gradient [9]
2010 Exponential natural ES - a simpler version of natural ES [10]
2014 Limited memory CMA-ES - time–memory complexity reduction by covariance matrix decomposition [11]
2016 Fitness inheritance CMA-ES - fitness evaluation computational cost reduction using fitness inheritance [12]
2017 RS-CMSA ES - usage of subpopulations [13]
2017 MA-ES - COV update and COV matrix square root are not used [14]
2018 Weighted ES - weighted recombination of general convex quadratic functions [15]

Methods

Evolution strategies use natural problem-dependent representations, so problem space and search space are identical. In common with evolutionary algorithms, the operators are applied in a loop. An iteration of the loop is called a generation. The sequence of generations is continued until a termination criterion is met.

The special feature of the ES is the self-adaptation of mutation step sizes and the coevolution associated with it. The ES is briefly presented using the standard form,[16][17][18] pointing out that there are many variants.[19][20][21][22] The real-valued chromosome contains, in addition to the n decision variables, n mutation step sizes σj, where: 1jnn. Often one mutation step size is used for all decision variables or each has its own step size. Mate selection to produce λ offspring is random, i.e. independent of fitness. First, new mutation step sizes are generated per mating by intermediate recombination of the parental σj with subsequent mutation as follows:

σ'j=σje(𝒩(0,1)𝒩j(0,1))

where 𝒩(0,1) is a normally distributed random variable with mean 0 and standard deviation 1. 𝒩(0,1) applies to all σ'j, while 𝒩j(0,1) is newly determined for each σ'j. Next, discrete recombination of the decision variables is followed by a mutation using the new mutation step sizes as standard deviations of the normal distribution. The new decision variables xj are calculated as follows:

xj=xj+𝒩j(0,σj)

This results in an evolutionary search on two levels: First, at the problem level itself and second, at the mutation step size level. In this way, it can be ensured that the ES searches for its target in ever finer steps. However, there is also the danger of being able to skip larger invalid areas in the search space only with difficulty.

Variants

The ES knows two variants of best selection for the generation of the next parent population (μ - number of parents, λ - number of offspring):[2]

  • (μ,λ): The μ best offspring are used for the next generation (usually μ=λ2).
  • (μ+λ): The best are selected from a union of μ parents and λ offspring.

Bäck and Schwefel recommend that the value of λ should be approximately seven times the μ,[17] whereby μ must not be chosen too small because of the strong selection pressure. Suitable values for μ are application-dependent and must be determined experimentally. The selection of the next generation in evolution strategies is deterministic and only based on the fitness rankings, not on the actual fitness values. The resulting algorithm is therefore invariant with respect to monotonic transformations of the objective function.

The simplest and oldest[1] evolution strategy (1+1) operates on a population of size two: the current point (parent) and the result of its mutation. Only if the mutant's fitness is at least as good as the parent one, it becomes the parent of the next generation. Otherwise the mutant is disregarded. More generally, λ mutants can be generated and compete with the parent, called (1+λ). In (1,λ) the best mutant becomes the parent of the next generation while the current parent is always disregarded. For some of these variants, proofs of linear convergence (in a stochastic sense) have been derived on unimodal objective functions.[23][24]

Individual step sizes for each coordinate, or correlations between coordinates, which are essentially defined by an underlying covariance matrix, are controlled in practice either by self-adaptation or by covariance matrix adaptation (CMA-ES).[21] When the mutation step is drawn from a multivariate normal distribution using an evolving covariance matrix, it has been hypothesized that this adapted matrix approximates the inverse Hessian of the search landscape. This hypothesis has been proven for a static model relying on a quadratic approximation.[25]

See also

References

Template:Reflist

Bibliography

Research centers

Template:Evolutionary computation