ES-2 Flashcards

1
Q

Covariance Matrix Adaptation

A
  • continuous optimization of non-linear, non-convex functions
  • handle ill-conditioned problems such as discontinuities, (sharp) ridges, or local optima
  • on convex-quadratic objective functions, setting the covariance matrix of the search distribution to the inverse Hessian matrix is equivalent to rescaling the ellipsoid function into a spherical one
  • increase the likelihood of repeating successful steps
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the objective of CMA ?

A

The final objective of covariance matrix adaptation is to closely approximate the contour lines of the objective function f

Increase the likelihood to repeat successful steps.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What method is similar to CMA ?

A

Learning the covariance matrix in the CMA-ES is similar to learning the inverse Hessian matrix in a quasi-Newton method.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

How offspring are generated in CMA-ES

A

offspring is generated by sampling a multivariate normal distribution

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

CMA-ES with convex-quadratic (ellipsoid) objective function.

A

-any convex-quadratic (ellipsoid) objective function is transformed into the spherical function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what is the purpose to Combining Rank-μ-Update and Cumulation

A

handle various of population seize.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Influence of population size in CMA-ES ?

A
  • Small population sizes usually lead to faster convergence,
  • large population sizes help to avoid local optima.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Invariance property of cam es

A
  • translation,
  • rotation, and
  • scaling of the coordinates

-strictly monotonic transformations
of function values

  • Invariance to order preserving (i.e. strictly monotonic) transformations of the objective function value. The algorithm only depends on the ranking of function values.
  • Invariance to angle preserving (rigid) transformations of the search space (rotation, re- flection, and translation) if the initial search point is transformed accordingly.
  • Scale invariance if the initial scaling, e.g. σ(0), and the initial search point, m(0), are chosen accordingly.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

(1+1)-CMA-ES

A
  • use rank-1 update
  • update the C using AA^T
  • 1+1 cma-es on convex quadratic functions, O(n2) iterations are required for the covariance matrix to approach the inverse Hessian
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

(μ/μ, λ)-CMA-ES

A
  • use rank-μ update

- using eigenvalue decompositions to obtain B and D such that C=AA^T where A=BDB^T

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

what are the two variants of restart CMA-ES

A

-IPOP-CMA-ES:
detect stagnation
restart with doubled population size

-BIPOP-CMA-ES interlaces two restart strategies:
one strategy uses an increasing population size, the other one a small population size
both have the same computational budget

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

what is differential evolution ?

A

differential evolution is a newer heuristic that uses differences between points in the population as mutation vectors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

what is Particle Swarm Optimization ?

A

candidate solutions are referred to as particles, the population as a swarm
each particle has a velocity; previous best positions bi are stored for each particle i

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Disadvantage of DE and PSO ?

A

like DE, PSO is not rotationally invariant

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

what is Information Geometric Optimization ?

A
  • optimization in a space of probability distributions choice of a statistical model
  • stochastic relaxation of the objective
  • stochastic natural gradient descent, where the gradient estimate is obtained through Monte Carlo estimation
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

relation of CMA-ES and IGO ?

A

the CMA-ES is in essence an IGO algorithm (plus rank-based utility weights, lowpass filtering to reduce noise, separate step size adaptation, . . . )

17
Q

A B C D in CMA ES

A

A is symmetric AA(inv) is I

B is an orthogonal and normal matrix

D is a diagonal matrix

C is the covariance matrix BD can obtain from wife value decomposition of C