1 / 39

Differential Evolution

Differential Evolution. Hossein Talebi Hassan nikoo. Variations to Basic Differential Evolution. Hybrid Differential Evolution Strategies Gradient-Based Hybrid Differential Evolution Evolutionary Algorithm-Based Hybrids DE Reproduction Process used as Cross over Operation of Simple GA

glenna-hunt
Download Presentation

Differential Evolution

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Differential Evolution HosseinTalebi Hassan nikoo

  2. Variations to Basic Differential Evolution • Hybrid Differential Evolution Strategies • Gradient-Based Hybrid Differential Evolution • Evolutionary Algorithm-Based Hybrids • DE Reproduction Process used as Cross over Operation of Simple GA • Ranked-Based Cross Over Operator for DE • Particle Swarm Optimization Hybrids • Population-Based Differential Evolution • Self-Adaptive Differential Evolution

  3. Variations to Basic Differential Evolution • Differential Evolution for Discrete-Valued Problems • Angle Modulated Differential Evolution • Binary Differential Evolution • Constraint Handling Approaches • Multi-Objective Optimization • Dynamic Environments • Applications

  4. Gradient-Based Hybrid Differential Evolution • acceleration operator to improve convergence speed – without decreasing diversity • migration operator • The Acceleration Operator uses gradient descent to adjust the best individual toward obtaining a better position

  5. Gradient-Based Hybrid Differential Evolution(Cont.) • Using Gradient Descent may results in getting stuck in a local optima or premature convergence • We can increase Population diversity by Migration Operator • This operator spawns new individuals from the best individual and replaces the current population with these new individuals

  6. Gradient-Based Hybrid Differential Evolution(Cont.) • The Migration Operator is applied only when diversity of population becomes too small

  7. Gradient-Based Hybrid Differential Evolution(Cont.)

  8. Using Stochastic Gradient Descent and DE for Neural Networks Training • Stochastic Gradient Descent

  9. Evolutionary Algorithm-Based Hybrids • Hrstka and Kucerov´a used the DE reproduction process as a crossover operator in a simple GA • Chang and Chang used standard mutation operators to increase DE population diversity by adding noise to the created trial vectors.

  10. Evolutionary Algorithm-Based Hybrids(Cont.) • Sarimveis and Nikolakopoulos [758] use rank-based selection to decide which individuals will take part to calculate difference vectors

  11. Ranked-base Mutation

  12. Particle Swarm Optimization Hybrids • Hendtlass proposed that the DE reproduction process be applied to the particles in a PSO swarm at specified intervals. • Kannanet al. apply DE to each particle for a number of iterations and replaces the best with particle

  13. Particle Swarm Optimization Hybrids(Cont.) • Another approach is to change only change best particle using • Where sigma is general difference vector

  14. Population-Based Differential Evolution • Ali and T¨orn proposed to use an auxiliary population • For each offspring created, if the fitness of the offspring is not better than the parent, instead of discarding the offspring, it is considered for inclusion in the auxiliary Population

  15. DETVSF (DE with Time Varying Scale Factor) • During the later stages it is important to adjust the movements of trial solutions finely so that they can explore the interior of a relatively small space in which the suspected global optimum lies • We can reduce the scale factor linearly with time from a (predetermined) maximum to a (predetermined) minimum value

  16. DETVSF (DE with Time Varying Scale Factor)(Cont.)

  17. Parameter Control in DE • Dynamic Parameters • Self-Adaptive

  18. Self-Adaptive Parameters • probability of recombination be self –adapted • Mu is the average of successful probablities • Abbass Proposed to use this formula :

  19. Self-Adaptive Parameters(Cont.) • Omranet al. propose a self-adaptive DE strategy that makes use of this formula for scale factor • For mutation operator

  20. Angle Modulated Differential Evolution • Pampar´aet al. proposed a DE algorithm to evolve solutions to binary-valued optimization problems, without having to change the operation of the original DE • They use a mapping between binary-valued and continuous-valued space to solve the problem in binary space

  21. Angle Modulated Differential Evolution(Cont.) • The objective is to evolve, in the abstracted continues space, a bitstring generating function will be used in the original space to produce bit-vector solutions • ‘a’, ’b’, ‘c’ and ‘d’ are continues space problem parameter

  22. Angle Modulated Differential Evolution(Cont.) • ‘a=0’ ‘b=1’ ‘c=1’ ‘d=0’

  23. Binary Differential Evolution • binDE borrows concepts from the binary particle swarm optimizer binPSO • binDE uses the floating-point DE individuals to determine a probability for each component • the corresponding bitstring solution will be calculated as follow :

  24. Binary Differential Evolution

  25. Constraint Handling Approaches • Penalty methods • adding a function to penalize solutions that violate constraints • Using F(x, t) = f(x, t) + λp(x, t) where λ is the penalty coefficient and p is time dependent penalty function • Converting the constrained problem to an unconstrained problem

  26. Constraint Handling Approaches(Cont.) • We can convert constrained problem to an unconstrained problem by defining the Lagrangian for the constrained problem • If primal problem is convex then defining dual problem and solving minmax problem

  27. Constraint Handling Approaches(Cont.) • By changing selection operator , infeasible solutions can be rejected and we can use a method for repairing of the infeasible solution

  28. Constraint Handling Approaches(Cont.) • Boundary constraints are easily enforced by clamping offspring to remain within the given boundaries

  29. Multi-Objective Optimization • Converting the problem into the • Weighted Aggregation Methods

  30. Multi-Objective Optimization(Cont.) • This method intends to define an aggregate objective function as a weighted sum of the objectives • Usually assumed that

  31. Multi-Objective Optimization(Cont.) • There is no guarantee that different solutions will be found • A niching strategy can be used to find multiple solutions • It is difficult to get the best weight values, ωk, since these are problem-dependent

  32. Multi-Objective Optimization(Cont.) • Vector evaluated DE is a population based method for MOO • If K objectives have to be optimized, K sub-populations are used, where each subpopulation optimizes one of the objectives. • Sub-populations are organized in a ring topology • The best individual of sub-population Ck migrates to population Ck+1 to produce the trial vectors for that population

  33. Dynamic Environments • Assumptions • the number of peaks, nX, to be found are know and these peaks are evenly distributed through the search space • Changes are small and gradual • DynDE uses multiple populations, with each population maintaining one of the peaks

  34. Dynamic Environments(Cont.) • At each iteration, the best individuals of each pair of sub-populations are compared if these global best positions are too close to one another, the sub-population with the worst global best solution is re-initialized

  35. Dynamic Environments(Cont.) • The following diversity increasing strategies • Re-initialize the sub-populations • Use quantum individuals :Some of the individuals are re-initialized to random points inside a ball centered at the global best individual • Use Brownian individuals: Some positions are initialized to random positions around global best individual • Some individuals are simply added noise

  36. Dynamic Environments(Cont.) • Initialization of Quantum Individuals

  37. Applications • Mostly applied to optimize functions defined over continuous-valued landscapes • Clustering • Controllers • Filter design • Image analysis • Integer-Programming • Model selection • NN training

  38. References • Computational Intelligence, an introduction,2nd edition, AndriesEngelbercht, Wiley • Differential Evolution - A simple and efficient adaptive scheme for global optimization over continuous spaces, Rainer Storn,Kenneth Price,1995 • Particle Swarm Optimization and Differential Evolution Algorithms: Technical Analysis, Applications and Hybridization Perspectives,Swagatam Das1, Ajith Abraham2, and AmitKonar1,Springer 2008. • Differential Evolution, homepagehttp://www.icsi.berkeley.edu/~storn/code.html

  39. Thanks For Your Attention Any Question?

More Related