SlideShare ist ein Scribd-Unternehmen logo
1 von 108
Downloaden Sie, um offline zu lesen
Probabilistic Model-Building
Genetic Algorithms
      a.k.a. Estimation of Distribution Algorithms
      a.k.a. Iterated Density Estimation Algorithms

Martin Pelikan

Missouri Estimation of Distribution Algorithms Laboratory (MEDAL)
Department of Mathematics and Computer Science
University of Missouri - St. Louis
martin@martinpelikan.net
                                                 Copyright is held by the author/owner(s).
http://martinpelikan.net/
                                                    GECCO’12 Companion, July 7–11, 2012, Philadelphia, PA, USA.
                                                    ACM 978-1-4503-1178-6/12/07.
Foreword
n    Motivation
      ¨    Genetic and evolutionary computation (GEC) popular.
      ¨    Toy problems great, but difficulties in practice.
      ¨    Must design new representations, operators, tune, …
n    This talk
      ¨    Discuss a promising direction in GEC.
      ¨    Combine machine learning and GEC.
      ¨    Create practical and powerful optimizers.




                           Martin Pelikan, Probabilistic Model-Building GAs
                                                                              2
Overview
n    Introduction
      ¨  Black-box   optimization via probabilistic modeling.
n    Probabilistic Model-Building GAs
      ¨  Discreterepresentation
      ¨  Continuous representation
      ¨  Computer programs (PMBGP)
      ¨  Permutations
n    Conclusions



                        Martin Pelikan, Probabilistic Model-Building GAs
                                                                           3
Problem Formulation
n    Input
      ¨  How do potential solutions look like?
      ¨  How to evaluate quality of potential solutions?

n    Output
      ¨  Best   solution (the optimum).
n    Important
      ¨  No   additional knowledge about the problem.




                        Martin Pelikan, Probabilistic Model-Building GAs
                                                                           4
Why View Problem as Black Box?
n    Advantages
      ¨  Separate problem definition from optimizer.
      ¨  Easy to solve new problems.
      ¨  Economy argument.

n    Difficulties
      ¨  Almost no prior problem knowledge.
      ¨  Problem specifics must be learned automatically.
      ¨  Noise, multiple objectives, interactive evaluation.



                      Martin Pelikan, Probabilistic Model-Building GAs
                                                                         5
Representations Considered Here
n    Start with
      ¨  Solutions   are n-bit binary strings.


n    Later
      ¨  Real-valuedvectors.
      ¨  Program trees.
      ¨  Permutations




                         Martin Pelikan, Probabilistic Model-Building GAs
                                                                            6
Typical Situation
n    Previously visited solutions + their evaluation:
                #	
 Solution                 Evaluation
                1	
 00100	
                      1	
                2	
 11011	
                      4	
                3	
 01101	
                      0	
                4	
 10111	
                      3	

n    Question: What solution to generate next?

                    Martin Pelikan, Probabilistic Model-Building GAs
                                                                       7
Many Answers
n    Hill climber
      ¨  Start with a random solution.
      ¨  Flip bit that improves the solution most.
      ¨  Finish when no more improvement possible.
n    Simulated annealing
      ¨  Introduce   Metropolis.
n    Probabilistic model-building GAs
      ¨  Inspiration   from GAs and machine learning (ML).



                         Martin Pelikan, Probabilistic Model-Building GAs
                                                                            8
Probabilistic Model-Building GAs
 Current      Selected                                                  New
population   population                                               population
  11001	
       11001	
                                                  01111	
  11101	
       10101	
                 Probabilistic                    11001	
  01011	
       01011	
                        Model                     11011	
  11000	
       11000	
                                                  00111	




  …replace crossover+mutation with learning
   and sampling probabilistic model

                   Martin Pelikan, Probabilistic Model-Building GAs
                                                                                   9
Other Names for PMBGAs
n    Estimation of distribution algorithms (EDAs)
      (Mühlenbein & Paass, 1996)

n    Iterated density estimation algorithms (IDEA)
      (Bosman & Thierens, 2000)




                   Martin Pelikan, Probabilistic Model-Building GAs
                                                                      10
Implicit vs. Explicit Model
n    GAs and PMBGAs perform similar task
      ¨  Generate  new solutions using probability
         distribution based on selected solutions.
n    GAs
      ¨  Variationdefines implicit probability distribution of
         target population given original population and
         variation operators (crossover and mutation).
n    PMBGAs
      ¨  Explicit
                 probabilistic model of selected candidate
         solutions is built and sampled.


                      Martin Pelikan, Probabilistic Model-Building GAs
                                                                         11
What Models to Use?
n    Start with a simple example
      ¨  Probability   vector for binary strings.


n    Later
      ¨  Dependency  tree models (COMIT).
      ¨  Bayesian networks (BOA).
      ¨  Bayesian networks with local structures (hBOA).




                         Martin Pelikan, Probabilistic Model-Building GAs
                                                                            12
Probability Vector
n  Assume n-bit binary strings.
n  Model: Probability vector p=(p1, …, pn)
      ¨  pi= probability of 1 in position i
      ¨  Learn p: Compute proportion of 1 in each position.
      ¨  Sample p: Sample 1 in position i with prob. pi




                     Martin Pelikan, Probabilistic Model-Building GAs
                                                                        13
Example: Probability Vector
(Mühlenbein, Paass, 1996), (Baluja, 1994)
 Current      Selected                                                  New
population   population                                               population
                                         Probability
  11001        11001                       vector                       10101
  10101        10101                                                    10001
                                    1.0 0.5 0.5 0.0 1.0
  01011        01011                                                    11101
  11000        11000                                                    11001




                   Martin Pelikan, Probabilistic Model-Building GAs
                                                                                   14
Probability Vector PMBGAs
n    PBIL (Baluja, 1995)
      ¨  Incremental   updates to the prob. vector.
n    Compact GA (Harik, Lobo, Goldberg, 1998)
      ¨  Also
            incremental updates but better analogy with
        populations.
n    UMDA (Mühlenbein, Paass, 1996)
      ¨  What   we showed here.
n  DEUM (Shakya et al., 2004)
n  All variants perform similarly.


                        Martin Pelikan, Probabilistic Model-Building GAs
                                                                           15
Probability Vector Dynamics
n  Bitsthat perform better get more copies.
n  And are combined in new ways.
n  But context of each bit is ignored.


n  Example   problem 1: Onemax
                                                              n
               f ( X 1 , X 2 ,… , X n ) = ∑ X i
                                                             i =1

                 Martin Pelikan, Probabilistic Model-Building GAs
                                                                    16
Probability Vector on Onemax
                                1
                               0.9
  Probability vector entries


                               0.8
                               0.7
                               0.6
                               0.5
                               0.4
                               0.3
                               0.2
                               0.1
                                0
                                 0   10           20             30             40       50
                                                 Generation
                                      Martin Pelikan, Probabilistic Model-Building GAs
                                                                                              17
Probability Vector: Ideal Scale-up
n    O(n log n) evaluations until convergence
      ¨  (Harik,
                Cantú-Paz, Goldberg, & Miller, 1997)
      ¨  (Mühlenbein, Schlierkamp-Vosen, 1993)

n    Other algorithms
      ¨  Hill
             climber: O(n log n) (Mühlenbein, 1992)
      ¨  GA with uniform: approx. O(n log n)
      ¨  GA with one-point: slightly slower




                    Martin Pelikan, Probabilistic Model-Building GAs
                                                                       18
When Does Prob. Vector Fail?
n    Example problem 2: Concatenated traps
      ¨  Partition
                  input string into disjoint groups of 5 bits.
      ¨  Groups contribute via trap (ones=number of ones):


                            " 5        if ones = 5
            trap ( ones ) = #
                            $ 4 − ones otherwise
      ¨  Concatenated trap = sum of single traps
      ¨  Optimum: String 111…1


                      Martin Pelikan, Probabilistic Model-Building GAs
                                                                         19
Trap-5
              5

              4
    trap(u)




              3

              2

              1

              0
                  0       1           2            3            4        5
                              Number of ones, u
                      Martin Pelikan, Probabilistic Model-Building GAs
                                                                             20
Probability Vector on Traps
                                 1
                                0.9
   Probability vector entries


                                0.8
                                0.7
                                0.6
                                0.5
                                0.4
                                0.3
                                0.2
                                0.1
                                 0
                                  0   10            20             30             40      50
                                                  Generation
                                       Martin Pelikan, Probabilistic Model-Building GAs
                                                                                               21
Why Failure?
n    Onemax:
      ¨  Optimum   in 111…1
      ¨  1 outperforms 0 on average.


n    Traps: optimum in 11111, but
         n  f(0****) = 2
         n  f(1****) = 1.375




n    So single bits are misleading.

                       Martin Pelikan, Probabilistic Model-Building GAs
                                                                          22
How to Fix It?
n  Consider 5-bit statistics instead 1-bit ones.
n  Then, 11111 would outperform 00000.
n  Learn model
      ¨  Compute   p(00000), p(00001), …, p(11111)
n    Sample model
      ¨  Sample 5 bits at a time
      ¨  Generate 00000 with p(00000),
          00001 with p(00001), …


                      Martin Pelikan, Probabilistic Model-Building GAs
                                                                         23
Correct Model on Traps: Dynamics
                            1
                           0.9
  Probabilities of 11111


                           0.8
                           0.7
                           0.6
                           0.5
                           0.4
                           0.3
                           0.2
                           0.1
                            0
                             0   10            20              30              40    50
                                             Generation                                   24
                                  Martin Pelikan, Probabilistic Model-Building GAs
Good News: Good Stats Work Great!
n  Optimum in O(n log n) evaluations.
n  Same performance as on onemax!
n  Others
      ¨  Hill
             climber: O(n5 log n) = much worse.
      ¨  GA with uniform: O(2n) = intractable.
      ¨  GA with k-point xover: O(2n) (w/o tight linkage).




                     Martin Pelikan, Probabilistic Model-Building GAs
                                                                        25
Challenge
n    If we could learn and use relevant context for
      each position
      ¨  Find non-misleading statistics.
      ¨  Use those statistics as in probability vector.

n    Then we could solve problems decomposable
      into statistics of order at most k with at most
      O(n2) evaluations!
      ¨  And    there are many such problems (Simon, 1968).



                        Martin Pelikan, Probabilistic Model-Building GAs
                                                                           26
What’s Next?
n    COMIT
      ¨  Use   tree models


n    Extended compact GA
      ¨  Cluster   bits into groups.


n    Bayesian optimization algorithm (BOA)
      ¨  Use   Bayesian networks (more general).



                         Martin Pelikan, Probabilistic Model-Building GAs
                                                                            27
Beyond single bits: COMIT
(Baluja, Davies, 1997)
                               Model                           P(X=1)
                                                                     75 %

    String
                                                               X       P(Y=1|X)
                                                               0	
           30 %
                                                               1	
           25 %

                                                           X          P(Z=1|X)
                                                           0	
              86 %
                                                           1                75 %
                  Martin Pelikan, Probabilistic Model-Building GAs
                                                                                    28
How to Learn a Tree Model?
n    Mutual information:
                                                                     P(Xi = a, X j = b)
      I(Xi , X j ) = ∑ P(Xi = a, X j = b) log
                    a,b                                           P(Xi = a)P(X j = b)
n    Goal
      ¨  Find tree that maximizes mutual information
          between connected nodes.
      ¨  Will minimize Kullback-Leibler divergence.
n    Algorithm
      ¨  Prim’s   algorithm for maximum spanning trees.


                          Martin Pelikan, Probabilistic Model-Building GAs
                                                                                          29
Prim’s Algorithm
n  Start with a graph with no edges.
n  Add arbitrary node to the tree.
n  Iterate
      ¨  Hang  a new node to the current tree.
      ¨  Prefer addition of edges with large mutual
          information (greedy approach).
n    Complexity: O(n2)



                     Martin Pelikan, Probabilistic Model-Building GAs
                                                                        30
Variants of PMBGAs with Tree Models
n    COMIT (Baluja, Davies, 1997)
      ¨  Tree   models.


n    MIMIC (DeBonet, 1996)
      ¨  Chain    distributions.


n    BMDA (Pelikan, Mühlenbein, 1998)
      ¨  Forest   distribution (independent trees or tree)


                          Martin Pelikan, Probabilistic Model-Building GAs
                                                                             31
Beyond Pairwise Dependencies: ECGA
n  Extended Compact GA (ECGA) (Harik, 1999).
n  Consider groups of string positions.

      String                                  Model



               00	
   16 %	
           0	
         86 %	
                 000	
    17 %	
               01	
   45 %	
           1	
         14 %	
                 001	
     2 %	
               10	
   35 %	
                                                 ···
               11	
    4 %	
                                              111	
    24 %	
                       Martin Pelikan, Probabilistic Model-Building GAs
                                                                                            32
Learning the Model in ECGA
n  Start with each bit in a separate group.
n  Each iteration merges two groups for best
    improvement.




                Martin Pelikan, Probabilistic Model-Building GAs
                                                                   33
How to Compute Model Quality?
n  ECGA uses minimum description length.
n  Minimize number of bits to store model+data:

                 MDL( M , D) = DModel + DData
n    Each frequency needs (0.5 log N) bits:
                                          |g|−1
                  DModel = ∑ 2                    log N
                                 g ∈G

n    Each solution X needs -log p(X) bits:
                DData = − N ∑ p( X ) log p( X )
                                    X
                    Martin Pelikan, Probabilistic Model-Building GAs
                                                                       34
Sampling Model in ECGA
n    Sample groups of bits at a time.

n    Based on observed probabilities/proportions.

n    But can also apply population-based crossover
      similar to uniform but w.r.t. model.




                   Martin Pelikan, Probabilistic Model-Building GAs
                                                                      35
Building-Block-Wise Mutation in ECGA
n    Sastry, Goldberg (2004); Lima et al. (2005)
n    Basic idea
      ¨    Use ECGA model builder to identify decomposition
      ¨    Use the best solution for BB-wise mutation
      ¨    For each k-bit partition (building block)
            n    Evaluate the remaining 2k-1 instantiations of this BB
            n    Use the best instantiation of this BB
n    Result (for order-k separable problems)
      ¨                                     (             )
            BB-wise mutation is O k log n times faster than ECGA!
      ¨    But only for separable problems (and similar ones).


                               Martin Pelikan, Probabilistic Model-Building GAs
                                                                                  36
What’s Next?
n    We saw
      ¨  Probabilityvector (no edges).
      ¨  Tree models (some edges).
      ¨  Marginal product models (groups of variables).


n    Next: Bayesian networks
      ¨  Can   represent all above and more.




                        Martin Pelikan, Probabilistic Model-Building GAs
                                                                           37
Bayesian Optimization Algorithm (BOA)
n  Pelikan, Goldberg, & Cantú-Paz (1998)
n  Use a Bayesian network (BN) as a model.
n  Bayesian network
      ¨  Acyclicdirected graph.
      ¨  Nodes are variables (string positions).
      ¨  Conditional dependencies (edges).
      ¨  Conditional independencies (implicit).




                      Martin Pelikan, Probabilistic Model-Building GAs
                                                                         38
Example: Bayesian Network (BN)
n  Conditional dependencies.
n  Conditional independencies.




                Martin Pelikan, Probabilistic Model-Building GAs
                                                                   39
BOA
                                        Bayesian
 Current      Selected                                             New
                                        network
population   population                                          population




              Martin Pelikan, Probabilistic Model-Building GAs
                                                                              40
Learning BNs

n    Two things again:
      ¨  Scoringmetric (as MDL in ECGA).
      ¨  Search procedure (in ECGA done by merging).




                    Martin Pelikan, Probabilistic Model-Building GAs
                                                                       41
Learning BNs: Scoring Metrics
n    Bayesian metrics
      ¨  Bayesian-Dirichlet         with likelihood equivallence
                         n
                             Γ(m'(π i ))           Γ(m'(xi , π i ) + m(xi , π i ))
        BD(B) = p(B)∏ ∏                        ∏ Γ(m'(x ,π ))
                    i=1 π Γ(m'(π i ) + m(π i )) xi                 i  i
                               i




n    Minimum description length metrics
      ¨  Bayesian   information criterion (BIC)
                                            Πi log 2 N &
                     n#
        BIC( B) = ∑ % − H ( X i | Πi )N − 2
                  i=1 $                           2 (  '

                         Martin Pelikan, Probabilistic Model-Building GAs
                                                                               42
Learning BNs: Search Procedure
n  Start with empty network (like ECGA).
n  Execute primitive operator that improves the
    metric the most (greedy).
n  Until no more improvement possible.
n  Primitive operators
      ¨  Edge addition (most important).
      ¨  Edge removal.
      ¨  Edge reversal.



                     Martin Pelikan, Probabilistic Model-Building GAs
                                                                        43
Learning BNs: Example




         Martin Pelikan, Probabilistic Model-Building GAs
                                                            44
BOA and Problem Decomposition
n  Conditions for factoring problem decomposition
    into a product of prior and conditional
    probabilities of small order in Mühlenbein,
    Mahnig, & Rodriguez (1999).
n  In practice, approximate factorization sufficient
    that can be learned automatically.
n  Learning makes complete theory intractable.




                Martin Pelikan, Probabilistic Model-Building GAs
                                                                   45
BOA Theory: Population Sizing
n    Initial supply (Goldberg et al., 2001)
      ¨    Have enough stuff to combine.                                     O 2k( )
n    Decision making (Harik et al, 1997)
      ¨    Decide well between competing partial sols.                       O   (   n log n   )
n    Drift (Thierens, Goldberg, Pereira, 1998)
      ¨    Don’t lose less salient stuff prematurely.                        O n ()
n    Model building (Pelikan et al., 2000, 2002)
      ¨    Find a good model.                                                    ( )
                                                                              O n1.05



                           Martin Pelikan, Probabilistic Model-Building GAs
                                                                                           46
BOA Theory: Num. of Generations
n  Two extreme cases, everything in the middle.
n  Uniform scaling
      ¨    Onemax model (Muehlenbein & Schlierkamp-Voosen, 1993)

                                       O    ( n)
n    Exponential scaling
      ¨    Domino convergence (Thierens, Goldberg, Pereira, 1998)

                                        O (n)
                          Martin Pelikan, Probabilistic Model-Building GAs
                                                                             47
Good News: Challenge Met!
n    Theory
      ¨    Population sizing (Pelikan et al., 2000, 2002)
            n    Initial supply.
            n    Decision making.                              O(n) to O(n1.05)
            n    Drift.
            n    Model building.
      ¨    Number of iterations (Pelikan et al., 2000, 2002)
            n  Uniform scaling.
            n  Exponential scaling.                            O(n0.5) to O(n)

n    BOA solves order-k decomposable problems in O(n1.55) to
      O(n2) evaluations!

                              Martin Pelikan, Probabilistic Model-Building GAs
                                                                                   48
Theory vs. Experiment (5-bit Traps)
                          500000	
                          450000	
                                             Experiment
                          400000	
           Theory
                          350000	
  Number of Evaluations




                          300000	

                          250000	

                          200000	


                          150000	




                          100000	

                                     100	
       125	
         150	
        175	
      200	
    225	
   250	
                                                          Problem Size                                          49
                                             Martin Pelikan, Probabilistic Model-Building GAs
BOA Siblings
n    Estimation of Bayesian Networks Algorithm
      (EBNA) (Etxeberria, Larrañaga, 1999).

n    Learning Factorized Distribution Algorithm
      (LFDA) (Mühlenbein, Mahnig, Rodriguez, 1999).




                   Martin Pelikan, Probabilistic Model-Building GAs
                                                                      50
Another Option: Markov Networks
n  MN-FDA, MN-EDA (Santana; 2003, 2005)
n  Similar to Bayes nets but with undirected edges.




                Martin Pelikan, Probabilistic Model-Building GAs
                                                                   51
Yet Another Option: Dependency Networks

n    Estimation of dependency networks algorithm (EDNA)
      ¨    Gamez, Mateo, Puerta (2007).
      ¨    Use dependency network as a model.
      ¨    Dependency network learned from pairwise interactions.
      ¨    Use Gibbs sampling to generate new solutions.
n    Dependency network
      ¨    Parents of a variable= all variables influencing this variable.
      ¨    Dependency network can contain cycles.




                           Martin Pelikan, Probabilistic Model-Building GAs
                                                                              52
Model Comparison
   BMDA                          ECGA                             BOA




      Model Expressiveness Increases




               Martin Pelikan, Probabilistic Model-Building GAs
                                                                        53
From single level to hierarchy
n  Single-level decomposition powerful.
n  But what if single-level decomposition is not
    enough?
n  Learn from humans and nature
      ¨  Decompose   problem over multiple levels.
      ¨  Use solutions from lower level as basic building
          blocks.
      ¨  Solve problem hierarchically.



                      Martin Pelikan, Probabilistic Model-Building GAs
                                                                         54
Hierarchical Decomposition
    	
                         Car


  Engine      Braking system                          Electrical system




Fuel system      Valves                     Ignition system


                 Martin Pelikan, Probabilistic Model-Building GAs
                                                                          55
Three Keys to Hierarchy Success
n    Proper decomposition
      ¨  Must   decompose problem on each level properly.
n    Chunking
      ¨  Must   represent & manipulate large order solutions.
n    Preservation of alternative solutions
      ¨  Must
             preserve alternative partial solutions
        (chunks).




                        Martin Pelikan, Probabilistic Model-Building GAs
                                                                           56
Hierarchical BOA (hBOA)
n  Pelikan & Goldberg (2000, 2001)
n  Proper decomposition
      ¨  Use   Bayesian networks like BOA.
n    Chunking
      ¨  Use   local structures in Bayesian networks.
n    Preservation of alternative solutions.
      ¨  Use restricted tournament replacement (RTR).
      ¨  Can use other niching methods.



                        Martin Pelikan, Probabilistic Model-Building GAs
                                                                           57
Local Structures in BNs
n    Look at one conditional dependency.
      ¨  2k   probabilities for k parents.
n    Why not use more powerful representations
      for conditional probabilities?
                                                X2X3            P(X1=0|X2X3)
                      X1
                                                  00	
                   26 %	
                                                  01	
                   44 %	
                 X2        X3
                                                  10	
                   15 %	
                                                  11	
                   15 %	

                            Martin Pelikan, Probabilistic Model-Building GAs
                                                                                  58
Local Structures in BNs
n    Look at one conditional dependency.
      ¨  2k   probabilities for k parents.
n    Why not use more powerful representations
      for conditional probabilities?
                                                              X2
                      X1                                0	
        1	

                                                       X3         15%	
                 X2        X3
                                                 0	
        1	

                                             26%	
      44%	

                            Martin Pelikan, Probabilistic Model-Building GAs
                                                                               59
Restricted Tournament Replacement
n  Used in hBOA for niching.
n  Insert each new candidate solution x like this:
      ¨  Pick random subset of original population.
      ¨  Find solution y most similar to x in the subset.
      ¨  Replace y by x if x is better than y.




                      Martin Pelikan, Probabilistic Model-Building GAs
                                                                         60
Hierarchical Traps: The Ultimate Test
n  Combine traps on more levels.
n  Each level contributes to fitness.
n  Groups of bits map to next level.




                 Martin Pelikan, Probabilistic Model-Building GAs
                                                                    61
hBOA on Hierarchical Traps
                                     Experiment
                            6
                           10        O(n1.63 log(n))
   Number of Evaluations




                            5
                           10




                            4
                           10



                                27                81                       243               729
                                                        Problem Size
                                          Martin Pelikan, Probabilistic Model-Building GAs
                                                                                                   62
PMBGAs Are Not Just Optimizers
n    PMBGAs provide us with two things
      ¨  Optimum or its approximation.
      ¨  Sequence of probabilistic models.

n    Probabilistic models
      ¨  Encode   populations of increasing quality.
      ¨  Tell us a lot about the problem at hand.
      ¨  Can we use this information?




                      Martin Pelikan, Probabilistic Model-Building GAs
                                                                         63
Efficiency Enhancement for PMBGAs
n    Sometimes O(n2) is not enough
      ¨  High-dimensional problems (1000s of variables)
      ¨  Expensive evaluation (fitness) function

n    Solution
      ¨  Efficiency   enhancement techniques




                         Martin Pelikan, Probabilistic Model-Building GAs
                                                                            64
Efficiency Enhancement Types
n    7 efficiency enhancement types for PMBGAs
      ¨  Parallelization
      ¨  Hybridization
      ¨  Time  continuation
      ¨  Fitness evaluation relaxation
      ¨  Prior knowledge utilization
      ¨  Incremental and sporadic model building
      ¨  Learning from experience




                       Martin Pelikan, Probabilistic Model-Building GAs
                                                                          65
Multi-objective PMBGAs
n    Methods for multi-objective GAs adopted
      ¨    Multi-objective mixture-based IDEAs
            (Thierens, & Bosman, 2001)

      ¨    Another multi-objective BOA (from SPEA2 and mBOA)
            (Laumanns, & Ocenasek, 2002)

      ¨    Multi-objective hBOA (from NSGA-II and hBOA)
            (Khan, Goldberg, & Pelikan, 2002)
            (Pelikan, Sastry, & Goldberg, 2005)

      ¨    Regularity Model Based Multiobjective EDA (RM-MEDA)
            (Zhang, Zhou, Jin, 2008)

                          Martin Pelikan, Probabilistic Model-Building GAs
                                                                             66
Promising Results with Discrete PMBGAs

n  Artificial classes of problems
n  Physics
n  Bioinformatics
n  Computational complexity and AI
n  Others




               Martin Pelikan, Probabilistic Model-Building GAs
                                                                  67
Results: Artificial Problems
n    Decomposition
      ¨  Concatenated      traps (Pelikan et al., 1998).
n    Hierarchical decomposition
      ¨  Hierarchical   traps (Pelikan, Goldberg, 2001).
n    Other sources of difficulty
      ¨  Exponential    scaling, noise (Pelikan, 2002).




                         Martin Pelikan, Probabilistic Model-Building GAs
                                                                            68
BOA on Concatenated 5-bit Traps
                          500000	
                          450000	
                                             Experiment
                          400000	
           Theory
                          350000	
  Number of Evaluations




                          300000	

                          250000	

                          200000	


                          150000	




                          100000	

                                     100	
       125	
          150	
       175	
      200	
    225	
   250	
                                                           Problem Size
                                             Martin Pelikan, Probabilistic Model-Building GAs
                                                                                                                69
hBOA on Hierarchical Traps
                                     Experiment
                            6
                           10        O(n1.63 log(n))
   Number of Evaluations




                            5
                           10




                            4
                           10



                                27                81                       243               729
                                                        Problem Size
                                          Martin Pelikan, Probabilistic Model-Building GAs
                                                                                                   70
Results: Physics
n    Spin glasses (Pelikan et al., 2002, 2006, 2008)
      (Hoens, 2005) (Santana, 2005) (Shakya et al.,
      2006)
      ¨  ±J and Gaussian couplings
      ¨  2D and 3D spin glass
      ¨  Sherrington-Kirkpatrick (SK) spin glass


n    Silicon clusters (Sastry, 2001)
      ¨  Gong   potential (3-body)


                       Martin Pelikan, Probabilistic Model-Building GAs
                                                                          71
hBOA on Ising Spin Glasses (2D)
                                      hBOA
                                      O(n1.51)
   Number ofof Evaluations
   Number Evaluations




                             3
                       10




                                 64      100        144    196 256 324 400
                                                 Problem Size
                                                Number of Spins
                                       Martin Pelikan, Probabilistic Model-Building GAs
                                                                                          72
Results on 2D Spin Glasses
n  Number of evaluations is O(n 1.51).
n  Overall time is O(n 3.51).
n  Compare O(n3.51) to O(n3.5) for best method
    (Galluccio & Loebl, 1999)
n  Great also on Gaussians.	




                Martin Pelikan, Probabilistic Model-Building GAs
                                                                   73
hBOA on Ising Spin Glasses (3D)
                            6
                           10
                                     Experimental average
                                     O(n3.63 )
   Number of Evaluations
   Number of Evaluations




                            5
                           10




                            4
                           10




                            3
                           10
                                64                     125                    216            343
                                                    Problem Size
                                                   Number of Spins
                                          Martin Pelikan, Probabilistic Model-Building GAs
                                                                                                   74
hBOA on SK Spin Glass




         Martin Pelikan, Probabilistic Model-Building GAs
                                                            75
Results: Computational Complexity, AI
n    MAXSAT, SAT (Pelikan, 2002)
      ¨  Random  3CNF from phase transition.
      ¨  Morphed graph coloring.
      ¨  Conversion from spin glass.


n    Feature subset selection (Inza et al., 2001)
      (Cantu-Paz, 2004)



                     Martin Pelikan, Probabilistic Model-Building GAs
                                                                        76
Results: Some Others
n    Military antenna design (Santarelli et al., 2004)
n    Groundwater remediation design (Arst et al., 2004)
n    Forest management (Ducheyne et al., 2003)
n    Nurse scheduling (Li, Aickelin, 2004)
n    Telecommunication network design (Rothlauf, 2002)
n    Graph partitioning (Ocenasek, Schwarz, 1999; Muehlenbein, Mahnig,
      2002; Baluja, 2004)
n    Portfolio management (Lipinski, 2005, 2007)
n    Quantum excitation chemistry (Sastry et al., 2005)
n    Maximum clique (Zhang et al., 2005)
n    Cancer chemotherapy optimization (Petrovski et al., 2006)
n    Minimum vertex cover (Pelikan et al., 2007)
n    Protein folding (Santana et al., 2007)
n    Side chain placement (Santana et al., 2007)

                         Martin Pelikan, Probabilistic Model-Building GAs
                                                                            77
Discrete PMBGAs: Summary
n    No interactions
      ¨  Univariate   models; PBIL, UMDA, cGA.
n    Some pairwise interactions
      ¨  Tree   models; COMIT, MIMIC, BMDA.
n    Multivariate interactions
      ¨  Multivariate   models: BOA, EBNA, LFDA.
n    Hierarchical decomposition
      ¨  hBOA




                         Martin Pelikan, Probabilistic Model-Building GAs
                                                                            78
Discrete PMBGAs: Recommendations
n    Easy problems
      ¨  Use   univariate models; PBIL, UMDA, cGA.
n    Somewhat difficult problems
      ¨  Use   bivariate models; MIMIC, COMIT, BMDA.
n    Difficult problems
      ¨  Use   multivariate models; BOA, EBNA, LFDA.
n    Most difficult problems
      ¨  Use   hierarchical decomposition; hBOA.


                        Martin Pelikan, Probabilistic Model-Building GAs
                                                                           79
Real-Valued PMBGAs
n    New challenge
      ¨  Infinite
                 domain for each variable.
      ¨  How to model?


n    2 approaches
      ¨  Discretize
                   and apply discrete model/PMBGA
      ¨  Create model for real-valued variables
         n    Estimate pdf.



                           Martin Pelikan, Probabilistic Model-Building GAs
                                                                              80
PBIL Extensions: First Step
n    SHCwL: Stochastic hill climbing with learning
      (Rudlof, Köppen, 1996).
n    Model
      ¨    Single-peak Gaussian for each variable.
      ¨    Means evolve based on parents (promising solutions).
      ¨    Deviations equal, decreasing over time.
n    Problems
      ¨    No interactions.
      ¨    Single Gaussians=can model only one attractor.
      ¨    Same deviations for each variable.


                          Martin Pelikan, Probabilistic Model-Building GAs
                                                                             81
Use Different Deviations
n  Sebag, Ducoulombier (1998)
n  Some variables have higher variance.
n  Use special standard deviation for each
    variable.




                Martin Pelikan, Probabilistic Model-Building GAs
                                                                   82
Use Covariance
n  Covariance allows rotation of 1-peak Gaussians.
n  EGNA (Larrañaga et al., 2000)
n  IDEA (Bosman, Thierens, 2000)




                Martin Pelikan, Probabilistic Model-Building GAs
                                                                   83
How Many Peaks?
n  One Gaussian vs. kernel around each point.
n  Kernel distribution similar to ES.
n  IDEA (Bosman, Thierens, 2000)




               Martin Pelikan, Probabilistic Model-Building GAs
                                                                  84
Mixtures: Between One and Many
n    Mixture distributions provide transition between one
      Gaussian and Gaussian kernels.
n    Mixture types                     4


      ¨    Over one variable.                                   2

            n    Gallagher, Frean, & Downs (1999).
      ¨    Over all variables.                                  0


            n    Pelikan & Goldberg (2000).                    -2
            n    Bosman & Thierens (2000).
      ¨    Over partitions of variables.                       -4
                                                                 -4          -2   0   2   4

            n    Bosman & Thierens (2000).
            n    Ahn, Ramakrishna, and Goldberg (2004).



                               Martin Pelikan, Probabilistic Model-Building GAs
                                                                                          85
Mixed BOA (mBOA)
n  Mixed BOA (Ocenasek, Schwarz, 2002)
n  Local distributions
      ¨  A decision tree (DT) for every variable.
      ¨  Internal DT nodes encode tests on other variables
          n  Discrete:Equal to a constant
          n  Continuous: Less than a constant

      ¨  Discretevariables:
          DT leaves represent probabilities.
      ¨  Continuous variables:
          DT leaves contain a normal kernel distribution.

                          Martin Pelikan, Probabilistic Model-Building GAs
                                                                             86
Real-Coded BOA (rBOA)
n  Ahn, Ramakrishna, Goldberg (2003)
n  Probabilistic Model
      ¨  Underlying   structure: Bayesian network
      ¨  Local distributions: Mixtures of Gaussians

n    Also extended to multiobjective problems
      (Ahn, 2005)




                       Martin Pelikan, Probabilistic Model-Building GAs
                                                                          87
Aggregation Pheromone System (APS)
n  Tsutsui (2004)
n  Inspired by aggregation pheromones
n  Basic idea
      ¨  Good   solutions emit aggregation pheromones
      ¨  New candidate solutions based on the density of
          aggregation pheromones
      ¨  Aggregation pheromone density encodes a mixture
          distribution



                    Martin Pelikan, Probabilistic Model-Building GAs
                                                                       88
Adaptive Variance Scaling
n    Adaptive variance in mBOA
      ¨  Ocenasek   et al. (2004)
n    Normal IDEAs
      ¨  Bosman   et al. (2006, 2007)
      ¨  Correlation-triggered adaptive variance scaling
      ¨  Standard-deviation ratio (SDR) triggered variance
          scaling




                       Martin Pelikan, Probabilistic Model-Building GAs
                                                                          89
Real-Valued PMBGAs: Discretization
n    Idea: Transform into discrete domain.
n    Fixed models
      ¨    2k equal-width bins with k-bit binary string.
      ¨    Goldberg (1989).
      ¨    Bosman & Thierens (2000); Pelikan et al. (2003).
n    Adaptive models
      ¨    Equal-height histograms of 2k bins.
      ¨    k-means clustering on each variable.
      ¨    Pelikan, Goldberg, & Tsutsui (2003); Cantu-Paz (2001).



                          Martin Pelikan, Probabilistic Model-Building GAs
                                                                             90
Real-Valued PMBGAs: Summary
n    Discretization
      ¨  Fixed
      ¨  Adaptive
n    Real-valued models
      ¨  Single or multiple peaks?
      ¨  Same variance or different variance?
      ¨  Covariance or no covariance?
      ¨  Mixtures?
      ¨  Treat entire vectors, subsets of variables, or single
          variables?

                       Martin Pelikan, Probabilistic Model-Building GAs
                                                                          91
Real-Valued PMBGAs: Recommendations
n    Multimodality?
      ¨  Use    multiple peaks.
n    Decomposability?
      ¨  All   variables, subsets, or single variables.
n    Strong linear dependencies?
      ¨  Covariance.

n    Partial differentiability?
      ¨  Combine     with gradient search.


                          Martin Pelikan, Probabilistic Model-Building GAs
                                                                             92
PMBGP (Genetic Programming)
n    New challenge
      ¨  Structured,  variable length representation.
      ¨  Possibly infinitely many values.
      ¨  Position independence (or not).
      ¨  Low correlation between solution quality and
          solution structure (Looks, 2006).

n    Approaches
      ¨  Use explicit probabilistic models for trees.
      ¨  Use models based on grammars.


                        Martin Pelikan, Probabilistic Model-Building GAs
                                                                           93
PIPE
n    Probabilistic incremental
      program evolution
      (Salustowicz &                                                     X    P(X)
      Schmidhuber, 1997)                                                sin   0.15
n    Store frequencies of                                               +    0.35
      operators/terminals in
                                                                         -    0.35
      nodes of a maximum tree.
                                                                         X    0.15
n    Sampling generates tree
      from top to bottom



                     Martin Pelikan, Probabilistic Model-Building GAs
                                                                                 94
eCGP
n  Sastry & Goldberg (2003)
n  ECGA adapted to program trees.
n  Maximum tree as in PIPE.
n  But nodes partitioned into groups.




                Martin Pelikan, Probabilistic Model-Building GAs
                                                                   95
BOA for GP
n  Looks, Goertzel, & Pennachin (2004)
n  Combinatory logic + BOA
      ¨  Trees translated into uniform structures.
      ¨  Labels only in leaves.
      ¨  BOA builds model over symbols in different nodes.

n    Complexity build-up
      ¨  Modeling limited to max. sized structure seen.
      ¨  Complexity builds up by special operator.



                     Martin Pelikan, Probabilistic Model-Building GAs
                                                                        96
MOSES
n  Looks (2006).
n  Evolve demes of programs.
n  Each deme represents similar structures.
n  Apply PMBGA to each deme (e.g. hBOA).
n  Introduce new demes/delete old ones.
n  Use normal forms to reduce complexity.




                Martin Pelikan, Probabilistic Model-Building GAs
                                                                   97
PMBGP with Grammars
n    Use grammars/stochastic grammars as models.
n    Grammars restrict the class of programs.

n    Some representatives
      ¨    Program evolution with explicit learning (Shan et al., 2003)
      ¨    Grammar-based EDA for GP (Bosman, de Jong, 2004)
      ¨    Stochastic grammar GP (Tanev, 2004)
      ¨    Adaptive constrained GP (Janikow, 2004)




                           Martin Pelikan, Probabilistic Model-Building GAs
                                                                              98
PMBGP: Summary
n  Interesting starting points available.
n  But still lot of work to be done.
n  Much to learn from discrete domain, but some
    completely new challenges.
n  Research in progress




               Martin Pelikan, Probabilistic Model-Building GAs
                                                                  99
PMBGAs for Permutations
n    New challenges
      ¨  Relativeorder
      ¨  Absolute order
      ¨  Permutation constraints

n    Two basic approaches
      ¨  Random-key    and real-valued PMBGAs
      ¨  Explicit probabilistic models for permutations




                       Martin Pelikan, Probabilistic Model-Building GAs
                                                                          100
Random Keys and PMBGAs
n    Bengoetxea et al. (2000); Bosman et al. (2001)
n    Random keys (Bean, 1997)
      ¨    Candidate solution = vector of real values
      ¨    Ascending ordering gives a permutation
n    Can use any real-valued PMBGA (or GEA)
      ¨    IDEAs (Bosman, Thierens, 2002)
      ¨    EGNA (Larranaga et al., 2001)
n    Strengths and weaknesses
      ¨    Good: Can use any real-valued PMBGA.
      ¨    Bad: Redundancy of the encoding.


                           Martin Pelikan, Probabilistic Model-Building GAs
                                                                              101
Direct Modeling of Permutations
n    Edge-histogram based sampling algorithm
      (EHBSA) (Tsutsui, Pelikan, Goldberg, 2003)
      ¨  Permutations   of n elements
      ¨  Model is a matrix A=(ai,j)i,j=1, 2, …, n
      ¨  ai,j represents the probability of edge (i, j)
      ¨  Uses template to reduce exploration
      ¨  Applicable also to scheduling




                       Martin Pelikan, Probabilistic Model-Building GAs
                                                                          102
ICE: Modify Crossover from Model
n    ICE
      ¨  Bosman,  Thierens (2001).
      ¨  Represent permutations with random keys.
      ¨  Learn multivariate model to factorize the problem.
      ¨  Use the learned model to modify crossover.

n    Performance
      ¨  Typically
                  outperforms IDEAs and other PMBGAs
        that learn and sample random keys.



                      Martin Pelikan, Probabilistic Model-Building GAs
                                                                         103
Multivariate Permutation Models
n    Basic approach
      ¨    Use any standard multivariate discrete model.
      ¨    Restrict sampling to permutations in some way.
      ¨    Bengoetxea et al. (2000), Pelikan et al. (2007).
n    Strengths and weaknesses
      ¨    Use explicit multivariate models to find regularities.
      ¨    High-order alphabet requires big samples for good models.
      ¨    Sampling can introduce unwanted bias.
      ¨    Inefficient encoding for only relative ordering constraints,
            which can be encoded simpler.




                           Martin Pelikan, Probabilistic Model-Building GAs
                                                                              104
Conclusions
n    Competent PMBGAs exist
      ¨    Scalable solution to broad classes of problems.
      ¨    Solution to previously intractable problems.
      ¨    Algorithms ready for new applications.
n    PMBGAs do more than just solve the problem
      ¨    They provide us with sequences of probabilistic models.
      ¨    The probabilistic models tell us a lot about the problem.
n    Consequences for practitioners
      ¨    Robust methods with few or no parameters.
      ¨    Capable of learning how to solve problem.
      ¨    But can incorporate prior knowledge as well.
      ¨    Can solve previously intractable problems.


                           Martin Pelikan, Probabilistic Model-Building GAs
                                                                              105
Starting Points
n    World wide web
n    Books and surveys
       ¨  Larrañaga & Lozano (eds.) (2001). Estimation of distribution
           algorithms: A new tool for evolutionary computation. Kluwer.
       ¨  Pelikan et al. (2002). A survey to optimization by building and
           using probabilistic models. Computational optimization and
           applications, 21(1), pp. 5-20.
       ¨  Pelikan (2005). Hierarchical BOA: Towards a New Generation of
           Evolutionary Algorithms. Springer.
       ¨  Lozano, Larrañaga, Inza, Bengoetxea (2006). Towards a New
           Evolutionary Computation: Advances on Estimation of Distribution
           Algorithms, Springer.
       ¨  Pelikan, Sastry, Cantu-Paz (eds.) (2006). Scalable Optimization via
           Probabilistic Modeling: From Algorithms to Applications, Springer.


                           Martin Pelikan, Probabilistic Model-Building GAs
                                                                              106
Online Code (1/2)
n    BOA, BOA with decision graphs, dependency-tree EDA
      http://medal-lab.org/

n    ECGA, xi-ary ECGA, BOA, and BOA with decision trees/graphs
      http://www.illigal.org/

n    mBOA
      http://jiri.ocenasek.com/

n    PIPE
      http://www.idsia.ch/~rafal/

n    Real-coded BOA
      http://www.evolution.re.kr/

                         Martin Pelikan, Probabilistic Model-Building GAs
                                                                            107
Online Code (2/2)
n    Demos of APS and EHBSA
      http://www.hannan-u.ac.jp/~tsutsui/research-e.html

n    RM-MEDA: A Regularity Model Based Multiobjective EDA
      Differential Evolution + EDA hybrid
      http://cswww.essex.ac.uk/staff/qzhang/mypublication.htm

n    Naive Multi-objective Mixture-based IDEA (MIDEA)
      Normal IDEA-Induced Chromosome Elements Exchanger (ICE)
      Normal Iterated Density-Estimation Evolutionary Algorithm (IDEA)
      http://homepages.cwi.nl/~bosman/code.html




                          Martin Pelikan, Probabilistic Model-Building GAs
                                                                             108

Weitere ähnliche Inhalte

Was ist angesagt?

Hyperparameter Tuning
Hyperparameter TuningHyperparameter Tuning
Hyperparameter TuningJon Lederman
 
Hexagonal Boron Nitride - Ubiquitous Layered dielectric for Two-Dimensional E...
Hexagonal Boron Nitride - Ubiquitous Layered dielectric for Two-Dimensional E...Hexagonal Boron Nitride - Ubiquitous Layered dielectric for Two-Dimensional E...
Hexagonal Boron Nitride - Ubiquitous Layered dielectric for Two-Dimensional E...Nikhil Jain
 
Genetic Algorithm (GA) Optimization - Step-by-Step Example
Genetic Algorithm (GA) Optimization - Step-by-Step ExampleGenetic Algorithm (GA) Optimization - Step-by-Step Example
Genetic Algorithm (GA) Optimization - Step-by-Step ExampleAhmed Gad
 
Hydrothermal Treatment Of Solid Waste
Hydrothermal Treatment Of Solid WasteHydrothermal Treatment Of Solid Waste
Hydrothermal Treatment Of Solid WasteKetan Wadodkar
 
Photoelectron spectroscopy
Photoelectron spectroscopyPhotoelectron spectroscopy
Photoelectron spectroscopytesfayehh
 
P05 deep boltzmann machines cvpr2012 deep learning methods for vision
P05 deep boltzmann machines cvpr2012 deep learning methods for visionP05 deep boltzmann machines cvpr2012 deep learning methods for vision
P05 deep boltzmann machines cvpr2012 deep learning methods for visionzukun
 
Rishabh Mehrotra - Recommendations in a Marketplace: Personalizing Explainabl...
Rishabh Mehrotra - Recommendations in a Marketplace: Personalizing Explainabl...Rishabh Mehrotra - Recommendations in a Marketplace: Personalizing Explainabl...
Rishabh Mehrotra - Recommendations in a Marketplace: Personalizing Explainabl...MLconf
 
Deep Learning for Recommender Systems RecSys2017 Tutorial
Deep Learning for Recommender Systems RecSys2017 Tutorial Deep Learning for Recommender Systems RecSys2017 Tutorial
Deep Learning for Recommender Systems RecSys2017 Tutorial Alexandros Karatzoglou
 
Cluster Analysis
Cluster AnalysisCluster Analysis
Cluster Analysisguest0edcaf
 
Coursera Machine Learning (by Andrew Ng)_강의정리
Coursera Machine Learning (by Andrew Ng)_강의정리Coursera Machine Learning (by Andrew Ng)_강의정리
Coursera Machine Learning (by Andrew Ng)_강의정리SANG WON PARK
 
Deep Dive into Hyperparameter Tuning
Deep Dive into Hyperparameter TuningDeep Dive into Hyperparameter Tuning
Deep Dive into Hyperparameter TuningShubhmay Potdar
 
Learning to Rank Presentation (v2) at LexisNexis Search Guild
Learning to Rank Presentation (v2) at LexisNexis Search GuildLearning to Rank Presentation (v2) at LexisNexis Search Guild
Learning to Rank Presentation (v2) at LexisNexis Search GuildSujit Pal
 
Computational materials design with high-throughput and machine learning methods
Computational materials design with high-throughput and machine learning methodsComputational materials design with high-throughput and machine learning methods
Computational materials design with high-throughput and machine learning methodsAnubhav Jain
 
Introduction to Optimization with Genetic Algorithm (GA)
Introduction to Optimization with Genetic Algorithm (GA)Introduction to Optimization with Genetic Algorithm (GA)
Introduction to Optimization with Genetic Algorithm (GA)Ahmed Gad
 
Counterfactual evaluation of machine learning models
Counterfactual evaluation of machine learning modelsCounterfactual evaluation of machine learning models
Counterfactual evaluation of machine learning modelsMichael Manapat
 
Content based recommendation systems
Content based recommendation systemsContent based recommendation systems
Content based recommendation systemsAravindharamanan S
 
Netflix Recommendations - Beyond the 5 Stars
Netflix Recommendations - Beyond the 5 StarsNetflix Recommendations - Beyond the 5 Stars
Netflix Recommendations - Beyond the 5 StarsXavier Amatriain
 
Engagement, metrics and "recommenders"
Engagement, metrics and "recommenders"Engagement, metrics and "recommenders"
Engagement, metrics and "recommenders"Mounia Lalmas-Roelleke
 
GRU4Rec v2 - Recurrent Neural Networks with Top-k Gains for Session-based Rec...
GRU4Rec v2 - Recurrent Neural Networks with Top-k Gains for Session-based Rec...GRU4Rec v2 - Recurrent Neural Networks with Top-k Gains for Session-based Rec...
GRU4Rec v2 - Recurrent Neural Networks with Top-k Gains for Session-based Rec...Balázs Hidasi
 
Tutorial: Context In Recommender Systems
Tutorial: Context In Recommender SystemsTutorial: Context In Recommender Systems
Tutorial: Context In Recommender SystemsYONG ZHENG
 

Was ist angesagt? (20)

Hyperparameter Tuning
Hyperparameter TuningHyperparameter Tuning
Hyperparameter Tuning
 
Hexagonal Boron Nitride - Ubiquitous Layered dielectric for Two-Dimensional E...
Hexagonal Boron Nitride - Ubiquitous Layered dielectric for Two-Dimensional E...Hexagonal Boron Nitride - Ubiquitous Layered dielectric for Two-Dimensional E...
Hexagonal Boron Nitride - Ubiquitous Layered dielectric for Two-Dimensional E...
 
Genetic Algorithm (GA) Optimization - Step-by-Step Example
Genetic Algorithm (GA) Optimization - Step-by-Step ExampleGenetic Algorithm (GA) Optimization - Step-by-Step Example
Genetic Algorithm (GA) Optimization - Step-by-Step Example
 
Hydrothermal Treatment Of Solid Waste
Hydrothermal Treatment Of Solid WasteHydrothermal Treatment Of Solid Waste
Hydrothermal Treatment Of Solid Waste
 
Photoelectron spectroscopy
Photoelectron spectroscopyPhotoelectron spectroscopy
Photoelectron spectroscopy
 
P05 deep boltzmann machines cvpr2012 deep learning methods for vision
P05 deep boltzmann machines cvpr2012 deep learning methods for visionP05 deep boltzmann machines cvpr2012 deep learning methods for vision
P05 deep boltzmann machines cvpr2012 deep learning methods for vision
 
Rishabh Mehrotra - Recommendations in a Marketplace: Personalizing Explainabl...
Rishabh Mehrotra - Recommendations in a Marketplace: Personalizing Explainabl...Rishabh Mehrotra - Recommendations in a Marketplace: Personalizing Explainabl...
Rishabh Mehrotra - Recommendations in a Marketplace: Personalizing Explainabl...
 
Deep Learning for Recommender Systems RecSys2017 Tutorial
Deep Learning for Recommender Systems RecSys2017 Tutorial Deep Learning for Recommender Systems RecSys2017 Tutorial
Deep Learning for Recommender Systems RecSys2017 Tutorial
 
Cluster Analysis
Cluster AnalysisCluster Analysis
Cluster Analysis
 
Coursera Machine Learning (by Andrew Ng)_강의정리
Coursera Machine Learning (by Andrew Ng)_강의정리Coursera Machine Learning (by Andrew Ng)_강의정리
Coursera Machine Learning (by Andrew Ng)_강의정리
 
Deep Dive into Hyperparameter Tuning
Deep Dive into Hyperparameter TuningDeep Dive into Hyperparameter Tuning
Deep Dive into Hyperparameter Tuning
 
Learning to Rank Presentation (v2) at LexisNexis Search Guild
Learning to Rank Presentation (v2) at LexisNexis Search GuildLearning to Rank Presentation (v2) at LexisNexis Search Guild
Learning to Rank Presentation (v2) at LexisNexis Search Guild
 
Computational materials design with high-throughput and machine learning methods
Computational materials design with high-throughput and machine learning methodsComputational materials design with high-throughput and machine learning methods
Computational materials design with high-throughput and machine learning methods
 
Introduction to Optimization with Genetic Algorithm (GA)
Introduction to Optimization with Genetic Algorithm (GA)Introduction to Optimization with Genetic Algorithm (GA)
Introduction to Optimization with Genetic Algorithm (GA)
 
Counterfactual evaluation of machine learning models
Counterfactual evaluation of machine learning modelsCounterfactual evaluation of machine learning models
Counterfactual evaluation of machine learning models
 
Content based recommendation systems
Content based recommendation systemsContent based recommendation systems
Content based recommendation systems
 
Netflix Recommendations - Beyond the 5 Stars
Netflix Recommendations - Beyond the 5 StarsNetflix Recommendations - Beyond the 5 Stars
Netflix Recommendations - Beyond the 5 Stars
 
Engagement, metrics and "recommenders"
Engagement, metrics and "recommenders"Engagement, metrics and "recommenders"
Engagement, metrics and "recommenders"
 
GRU4Rec v2 - Recurrent Neural Networks with Top-k Gains for Session-based Rec...
GRU4Rec v2 - Recurrent Neural Networks with Top-k Gains for Session-based Rec...GRU4Rec v2 - Recurrent Neural Networks with Top-k Gains for Session-based Rec...
GRU4Rec v2 - Recurrent Neural Networks with Top-k Gains for Session-based Rec...
 
Tutorial: Context In Recommender Systems
Tutorial: Context In Recommender SystemsTutorial: Context In Recommender Systems
Tutorial: Context In Recommender Systems
 

Ähnlich wie Estimation of Distribution Algorithms Tutorial

Variational Autoencoders VAE - Santiago Pascual - UPC Barcelona 2018
Variational Autoencoders VAE - Santiago Pascual - UPC Barcelona 2018Variational Autoencoders VAE - Santiago Pascual - UPC Barcelona 2018
Variational Autoencoders VAE - Santiago Pascual - UPC Barcelona 2018Universitat Politècnica de Catalunya
 
Using Problem-Specific Knowledge and Learning from Experience in Estimation o...
Using Problem-Specific Knowledge and Learning from Experience in Estimation o...Using Problem-Specific Knowledge and Learning from Experience in Estimation o...
Using Problem-Specific Knowledge and Learning from Experience in Estimation o...Martin Pelikan
 
Score based generative model
Score based generative modelScore based generative model
Score based generative modelsangyun lee
 
Methods of Manifold Learning for Dimension Reduction of Large Data Sets
Methods of Manifold Learning for Dimension Reduction of Large Data SetsMethods of Manifold Learning for Dimension Reduction of Large Data Sets
Methods of Manifold Learning for Dimension Reduction of Large Data SetsRyan B Harvey, CSDP, CSM
 
A New Model for Credit Approval Problems: A Neuro-Genetic System with Quantum...
A New Model for Credit Approval Problems: A Neuro-Genetic System with Quantum...A New Model for Credit Approval Problems: A Neuro-Genetic System with Quantum...
A New Model for Credit Approval Problems: A Neuro-Genetic System with Quantum...Anderson Pinho
 
Ordinal Regression and Machine Learning: Applications, Methods, Metrics
Ordinal Regression and Machine Learning: Applications, Methods, MetricsOrdinal Regression and Machine Learning: Applications, Methods, Metrics
Ordinal Regression and Machine Learning: Applications, Methods, MetricsFrancesco Casalegno
 

Ähnlich wie Estimation of Distribution Algorithms Tutorial (7)

Variational Autoencoders VAE - Santiago Pascual - UPC Barcelona 2018
Variational Autoencoders VAE - Santiago Pascual - UPC Barcelona 2018Variational Autoencoders VAE - Santiago Pascual - UPC Barcelona 2018
Variational Autoencoders VAE - Santiago Pascual - UPC Barcelona 2018
 
Using Problem-Specific Knowledge and Learning from Experience in Estimation o...
Using Problem-Specific Knowledge and Learning from Experience in Estimation o...Using Problem-Specific Knowledge and Learning from Experience in Estimation o...
Using Problem-Specific Knowledge and Learning from Experience in Estimation o...
 
Deep Learning Opening Workshop - Improving Generative Models - Junier Oliva, ...
Deep Learning Opening Workshop - Improving Generative Models - Junier Oliva, ...Deep Learning Opening Workshop - Improving Generative Models - Junier Oliva, ...
Deep Learning Opening Workshop - Improving Generative Models - Junier Oliva, ...
 
Score based generative model
Score based generative modelScore based generative model
Score based generative model
 
Methods of Manifold Learning for Dimension Reduction of Large Data Sets
Methods of Manifold Learning for Dimension Reduction of Large Data SetsMethods of Manifold Learning for Dimension Reduction of Large Data Sets
Methods of Manifold Learning for Dimension Reduction of Large Data Sets
 
A New Model for Credit Approval Problems: A Neuro-Genetic System with Quantum...
A New Model for Credit Approval Problems: A Neuro-Genetic System with Quantum...A New Model for Credit Approval Problems: A Neuro-Genetic System with Quantum...
A New Model for Credit Approval Problems: A Neuro-Genetic System with Quantum...
 
Ordinal Regression and Machine Learning: Applications, Methods, Metrics
Ordinal Regression and Machine Learning: Applications, Methods, MetricsOrdinal Regression and Machine Learning: Applications, Methods, Metrics
Ordinal Regression and Machine Learning: Applications, Methods, Metrics
 

Mehr von Martin Pelikan

Transfer Learning, Soft Distance-Based Bias, and the Hierarchical BOA
Transfer Learning, Soft Distance-Based Bias, and the Hierarchical BOATransfer Learning, Soft Distance-Based Bias, and the Hierarchical BOA
Transfer Learning, Soft Distance-Based Bias, and the Hierarchical BOAMartin Pelikan
 
Population Dynamics in Conway’s Game of Life and its Variants
Population Dynamics in Conway’s Game of Life and its VariantsPopulation Dynamics in Conway’s Game of Life and its Variants
Population Dynamics in Conway’s Game of Life and its VariantsMartin Pelikan
 
Image segmentation using a genetic algorithm and hierarchical local search
Image segmentation using a genetic algorithm and hierarchical local searchImage segmentation using a genetic algorithm and hierarchical local search
Image segmentation using a genetic algorithm and hierarchical local searchMartin Pelikan
 
Distance-based bias in model-directed optimization of additively decomposable...
Distance-based bias in model-directed optimization of additively decomposable...Distance-based bias in model-directed optimization of additively decomposable...
Distance-based bias in model-directed optimization of additively decomposable...Martin Pelikan
 
Pairwise and Problem-Specific Distance Metrics in the Linkage Tree Genetic Al...
Pairwise and Problem-Specific Distance Metrics in the Linkage Tree Genetic Al...Pairwise and Problem-Specific Distance Metrics in the Linkage Tree Genetic Al...
Pairwise and Problem-Specific Distance Metrics in the Linkage Tree Genetic Al...Martin Pelikan
 
Spurious Dependencies and EDA Scalability
Spurious Dependencies and EDA ScalabilitySpurious Dependencies and EDA Scalability
Spurious Dependencies and EDA ScalabilityMartin Pelikan
 
Effects of a Deterministic Hill climber on hBOA
Effects of a Deterministic Hill climber on hBOAEffects of a Deterministic Hill climber on hBOA
Effects of a Deterministic Hill climber on hBOAMartin Pelikan
 
Intelligent Bias of Network Structures in the Hierarchical BOA
Intelligent Bias of Network Structures in the Hierarchical BOAIntelligent Bias of Network Structures in the Hierarchical BOA
Intelligent Bias of Network Structures in the Hierarchical BOAMartin Pelikan
 
Analysis of Evolutionary Algorithms on the One-Dimensional Spin Glass with Po...
Analysis of Evolutionary Algorithms on the One-Dimensional Spin Glass with Po...Analysis of Evolutionary Algorithms on the One-Dimensional Spin Glass with Po...
Analysis of Evolutionary Algorithms on the One-Dimensional Spin Glass with Po...Martin Pelikan
 
Performance of Evolutionary Algorithms on NK Landscapes with Nearest Neighbor...
Performance of Evolutionary Algorithms on NK Landscapes with Nearest Neighbor...Performance of Evolutionary Algorithms on NK Landscapes with Nearest Neighbor...
Performance of Evolutionary Algorithms on NK Landscapes with Nearest Neighbor...Martin Pelikan
 
Initial-Population Bias in the Univariate Estimation of Distribution Algorithm
Initial-Population Bias in the Univariate Estimation of Distribution AlgorithmInitial-Population Bias in the Univariate Estimation of Distribution Algorithm
Initial-Population Bias in the Univariate Estimation of Distribution AlgorithmMartin Pelikan
 
Using Previous Models to Bias Structural Learning in the Hierarchical BOA
Using Previous Models to Bias Structural Learning in the Hierarchical BOAUsing Previous Models to Bias Structural Learning in the Hierarchical BOA
Using Previous Models to Bias Structural Learning in the Hierarchical BOAMartin Pelikan
 
Efficiency Enhancement of Estimation of Distribution Algorithms
Efficiency Enhancement of Estimation of Distribution AlgorithmsEfficiency Enhancement of Estimation of Distribution Algorithms
Efficiency Enhancement of Estimation of Distribution AlgorithmsMartin Pelikan
 
Finding Ground States of Sherrington-Kirkpatrick Spin Glasses with Hierarchic...
Finding Ground States of Sherrington-Kirkpatrick Spin Glasses with Hierarchic...Finding Ground States of Sherrington-Kirkpatrick Spin Glasses with Hierarchic...
Finding Ground States of Sherrington-Kirkpatrick Spin Glasses with Hierarchic...Martin Pelikan
 
iBOA: The Incremental Bayesian Optimization Algorithm
iBOA: The Incremental Bayesian Optimization AlgorithmiBOA: The Incremental Bayesian Optimization Algorithm
iBOA: The Incremental Bayesian Optimization AlgorithmMartin Pelikan
 
Fitness inheritance in the Bayesian optimization algorithm
Fitness inheritance in the Bayesian optimization algorithmFitness inheritance in the Bayesian optimization algorithm
Fitness inheritance in the Bayesian optimization algorithmMartin Pelikan
 
Computational complexity and simulation of rare events of Ising spin glasses
Computational complexity and simulation of rare events of Ising spin glasses Computational complexity and simulation of rare events of Ising spin glasses
Computational complexity and simulation of rare events of Ising spin glasses Martin Pelikan
 
The Bayesian Optimization Algorithm with Substructural Local Search
The Bayesian Optimization Algorithm with Substructural Local SearchThe Bayesian Optimization Algorithm with Substructural Local Search
The Bayesian Optimization Algorithm with Substructural Local SearchMartin Pelikan
 
Analyzing Probabilistic Models in Hierarchical BOA on Traps and Spin Glasses
Analyzing Probabilistic Models in Hierarchical BOA on Traps and Spin GlassesAnalyzing Probabilistic Models in Hierarchical BOA on Traps and Spin Glasses
Analyzing Probabilistic Models in Hierarchical BOA on Traps and Spin GlassesMartin Pelikan
 
Hybrid Evolutionary Algorithms on Minimum Vertex Cover for Random Graphs
Hybrid Evolutionary Algorithms on Minimum Vertex Cover for Random GraphsHybrid Evolutionary Algorithms on Minimum Vertex Cover for Random Graphs
Hybrid Evolutionary Algorithms on Minimum Vertex Cover for Random GraphsMartin Pelikan
 

Mehr von Martin Pelikan (20)

Transfer Learning, Soft Distance-Based Bias, and the Hierarchical BOA
Transfer Learning, Soft Distance-Based Bias, and the Hierarchical BOATransfer Learning, Soft Distance-Based Bias, and the Hierarchical BOA
Transfer Learning, Soft Distance-Based Bias, and the Hierarchical BOA
 
Population Dynamics in Conway’s Game of Life and its Variants
Population Dynamics in Conway’s Game of Life and its VariantsPopulation Dynamics in Conway’s Game of Life and its Variants
Population Dynamics in Conway’s Game of Life and its Variants
 
Image segmentation using a genetic algorithm and hierarchical local search
Image segmentation using a genetic algorithm and hierarchical local searchImage segmentation using a genetic algorithm and hierarchical local search
Image segmentation using a genetic algorithm and hierarchical local search
 
Distance-based bias in model-directed optimization of additively decomposable...
Distance-based bias in model-directed optimization of additively decomposable...Distance-based bias in model-directed optimization of additively decomposable...
Distance-based bias in model-directed optimization of additively decomposable...
 
Pairwise and Problem-Specific Distance Metrics in the Linkage Tree Genetic Al...
Pairwise and Problem-Specific Distance Metrics in the Linkage Tree Genetic Al...Pairwise and Problem-Specific Distance Metrics in the Linkage Tree Genetic Al...
Pairwise and Problem-Specific Distance Metrics in the Linkage Tree Genetic Al...
 
Spurious Dependencies and EDA Scalability
Spurious Dependencies and EDA ScalabilitySpurious Dependencies and EDA Scalability
Spurious Dependencies and EDA Scalability
 
Effects of a Deterministic Hill climber on hBOA
Effects of a Deterministic Hill climber on hBOAEffects of a Deterministic Hill climber on hBOA
Effects of a Deterministic Hill climber on hBOA
 
Intelligent Bias of Network Structures in the Hierarchical BOA
Intelligent Bias of Network Structures in the Hierarchical BOAIntelligent Bias of Network Structures in the Hierarchical BOA
Intelligent Bias of Network Structures in the Hierarchical BOA
 
Analysis of Evolutionary Algorithms on the One-Dimensional Spin Glass with Po...
Analysis of Evolutionary Algorithms on the One-Dimensional Spin Glass with Po...Analysis of Evolutionary Algorithms on the One-Dimensional Spin Glass with Po...
Analysis of Evolutionary Algorithms on the One-Dimensional Spin Glass with Po...
 
Performance of Evolutionary Algorithms on NK Landscapes with Nearest Neighbor...
Performance of Evolutionary Algorithms on NK Landscapes with Nearest Neighbor...Performance of Evolutionary Algorithms on NK Landscapes with Nearest Neighbor...
Performance of Evolutionary Algorithms on NK Landscapes with Nearest Neighbor...
 
Initial-Population Bias in the Univariate Estimation of Distribution Algorithm
Initial-Population Bias in the Univariate Estimation of Distribution AlgorithmInitial-Population Bias in the Univariate Estimation of Distribution Algorithm
Initial-Population Bias in the Univariate Estimation of Distribution Algorithm
 
Using Previous Models to Bias Structural Learning in the Hierarchical BOA
Using Previous Models to Bias Structural Learning in the Hierarchical BOAUsing Previous Models to Bias Structural Learning in the Hierarchical BOA
Using Previous Models to Bias Structural Learning in the Hierarchical BOA
 
Efficiency Enhancement of Estimation of Distribution Algorithms
Efficiency Enhancement of Estimation of Distribution AlgorithmsEfficiency Enhancement of Estimation of Distribution Algorithms
Efficiency Enhancement of Estimation of Distribution Algorithms
 
Finding Ground States of Sherrington-Kirkpatrick Spin Glasses with Hierarchic...
Finding Ground States of Sherrington-Kirkpatrick Spin Glasses with Hierarchic...Finding Ground States of Sherrington-Kirkpatrick Spin Glasses with Hierarchic...
Finding Ground States of Sherrington-Kirkpatrick Spin Glasses with Hierarchic...
 
iBOA: The Incremental Bayesian Optimization Algorithm
iBOA: The Incremental Bayesian Optimization AlgorithmiBOA: The Incremental Bayesian Optimization Algorithm
iBOA: The Incremental Bayesian Optimization Algorithm
 
Fitness inheritance in the Bayesian optimization algorithm
Fitness inheritance in the Bayesian optimization algorithmFitness inheritance in the Bayesian optimization algorithm
Fitness inheritance in the Bayesian optimization algorithm
 
Computational complexity and simulation of rare events of Ising spin glasses
Computational complexity and simulation of rare events of Ising spin glasses Computational complexity and simulation of rare events of Ising spin glasses
Computational complexity and simulation of rare events of Ising spin glasses
 
The Bayesian Optimization Algorithm with Substructural Local Search
The Bayesian Optimization Algorithm with Substructural Local SearchThe Bayesian Optimization Algorithm with Substructural Local Search
The Bayesian Optimization Algorithm with Substructural Local Search
 
Analyzing Probabilistic Models in Hierarchical BOA on Traps and Spin Glasses
Analyzing Probabilistic Models in Hierarchical BOA on Traps and Spin GlassesAnalyzing Probabilistic Models in Hierarchical BOA on Traps and Spin Glasses
Analyzing Probabilistic Models in Hierarchical BOA on Traps and Spin Glasses
 
Hybrid Evolutionary Algorithms on Minimum Vertex Cover for Random Graphs
Hybrid Evolutionary Algorithms on Minimum Vertex Cover for Random GraphsHybrid Evolutionary Algorithms on Minimum Vertex Cover for Random Graphs
Hybrid Evolutionary Algorithms on Minimum Vertex Cover for Random Graphs
 

Kürzlich hochgeladen

Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebUiPathCommunity
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024Lonnie McRorey
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .Alan Dix
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxhariprasad279825
 
Powerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time ClashPowerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time Clashcharlottematthew16
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024Stephanie Beckett
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr BaganFwdays
 
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DayH2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DaySri Ambati
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxNavinnSomaal
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsMark Billinghurst
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...Fwdays
 
Search Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfSearch Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfRankYa
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.Curtis Poe
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxLoriGlavin3
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationSlibray Presentation
 
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostLeverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostZilliz
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLScyllaDB
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 3652toLead Limited
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek SchlawackFwdays
 

Kürzlich hochgeladen (20)

Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio Web
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptx
 
Powerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time ClashPowerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time Clash
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan
 
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DayH2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptx
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR Systems
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
 
Search Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfSearch Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdf
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck Presentation
 
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostLeverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQL
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365
 
DMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special EditionDMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special Edition
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
 

Estimation of Distribution Algorithms Tutorial

  • 1. Probabilistic Model-Building Genetic Algorithms a.k.a. Estimation of Distribution Algorithms a.k.a. Iterated Density Estimation Algorithms Martin Pelikan Missouri Estimation of Distribution Algorithms Laboratory (MEDAL) Department of Mathematics and Computer Science University of Missouri - St. Louis martin@martinpelikan.net Copyright is held by the author/owner(s). http://martinpelikan.net/ GECCO’12 Companion, July 7–11, 2012, Philadelphia, PA, USA. ACM 978-1-4503-1178-6/12/07.
  • 2. Foreword n  Motivation ¨  Genetic and evolutionary computation (GEC) popular. ¨  Toy problems great, but difficulties in practice. ¨  Must design new representations, operators, tune, … n  This talk ¨  Discuss a promising direction in GEC. ¨  Combine machine learning and GEC. ¨  Create practical and powerful optimizers. Martin Pelikan, Probabilistic Model-Building GAs 2
  • 3. Overview n  Introduction ¨  Black-box optimization via probabilistic modeling. n  Probabilistic Model-Building GAs ¨  Discreterepresentation ¨  Continuous representation ¨  Computer programs (PMBGP) ¨  Permutations n  Conclusions Martin Pelikan, Probabilistic Model-Building GAs 3
  • 4. Problem Formulation n  Input ¨  How do potential solutions look like? ¨  How to evaluate quality of potential solutions? n  Output ¨  Best solution (the optimum). n  Important ¨  No additional knowledge about the problem. Martin Pelikan, Probabilistic Model-Building GAs 4
  • 5. Why View Problem as Black Box? n  Advantages ¨  Separate problem definition from optimizer. ¨  Easy to solve new problems. ¨  Economy argument. n  Difficulties ¨  Almost no prior problem knowledge. ¨  Problem specifics must be learned automatically. ¨  Noise, multiple objectives, interactive evaluation. Martin Pelikan, Probabilistic Model-Building GAs 5
  • 6. Representations Considered Here n  Start with ¨  Solutions are n-bit binary strings. n  Later ¨  Real-valuedvectors. ¨  Program trees. ¨  Permutations Martin Pelikan, Probabilistic Model-Building GAs 6
  • 7. Typical Situation n  Previously visited solutions + their evaluation: # Solution Evaluation 1 00100 1 2 11011 4 3 01101 0 4 10111 3 n  Question: What solution to generate next? Martin Pelikan, Probabilistic Model-Building GAs 7
  • 8. Many Answers n  Hill climber ¨  Start with a random solution. ¨  Flip bit that improves the solution most. ¨  Finish when no more improvement possible. n  Simulated annealing ¨  Introduce Metropolis. n  Probabilistic model-building GAs ¨  Inspiration from GAs and machine learning (ML). Martin Pelikan, Probabilistic Model-Building GAs 8
  • 9. Probabilistic Model-Building GAs Current Selected New population population population 11001 11001 01111 11101 10101 Probabilistic 11001 01011 01011 Model 11011 11000 11000 00111 …replace crossover+mutation with learning and sampling probabilistic model Martin Pelikan, Probabilistic Model-Building GAs 9
  • 10. Other Names for PMBGAs n  Estimation of distribution algorithms (EDAs) (Mühlenbein & Paass, 1996) n  Iterated density estimation algorithms (IDEA) (Bosman & Thierens, 2000) Martin Pelikan, Probabilistic Model-Building GAs 10
  • 11. Implicit vs. Explicit Model n  GAs and PMBGAs perform similar task ¨  Generate new solutions using probability distribution based on selected solutions. n  GAs ¨  Variationdefines implicit probability distribution of target population given original population and variation operators (crossover and mutation). n  PMBGAs ¨  Explicit probabilistic model of selected candidate solutions is built and sampled. Martin Pelikan, Probabilistic Model-Building GAs 11
  • 12. What Models to Use? n  Start with a simple example ¨  Probability vector for binary strings. n  Later ¨  Dependency tree models (COMIT). ¨  Bayesian networks (BOA). ¨  Bayesian networks with local structures (hBOA). Martin Pelikan, Probabilistic Model-Building GAs 12
  • 13. Probability Vector n  Assume n-bit binary strings. n  Model: Probability vector p=(p1, …, pn) ¨  pi= probability of 1 in position i ¨  Learn p: Compute proportion of 1 in each position. ¨  Sample p: Sample 1 in position i with prob. pi Martin Pelikan, Probabilistic Model-Building GAs 13
  • 14. Example: Probability Vector (Mühlenbein, Paass, 1996), (Baluja, 1994) Current Selected New population population population Probability 11001 11001 vector 10101 10101 10101 10001 1.0 0.5 0.5 0.0 1.0 01011 01011 11101 11000 11000 11001 Martin Pelikan, Probabilistic Model-Building GAs 14
  • 15. Probability Vector PMBGAs n  PBIL (Baluja, 1995) ¨  Incremental updates to the prob. vector. n  Compact GA (Harik, Lobo, Goldberg, 1998) ¨  Also incremental updates but better analogy with populations. n  UMDA (Mühlenbein, Paass, 1996) ¨  What we showed here. n  DEUM (Shakya et al., 2004) n  All variants perform similarly. Martin Pelikan, Probabilistic Model-Building GAs 15
  • 16. Probability Vector Dynamics n  Bitsthat perform better get more copies. n  And are combined in new ways. n  But context of each bit is ignored. n  Example problem 1: Onemax n f ( X 1 , X 2 ,… , X n ) = ∑ X i i =1 Martin Pelikan, Probabilistic Model-Building GAs 16
  • 17. Probability Vector on Onemax 1 0.9 Probability vector entries 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 0 10 20 30 40 50 Generation Martin Pelikan, Probabilistic Model-Building GAs 17
  • 18. Probability Vector: Ideal Scale-up n  O(n log n) evaluations until convergence ¨  (Harik, Cantú-Paz, Goldberg, & Miller, 1997) ¨  (Mühlenbein, Schlierkamp-Vosen, 1993) n  Other algorithms ¨  Hill climber: O(n log n) (Mühlenbein, 1992) ¨  GA with uniform: approx. O(n log n) ¨  GA with one-point: slightly slower Martin Pelikan, Probabilistic Model-Building GAs 18
  • 19. When Does Prob. Vector Fail? n  Example problem 2: Concatenated traps ¨  Partition input string into disjoint groups of 5 bits. ¨  Groups contribute via trap (ones=number of ones): " 5 if ones = 5 trap ( ones ) = # $ 4 − ones otherwise ¨  Concatenated trap = sum of single traps ¨  Optimum: String 111…1 Martin Pelikan, Probabilistic Model-Building GAs 19
  • 20. Trap-5 5 4 trap(u) 3 2 1 0 0 1 2 3 4 5 Number of ones, u Martin Pelikan, Probabilistic Model-Building GAs 20
  • 21. Probability Vector on Traps 1 0.9 Probability vector entries 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 0 10 20 30 40 50 Generation Martin Pelikan, Probabilistic Model-Building GAs 21
  • 22. Why Failure? n  Onemax: ¨  Optimum in 111…1 ¨  1 outperforms 0 on average. n  Traps: optimum in 11111, but n  f(0****) = 2 n  f(1****) = 1.375 n  So single bits are misleading. Martin Pelikan, Probabilistic Model-Building GAs 22
  • 23. How to Fix It? n  Consider 5-bit statistics instead 1-bit ones. n  Then, 11111 would outperform 00000. n  Learn model ¨  Compute p(00000), p(00001), …, p(11111) n  Sample model ¨  Sample 5 bits at a time ¨  Generate 00000 with p(00000), 00001 with p(00001), … Martin Pelikan, Probabilistic Model-Building GAs 23
  • 24. Correct Model on Traps: Dynamics 1 0.9 Probabilities of 11111 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 0 10 20 30 40 50 Generation 24 Martin Pelikan, Probabilistic Model-Building GAs
  • 25. Good News: Good Stats Work Great! n  Optimum in O(n log n) evaluations. n  Same performance as on onemax! n  Others ¨  Hill climber: O(n5 log n) = much worse. ¨  GA with uniform: O(2n) = intractable. ¨  GA with k-point xover: O(2n) (w/o tight linkage). Martin Pelikan, Probabilistic Model-Building GAs 25
  • 26. Challenge n  If we could learn and use relevant context for each position ¨  Find non-misleading statistics. ¨  Use those statistics as in probability vector. n  Then we could solve problems decomposable into statistics of order at most k with at most O(n2) evaluations! ¨  And there are many such problems (Simon, 1968). Martin Pelikan, Probabilistic Model-Building GAs 26
  • 27. What’s Next? n  COMIT ¨  Use tree models n  Extended compact GA ¨  Cluster bits into groups. n  Bayesian optimization algorithm (BOA) ¨  Use Bayesian networks (more general). Martin Pelikan, Probabilistic Model-Building GAs 27
  • 28. Beyond single bits: COMIT (Baluja, Davies, 1997) Model P(X=1) 75 % String X P(Y=1|X) 0 30 % 1 25 % X P(Z=1|X) 0 86 % 1 75 % Martin Pelikan, Probabilistic Model-Building GAs 28
  • 29. How to Learn a Tree Model? n  Mutual information: P(Xi = a, X j = b) I(Xi , X j ) = ∑ P(Xi = a, X j = b) log a,b P(Xi = a)P(X j = b) n  Goal ¨  Find tree that maximizes mutual information between connected nodes. ¨  Will minimize Kullback-Leibler divergence. n  Algorithm ¨  Prim’s algorithm for maximum spanning trees. Martin Pelikan, Probabilistic Model-Building GAs 29
  • 30. Prim’s Algorithm n  Start with a graph with no edges. n  Add arbitrary node to the tree. n  Iterate ¨  Hang a new node to the current tree. ¨  Prefer addition of edges with large mutual information (greedy approach). n  Complexity: O(n2) Martin Pelikan, Probabilistic Model-Building GAs 30
  • 31. Variants of PMBGAs with Tree Models n  COMIT (Baluja, Davies, 1997) ¨  Tree models. n  MIMIC (DeBonet, 1996) ¨  Chain distributions. n  BMDA (Pelikan, Mühlenbein, 1998) ¨  Forest distribution (independent trees or tree) Martin Pelikan, Probabilistic Model-Building GAs 31
  • 32. Beyond Pairwise Dependencies: ECGA n  Extended Compact GA (ECGA) (Harik, 1999). n  Consider groups of string positions. String Model 00 16 % 0 86 % 000 17 % 01 45 % 1 14 % 001 2 % 10 35 % ··· 11 4 % 111 24 % Martin Pelikan, Probabilistic Model-Building GAs 32
  • 33. Learning the Model in ECGA n  Start with each bit in a separate group. n  Each iteration merges two groups for best improvement. Martin Pelikan, Probabilistic Model-Building GAs 33
  • 34. How to Compute Model Quality? n  ECGA uses minimum description length. n  Minimize number of bits to store model+data: MDL( M , D) = DModel + DData n  Each frequency needs (0.5 log N) bits: |g|−1 DModel = ∑ 2 log N g ∈G n  Each solution X needs -log p(X) bits: DData = − N ∑ p( X ) log p( X ) X Martin Pelikan, Probabilistic Model-Building GAs 34
  • 35. Sampling Model in ECGA n  Sample groups of bits at a time. n  Based on observed probabilities/proportions. n  But can also apply population-based crossover similar to uniform but w.r.t. model. Martin Pelikan, Probabilistic Model-Building GAs 35
  • 36. Building-Block-Wise Mutation in ECGA n  Sastry, Goldberg (2004); Lima et al. (2005) n  Basic idea ¨  Use ECGA model builder to identify decomposition ¨  Use the best solution for BB-wise mutation ¨  For each k-bit partition (building block) n  Evaluate the remaining 2k-1 instantiations of this BB n  Use the best instantiation of this BB n  Result (for order-k separable problems) ¨  ( ) BB-wise mutation is O k log n times faster than ECGA! ¨  But only for separable problems (and similar ones). Martin Pelikan, Probabilistic Model-Building GAs 36
  • 37. What’s Next? n  We saw ¨  Probabilityvector (no edges). ¨  Tree models (some edges). ¨  Marginal product models (groups of variables). n  Next: Bayesian networks ¨  Can represent all above and more. Martin Pelikan, Probabilistic Model-Building GAs 37
  • 38. Bayesian Optimization Algorithm (BOA) n  Pelikan, Goldberg, & Cantú-Paz (1998) n  Use a Bayesian network (BN) as a model. n  Bayesian network ¨  Acyclicdirected graph. ¨  Nodes are variables (string positions). ¨  Conditional dependencies (edges). ¨  Conditional independencies (implicit). Martin Pelikan, Probabilistic Model-Building GAs 38
  • 39. Example: Bayesian Network (BN) n  Conditional dependencies. n  Conditional independencies. Martin Pelikan, Probabilistic Model-Building GAs 39
  • 40. BOA Bayesian Current Selected New network population population population Martin Pelikan, Probabilistic Model-Building GAs 40
  • 41. Learning BNs n  Two things again: ¨  Scoringmetric (as MDL in ECGA). ¨  Search procedure (in ECGA done by merging). Martin Pelikan, Probabilistic Model-Building GAs 41
  • 42. Learning BNs: Scoring Metrics n  Bayesian metrics ¨  Bayesian-Dirichlet with likelihood equivallence n Γ(m'(π i )) Γ(m'(xi , π i ) + m(xi , π i )) BD(B) = p(B)∏ ∏ ∏ Γ(m'(x ,π )) i=1 π Γ(m'(π i ) + m(π i )) xi i i i n  Minimum description length metrics ¨  Bayesian information criterion (BIC) Πi log 2 N & n# BIC( B) = ∑ % − H ( X i | Πi )N − 2 i=1 $ 2 ( ' Martin Pelikan, Probabilistic Model-Building GAs 42
  • 43. Learning BNs: Search Procedure n  Start with empty network (like ECGA). n  Execute primitive operator that improves the metric the most (greedy). n  Until no more improvement possible. n  Primitive operators ¨  Edge addition (most important). ¨  Edge removal. ¨  Edge reversal. Martin Pelikan, Probabilistic Model-Building GAs 43
  • 44. Learning BNs: Example Martin Pelikan, Probabilistic Model-Building GAs 44
  • 45. BOA and Problem Decomposition n  Conditions for factoring problem decomposition into a product of prior and conditional probabilities of small order in Mühlenbein, Mahnig, & Rodriguez (1999). n  In practice, approximate factorization sufficient that can be learned automatically. n  Learning makes complete theory intractable. Martin Pelikan, Probabilistic Model-Building GAs 45
  • 46. BOA Theory: Population Sizing n  Initial supply (Goldberg et al., 2001) ¨  Have enough stuff to combine. O 2k( ) n  Decision making (Harik et al, 1997) ¨  Decide well between competing partial sols. O ( n log n ) n  Drift (Thierens, Goldberg, Pereira, 1998) ¨  Don’t lose less salient stuff prematurely. O n () n  Model building (Pelikan et al., 2000, 2002) ¨  Find a good model. ( ) O n1.05 Martin Pelikan, Probabilistic Model-Building GAs 46
  • 47. BOA Theory: Num. of Generations n  Two extreme cases, everything in the middle. n  Uniform scaling ¨  Onemax model (Muehlenbein & Schlierkamp-Voosen, 1993) O ( n) n  Exponential scaling ¨  Domino convergence (Thierens, Goldberg, Pereira, 1998) O (n) Martin Pelikan, Probabilistic Model-Building GAs 47
  • 48. Good News: Challenge Met! n  Theory ¨  Population sizing (Pelikan et al., 2000, 2002) n  Initial supply. n  Decision making. O(n) to O(n1.05) n  Drift. n  Model building. ¨  Number of iterations (Pelikan et al., 2000, 2002) n  Uniform scaling. n  Exponential scaling. O(n0.5) to O(n) n  BOA solves order-k decomposable problems in O(n1.55) to O(n2) evaluations! Martin Pelikan, Probabilistic Model-Building GAs 48
  • 49. Theory vs. Experiment (5-bit Traps) 500000 450000 Experiment 400000 Theory 350000 Number of Evaluations 300000 250000 200000 150000 100000 100 125 150 175 200 225 250 Problem Size 49 Martin Pelikan, Probabilistic Model-Building GAs
  • 50. BOA Siblings n  Estimation of Bayesian Networks Algorithm (EBNA) (Etxeberria, Larrañaga, 1999). n  Learning Factorized Distribution Algorithm (LFDA) (Mühlenbein, Mahnig, Rodriguez, 1999). Martin Pelikan, Probabilistic Model-Building GAs 50
  • 51. Another Option: Markov Networks n  MN-FDA, MN-EDA (Santana; 2003, 2005) n  Similar to Bayes nets but with undirected edges. Martin Pelikan, Probabilistic Model-Building GAs 51
  • 52. Yet Another Option: Dependency Networks n  Estimation of dependency networks algorithm (EDNA) ¨  Gamez, Mateo, Puerta (2007). ¨  Use dependency network as a model. ¨  Dependency network learned from pairwise interactions. ¨  Use Gibbs sampling to generate new solutions. n  Dependency network ¨  Parents of a variable= all variables influencing this variable. ¨  Dependency network can contain cycles. Martin Pelikan, Probabilistic Model-Building GAs 52
  • 53. Model Comparison BMDA ECGA BOA Model Expressiveness Increases Martin Pelikan, Probabilistic Model-Building GAs 53
  • 54. From single level to hierarchy n  Single-level decomposition powerful. n  But what if single-level decomposition is not enough? n  Learn from humans and nature ¨  Decompose problem over multiple levels. ¨  Use solutions from lower level as basic building blocks. ¨  Solve problem hierarchically. Martin Pelikan, Probabilistic Model-Building GAs 54
  • 55. Hierarchical Decomposition Car Engine Braking system Electrical system Fuel system Valves Ignition system Martin Pelikan, Probabilistic Model-Building GAs 55
  • 56. Three Keys to Hierarchy Success n  Proper decomposition ¨  Must decompose problem on each level properly. n  Chunking ¨  Must represent & manipulate large order solutions. n  Preservation of alternative solutions ¨  Must preserve alternative partial solutions (chunks). Martin Pelikan, Probabilistic Model-Building GAs 56
  • 57. Hierarchical BOA (hBOA) n  Pelikan & Goldberg (2000, 2001) n  Proper decomposition ¨  Use Bayesian networks like BOA. n  Chunking ¨  Use local structures in Bayesian networks. n  Preservation of alternative solutions. ¨  Use restricted tournament replacement (RTR). ¨  Can use other niching methods. Martin Pelikan, Probabilistic Model-Building GAs 57
  • 58. Local Structures in BNs n  Look at one conditional dependency. ¨  2k probabilities for k parents. n  Why not use more powerful representations for conditional probabilities? X2X3 P(X1=0|X2X3) X1 00 26 % 01 44 % X2 X3 10 15 % 11 15 % Martin Pelikan, Probabilistic Model-Building GAs 58
  • 59. Local Structures in BNs n  Look at one conditional dependency. ¨  2k probabilities for k parents. n  Why not use more powerful representations for conditional probabilities? X2 X1 0 1 X3 15% X2 X3 0 1 26% 44% Martin Pelikan, Probabilistic Model-Building GAs 59
  • 60. Restricted Tournament Replacement n  Used in hBOA for niching. n  Insert each new candidate solution x like this: ¨  Pick random subset of original population. ¨  Find solution y most similar to x in the subset. ¨  Replace y by x if x is better than y. Martin Pelikan, Probabilistic Model-Building GAs 60
  • 61. Hierarchical Traps: The Ultimate Test n  Combine traps on more levels. n  Each level contributes to fitness. n  Groups of bits map to next level. Martin Pelikan, Probabilistic Model-Building GAs 61
  • 62. hBOA on Hierarchical Traps Experiment 6 10 O(n1.63 log(n)) Number of Evaluations 5 10 4 10 27 81 243 729 Problem Size Martin Pelikan, Probabilistic Model-Building GAs 62
  • 63. PMBGAs Are Not Just Optimizers n  PMBGAs provide us with two things ¨  Optimum or its approximation. ¨  Sequence of probabilistic models. n  Probabilistic models ¨  Encode populations of increasing quality. ¨  Tell us a lot about the problem at hand. ¨  Can we use this information? Martin Pelikan, Probabilistic Model-Building GAs 63
  • 64. Efficiency Enhancement for PMBGAs n  Sometimes O(n2) is not enough ¨  High-dimensional problems (1000s of variables) ¨  Expensive evaluation (fitness) function n  Solution ¨  Efficiency enhancement techniques Martin Pelikan, Probabilistic Model-Building GAs 64
  • 65. Efficiency Enhancement Types n  7 efficiency enhancement types for PMBGAs ¨  Parallelization ¨  Hybridization ¨  Time continuation ¨  Fitness evaluation relaxation ¨  Prior knowledge utilization ¨  Incremental and sporadic model building ¨  Learning from experience Martin Pelikan, Probabilistic Model-Building GAs 65
  • 66. Multi-objective PMBGAs n  Methods for multi-objective GAs adopted ¨  Multi-objective mixture-based IDEAs (Thierens, & Bosman, 2001) ¨  Another multi-objective BOA (from SPEA2 and mBOA) (Laumanns, & Ocenasek, 2002) ¨  Multi-objective hBOA (from NSGA-II and hBOA) (Khan, Goldberg, & Pelikan, 2002) (Pelikan, Sastry, & Goldberg, 2005) ¨  Regularity Model Based Multiobjective EDA (RM-MEDA) (Zhang, Zhou, Jin, 2008) Martin Pelikan, Probabilistic Model-Building GAs 66
  • 67. Promising Results with Discrete PMBGAs n  Artificial classes of problems n  Physics n  Bioinformatics n  Computational complexity and AI n  Others Martin Pelikan, Probabilistic Model-Building GAs 67
  • 68. Results: Artificial Problems n  Decomposition ¨  Concatenated traps (Pelikan et al., 1998). n  Hierarchical decomposition ¨  Hierarchical traps (Pelikan, Goldberg, 2001). n  Other sources of difficulty ¨  Exponential scaling, noise (Pelikan, 2002). Martin Pelikan, Probabilistic Model-Building GAs 68
  • 69. BOA on Concatenated 5-bit Traps 500000 450000 Experiment 400000 Theory 350000 Number of Evaluations 300000 250000 200000 150000 100000 100 125 150 175 200 225 250 Problem Size Martin Pelikan, Probabilistic Model-Building GAs 69
  • 70. hBOA on Hierarchical Traps Experiment 6 10 O(n1.63 log(n)) Number of Evaluations 5 10 4 10 27 81 243 729 Problem Size Martin Pelikan, Probabilistic Model-Building GAs 70
  • 71. Results: Physics n  Spin glasses (Pelikan et al., 2002, 2006, 2008) (Hoens, 2005) (Santana, 2005) (Shakya et al., 2006) ¨  ±J and Gaussian couplings ¨  2D and 3D spin glass ¨  Sherrington-Kirkpatrick (SK) spin glass n  Silicon clusters (Sastry, 2001) ¨  Gong potential (3-body) Martin Pelikan, Probabilistic Model-Building GAs 71
  • 72. hBOA on Ising Spin Glasses (2D) hBOA O(n1.51) Number ofof Evaluations Number Evaluations 3 10 64 100 144 196 256 324 400 Problem Size Number of Spins Martin Pelikan, Probabilistic Model-Building GAs 72
  • 73. Results on 2D Spin Glasses n  Number of evaluations is O(n 1.51). n  Overall time is O(n 3.51). n  Compare O(n3.51) to O(n3.5) for best method (Galluccio & Loebl, 1999) n  Great also on Gaussians. Martin Pelikan, Probabilistic Model-Building GAs 73
  • 74. hBOA on Ising Spin Glasses (3D) 6 10 Experimental average O(n3.63 ) Number of Evaluations Number of Evaluations 5 10 4 10 3 10 64 125 216 343 Problem Size Number of Spins Martin Pelikan, Probabilistic Model-Building GAs 74
  • 75. hBOA on SK Spin Glass Martin Pelikan, Probabilistic Model-Building GAs 75
  • 76. Results: Computational Complexity, AI n  MAXSAT, SAT (Pelikan, 2002) ¨  Random 3CNF from phase transition. ¨  Morphed graph coloring. ¨  Conversion from spin glass. n  Feature subset selection (Inza et al., 2001) (Cantu-Paz, 2004) Martin Pelikan, Probabilistic Model-Building GAs 76
  • 77. Results: Some Others n  Military antenna design (Santarelli et al., 2004) n  Groundwater remediation design (Arst et al., 2004) n  Forest management (Ducheyne et al., 2003) n  Nurse scheduling (Li, Aickelin, 2004) n  Telecommunication network design (Rothlauf, 2002) n  Graph partitioning (Ocenasek, Schwarz, 1999; Muehlenbein, Mahnig, 2002; Baluja, 2004) n  Portfolio management (Lipinski, 2005, 2007) n  Quantum excitation chemistry (Sastry et al., 2005) n  Maximum clique (Zhang et al., 2005) n  Cancer chemotherapy optimization (Petrovski et al., 2006) n  Minimum vertex cover (Pelikan et al., 2007) n  Protein folding (Santana et al., 2007) n  Side chain placement (Santana et al., 2007) Martin Pelikan, Probabilistic Model-Building GAs 77
  • 78. Discrete PMBGAs: Summary n  No interactions ¨  Univariate models; PBIL, UMDA, cGA. n  Some pairwise interactions ¨  Tree models; COMIT, MIMIC, BMDA. n  Multivariate interactions ¨  Multivariate models: BOA, EBNA, LFDA. n  Hierarchical decomposition ¨  hBOA Martin Pelikan, Probabilistic Model-Building GAs 78
  • 79. Discrete PMBGAs: Recommendations n  Easy problems ¨  Use univariate models; PBIL, UMDA, cGA. n  Somewhat difficult problems ¨  Use bivariate models; MIMIC, COMIT, BMDA. n  Difficult problems ¨  Use multivariate models; BOA, EBNA, LFDA. n  Most difficult problems ¨  Use hierarchical decomposition; hBOA. Martin Pelikan, Probabilistic Model-Building GAs 79
  • 80. Real-Valued PMBGAs n  New challenge ¨  Infinite domain for each variable. ¨  How to model? n  2 approaches ¨  Discretize and apply discrete model/PMBGA ¨  Create model for real-valued variables n  Estimate pdf. Martin Pelikan, Probabilistic Model-Building GAs 80
  • 81. PBIL Extensions: First Step n  SHCwL: Stochastic hill climbing with learning (Rudlof, Köppen, 1996). n  Model ¨  Single-peak Gaussian for each variable. ¨  Means evolve based on parents (promising solutions). ¨  Deviations equal, decreasing over time. n  Problems ¨  No interactions. ¨  Single Gaussians=can model only one attractor. ¨  Same deviations for each variable. Martin Pelikan, Probabilistic Model-Building GAs 81
  • 82. Use Different Deviations n  Sebag, Ducoulombier (1998) n  Some variables have higher variance. n  Use special standard deviation for each variable. Martin Pelikan, Probabilistic Model-Building GAs 82
  • 83. Use Covariance n  Covariance allows rotation of 1-peak Gaussians. n  EGNA (Larrañaga et al., 2000) n  IDEA (Bosman, Thierens, 2000) Martin Pelikan, Probabilistic Model-Building GAs 83
  • 84. How Many Peaks? n  One Gaussian vs. kernel around each point. n  Kernel distribution similar to ES. n  IDEA (Bosman, Thierens, 2000) Martin Pelikan, Probabilistic Model-Building GAs 84
  • 85. Mixtures: Between One and Many n  Mixture distributions provide transition between one Gaussian and Gaussian kernels. n  Mixture types 4 ¨  Over one variable. 2 n  Gallagher, Frean, & Downs (1999). ¨  Over all variables. 0 n  Pelikan & Goldberg (2000). -2 n  Bosman & Thierens (2000). ¨  Over partitions of variables. -4 -4 -2 0 2 4 n  Bosman & Thierens (2000). n  Ahn, Ramakrishna, and Goldberg (2004). Martin Pelikan, Probabilistic Model-Building GAs 85
  • 86. Mixed BOA (mBOA) n  Mixed BOA (Ocenasek, Schwarz, 2002) n  Local distributions ¨  A decision tree (DT) for every variable. ¨  Internal DT nodes encode tests on other variables n  Discrete:Equal to a constant n  Continuous: Less than a constant ¨  Discretevariables: DT leaves represent probabilities. ¨  Continuous variables: DT leaves contain a normal kernel distribution. Martin Pelikan, Probabilistic Model-Building GAs 86
  • 87. Real-Coded BOA (rBOA) n  Ahn, Ramakrishna, Goldberg (2003) n  Probabilistic Model ¨  Underlying structure: Bayesian network ¨  Local distributions: Mixtures of Gaussians n  Also extended to multiobjective problems (Ahn, 2005) Martin Pelikan, Probabilistic Model-Building GAs 87
  • 88. Aggregation Pheromone System (APS) n  Tsutsui (2004) n  Inspired by aggregation pheromones n  Basic idea ¨  Good solutions emit aggregation pheromones ¨  New candidate solutions based on the density of aggregation pheromones ¨  Aggregation pheromone density encodes a mixture distribution Martin Pelikan, Probabilistic Model-Building GAs 88
  • 89. Adaptive Variance Scaling n  Adaptive variance in mBOA ¨  Ocenasek et al. (2004) n  Normal IDEAs ¨  Bosman et al. (2006, 2007) ¨  Correlation-triggered adaptive variance scaling ¨  Standard-deviation ratio (SDR) triggered variance scaling Martin Pelikan, Probabilistic Model-Building GAs 89
  • 90. Real-Valued PMBGAs: Discretization n  Idea: Transform into discrete domain. n  Fixed models ¨  2k equal-width bins with k-bit binary string. ¨  Goldberg (1989). ¨  Bosman & Thierens (2000); Pelikan et al. (2003). n  Adaptive models ¨  Equal-height histograms of 2k bins. ¨  k-means clustering on each variable. ¨  Pelikan, Goldberg, & Tsutsui (2003); Cantu-Paz (2001). Martin Pelikan, Probabilistic Model-Building GAs 90
  • 91. Real-Valued PMBGAs: Summary n  Discretization ¨  Fixed ¨  Adaptive n  Real-valued models ¨  Single or multiple peaks? ¨  Same variance or different variance? ¨  Covariance or no covariance? ¨  Mixtures? ¨  Treat entire vectors, subsets of variables, or single variables? Martin Pelikan, Probabilistic Model-Building GAs 91
  • 92. Real-Valued PMBGAs: Recommendations n  Multimodality? ¨  Use multiple peaks. n  Decomposability? ¨  All variables, subsets, or single variables. n  Strong linear dependencies? ¨  Covariance. n  Partial differentiability? ¨  Combine with gradient search. Martin Pelikan, Probabilistic Model-Building GAs 92
  • 93. PMBGP (Genetic Programming) n  New challenge ¨  Structured, variable length representation. ¨  Possibly infinitely many values. ¨  Position independence (or not). ¨  Low correlation between solution quality and solution structure (Looks, 2006). n  Approaches ¨  Use explicit probabilistic models for trees. ¨  Use models based on grammars. Martin Pelikan, Probabilistic Model-Building GAs 93
  • 94. PIPE n  Probabilistic incremental program evolution (Salustowicz & X P(X) Schmidhuber, 1997) sin 0.15 n  Store frequencies of + 0.35 operators/terminals in - 0.35 nodes of a maximum tree. X 0.15 n  Sampling generates tree from top to bottom Martin Pelikan, Probabilistic Model-Building GAs 94
  • 95. eCGP n  Sastry & Goldberg (2003) n  ECGA adapted to program trees. n  Maximum tree as in PIPE. n  But nodes partitioned into groups. Martin Pelikan, Probabilistic Model-Building GAs 95
  • 96. BOA for GP n  Looks, Goertzel, & Pennachin (2004) n  Combinatory logic + BOA ¨  Trees translated into uniform structures. ¨  Labels only in leaves. ¨  BOA builds model over symbols in different nodes. n  Complexity build-up ¨  Modeling limited to max. sized structure seen. ¨  Complexity builds up by special operator. Martin Pelikan, Probabilistic Model-Building GAs 96
  • 97. MOSES n  Looks (2006). n  Evolve demes of programs. n  Each deme represents similar structures. n  Apply PMBGA to each deme (e.g. hBOA). n  Introduce new demes/delete old ones. n  Use normal forms to reduce complexity. Martin Pelikan, Probabilistic Model-Building GAs 97
  • 98. PMBGP with Grammars n  Use grammars/stochastic grammars as models. n  Grammars restrict the class of programs. n  Some representatives ¨  Program evolution with explicit learning (Shan et al., 2003) ¨  Grammar-based EDA for GP (Bosman, de Jong, 2004) ¨  Stochastic grammar GP (Tanev, 2004) ¨  Adaptive constrained GP (Janikow, 2004) Martin Pelikan, Probabilistic Model-Building GAs 98
  • 99. PMBGP: Summary n  Interesting starting points available. n  But still lot of work to be done. n  Much to learn from discrete domain, but some completely new challenges. n  Research in progress Martin Pelikan, Probabilistic Model-Building GAs 99
  • 100. PMBGAs for Permutations n  New challenges ¨  Relativeorder ¨  Absolute order ¨  Permutation constraints n  Two basic approaches ¨  Random-key and real-valued PMBGAs ¨  Explicit probabilistic models for permutations Martin Pelikan, Probabilistic Model-Building GAs 100
  • 101. Random Keys and PMBGAs n  Bengoetxea et al. (2000); Bosman et al. (2001) n  Random keys (Bean, 1997) ¨  Candidate solution = vector of real values ¨  Ascending ordering gives a permutation n  Can use any real-valued PMBGA (or GEA) ¨  IDEAs (Bosman, Thierens, 2002) ¨  EGNA (Larranaga et al., 2001) n  Strengths and weaknesses ¨  Good: Can use any real-valued PMBGA. ¨  Bad: Redundancy of the encoding. Martin Pelikan, Probabilistic Model-Building GAs 101
  • 102. Direct Modeling of Permutations n  Edge-histogram based sampling algorithm (EHBSA) (Tsutsui, Pelikan, Goldberg, 2003) ¨  Permutations of n elements ¨  Model is a matrix A=(ai,j)i,j=1, 2, …, n ¨  ai,j represents the probability of edge (i, j) ¨  Uses template to reduce exploration ¨  Applicable also to scheduling Martin Pelikan, Probabilistic Model-Building GAs 102
  • 103. ICE: Modify Crossover from Model n  ICE ¨  Bosman, Thierens (2001). ¨  Represent permutations with random keys. ¨  Learn multivariate model to factorize the problem. ¨  Use the learned model to modify crossover. n  Performance ¨  Typically outperforms IDEAs and other PMBGAs that learn and sample random keys. Martin Pelikan, Probabilistic Model-Building GAs 103
  • 104. Multivariate Permutation Models n  Basic approach ¨  Use any standard multivariate discrete model. ¨  Restrict sampling to permutations in some way. ¨  Bengoetxea et al. (2000), Pelikan et al. (2007). n  Strengths and weaknesses ¨  Use explicit multivariate models to find regularities. ¨  High-order alphabet requires big samples for good models. ¨  Sampling can introduce unwanted bias. ¨  Inefficient encoding for only relative ordering constraints, which can be encoded simpler. Martin Pelikan, Probabilistic Model-Building GAs 104
  • 105. Conclusions n  Competent PMBGAs exist ¨  Scalable solution to broad classes of problems. ¨  Solution to previously intractable problems. ¨  Algorithms ready for new applications. n  PMBGAs do more than just solve the problem ¨  They provide us with sequences of probabilistic models. ¨  The probabilistic models tell us a lot about the problem. n  Consequences for practitioners ¨  Robust methods with few or no parameters. ¨  Capable of learning how to solve problem. ¨  But can incorporate prior knowledge as well. ¨  Can solve previously intractable problems. Martin Pelikan, Probabilistic Model-Building GAs 105
  • 106. Starting Points n  World wide web n  Books and surveys ¨  Larrañaga & Lozano (eds.) (2001). Estimation of distribution algorithms: A new tool for evolutionary computation. Kluwer. ¨  Pelikan et al. (2002). A survey to optimization by building and using probabilistic models. Computational optimization and applications, 21(1), pp. 5-20. ¨  Pelikan (2005). Hierarchical BOA: Towards a New Generation of Evolutionary Algorithms. Springer. ¨  Lozano, Larrañaga, Inza, Bengoetxea (2006). Towards a New Evolutionary Computation: Advances on Estimation of Distribution Algorithms, Springer. ¨  Pelikan, Sastry, Cantu-Paz (eds.) (2006). Scalable Optimization via Probabilistic Modeling: From Algorithms to Applications, Springer. Martin Pelikan, Probabilistic Model-Building GAs 106
  • 107. Online Code (1/2) n  BOA, BOA with decision graphs, dependency-tree EDA http://medal-lab.org/ n  ECGA, xi-ary ECGA, BOA, and BOA with decision trees/graphs http://www.illigal.org/ n  mBOA http://jiri.ocenasek.com/ n  PIPE http://www.idsia.ch/~rafal/ n  Real-coded BOA http://www.evolution.re.kr/ Martin Pelikan, Probabilistic Model-Building GAs 107
  • 108. Online Code (2/2) n  Demos of APS and EHBSA http://www.hannan-u.ac.jp/~tsutsui/research-e.html n  RM-MEDA: A Regularity Model Based Multiobjective EDA Differential Evolution + EDA hybrid http://cswww.essex.ac.uk/staff/qzhang/mypublication.htm n  Naive Multi-objective Mixture-based IDEA (MIDEA) Normal IDEA-Induced Chromosome Elements Exchanger (ICE) Normal Iterated Density-Estimation Evolutionary Algorithm (IDEA) http://homepages.cwi.nl/~bosman/code.html Martin Pelikan, Probabilistic Model-Building GAs 108