SlideShare ist ein Scribd-Unternehmen logo
1 von 53
Downloaden Sie, um offline zu lesen
Lyapunov functions, value functions,
               and performance bounds
Sean Meyn

Department of Electrical and Computer Engineering
University of Illinois
and the Coordinated Science Laboratory




   Joint work with R. Tweedie, I. Kontoyiannis, and P. Mehta
   Supported in part by NSF (ECS 05 23620, and prior funding), and AFOSR
Objectives


Nonlinear state space model ≡ (controlled) Markov process,
                                                  state process X


Typical form:

     dX(t) = f (X(t), U (t)) dt + σ(X(t), U (t)) dW (t)
                                                          noise
                              control
Objectives


Nonlinear state space model ≡ (controlled) Markov process,
                                                  state process X
Typical form:
     dX(t) = f (X(t), U (t)) dt + σ(X(t), U (t)) dW (t)
                                                          noise
                                  control


Questions: For a given feedback law,
• Is the state process stable?
• Is the average cost finite?    E[c(X(t), U (t))]
• Can we solve the DP equations?       min c(x, u) + Du h∗ (x) = η ∗
                                            u
• Can we approximate the average cost η∗? The value function h∗ ?
Outline


Markov Models                 P t (x, · ) − π   f   →0




                                                         sup Ex [SτC (f )] < ∞
                                                         C
                  π(f ) < ∞
Representations


Lyapunov Theory           DV (x) ≤ −f (x) + bIC (x)




Conclusions
I
Markov Models
Notation


Markov chain: X = X(t) : t ≥ 0
Countable state space, X
Transition semigroup,
     P t (x, y) = P X(s + t) = y X(s) = x ,   x, y ∈ X
Notation: Generators & Resolvents


Markov chain: X = X(t) : t ≥ 0
Countable state space, X
Transition semigroup,
     P t (x, y) = P X(s + t) = y X(s) = x ,   x, y ∈ X

Generator: For some domain of functions h,
                     1
      Dh (x) = lim E[h(X(s + t)) − h(X(s)) X(s) = x]
                 t→0 t

                    1 t
              = lim (P h (x) − h(x))
                t→0 t
Notation: Generators & Resolvents


Generator: For some domain of functions h,
                     1
      Dh (x) = lim E[h(X(s + t)) − h(X(s)) X(s) = x]
                 t→0 t

                     1 t
               = lim (P h (x) − h(x))
                 t→0 t

Rate matrix:
       Dh (x) =       Q(x, y)h(y)   P t = eQt
                  y
α             µ
Example: MM1 Queue

                          
                          x + 1 Prob εα
                          
Sample paths:   X(t + ε) ≈ x − 1 Prob εµ
                          
                          
                            x    Prob 1 − ε(α + µ)


Rate matrix:
                                                           
     −α      α   0      0      0      0                  ···
    µ −α − µ
                α      0      0      0                  · · ·
                                                              
    0
            µ −α − µ   α      0      0                  · · ·
                                                              
   
Q= 0        0   µ    −α − µ   α      0                  · · ·
                                                              
    0
            0   0      µ    −α − µ   α                  · · ·
                                                              
    0
            0   0      0      µ    −α − µ               · · ·
                                                              
      .
      .      .
             .   .
                 .      .
                        .      .
                               .      .
                                      .
      .      .   .      .      .      .
2        2
                                               σW = 0   σW = 1




Example: O-U Model



Sample paths:    dX(t) = AX(t) dt + B dW (t)
                                A n × n, Bn × 1, W standard BM

                                        2
Generator: Dh (x) = (Ax)T h (x) + B T       h (x)B
2        2
                                                       σW = 0   σW = 1




Example: O-U Model



Sample paths:         dX(t) = AX(t) dt + B dW (t)
                                      A n × n, Bn × 1, W standard BM

                                                2
Generator: Dh (x) = (Ax)T h (x) + B T               h (x)B


h quadratic,      h(x) = 1 xT P x
                         2               h (x) = P x
                                         2
                                             h (x) = P


               Dh (x) = 1 xT (P A + AT P )x + B T P B
                        2
Notation: Generators & Resolvents


Generator: For some domain of functions h,
                     1
      Dh (x) = lim E[h(X(s + t)) − h(X(s)) X(s) = x]
                 t→0 t

                           1 t
                     = lim (P h (x) − h(x))
                       t→0 t

Rate matrix:
       Dh (x) =              Q(x, y)h(y)   P t = eQt
                         y

Resolvent:
                 ∞
   Rα =              e−αt P t
             0
Notation: Generators & Resolvents


Generator: For some domain of functions h,
                     1
      Dh (x) = lim E[h(X(s + t)) − h(X(s)) X(s) = x]
                 t→0 t

                           1 t
                     = lim (P h (x) − h(x))
                       t→0 t

Rate matrix:
       Dh (x) =              Q(x, y)h(y)        P t = eQt
                         y

Resolvent:                                 Resolvent equations:
                 ∞
   Rα =              e−αt P t              Rα = [ Iα − Q]−1
             0
                                            QRα = Rα Q = αRα − I
Notation: Generators & Resolvents


Motivation: Dynamic programming. For a cost function c,

hα (x) = Rα c (x) =            Rα (x, y)c(y)
                         y∈X
               ∞
       =           eαt E[c(X(t)) X(0) = x] dt
           0
                                               Discounted-cost value function
Notation: Generators & Resolvents


Motivation: Dynamic programming. For a cost function c,

hα (x) = Rα c (x) =            Rα (x, y)c(y)
                         y∈X
               ∞
       =           eαt E[c(X(t)) X(0) = x] dt
           0
                                               Discounted-cost value function


Resolvent equation = dynamic programming equation,

                          c + Dhα      = αhα
Notation: Steady State Distribution


Invariant (probability) measure π: X is stationary. In particular,

                                         X(t) ∼ π,       t≥0
Notation: Steady State Distribution


Invariant (probability) measure π: X is stationary. In particular,

                                              X(t) ∼ π,     t≥0


Characterizations:

                     π(x)P t (x, y)   =   π(y)
               x∈X

           α         π(x)Rα (x, y)    =   π(y), α > 0
               x∈X

                      π(x)Q(x, y)     =   0
                x∈X                                  y ∈X
Notation: Relative Value Function


Invariant measure π, cost function c , steady-state mean η

Relative value function:
                        ∞
          h(x) =            E[c(X(t)) − η X(0) = x] dt
                    0
Notation: Relative Value Function


Invariant measure π, cost function c , steady-state mean η

Relative value function:
                        ∞
          h(x) =            E[c(X(t)) − η X(0) = x] dt
                    0


Solution to Poisson’s equation (average-cost DP equation):

                             c + Dh = η
II
    Representations
π    ∝ ν[I − (R − s ⊗ ν)]−1

h    = [I − (R − s ⊗ ν)]−1 c
                           ˜
Irreducibility


ψ-Irreducibility:
        ψ(y) > 0 =⇒ P X(t) reaches y X(0) = x > 0 all x

        ψ(y) > 0 =⇒ R(x, y) > 0 all x
Small Functions and Small Measures


ψ-Irreducibility:
        ψ(y) > 0 =⇒ P X(t) reaches y X(0) = x > 0 all x

        ψ(y) > 0 =⇒ R(x, y) > 0 all x

Small functions and measures: For a function s and probability ν,
                R(x, y) ≥ s(x)ν(y),     x, y ∈ X             ∞
                                                    R=           e−t P t dt
                                                         0
Small Functions and Small Measures


ψ-Irreducibility:
        ψ(y) > 0 =⇒ P X(t) reaches y X(0) = x > 0 all x

        ψ(y) > 0 =⇒ R(x, y) > 0 all x

Small functions and measures: For a function s and probability ν,
                R(x, y) ≥ s(x)ν(y),     x, y ∈ X
Resolvent dominates rank-one matrix,                  R=
                                                               ∞
                                                                   e−t P t dt
                                                           0
                     R ≥s⊗ν
Small Functions and Small Measures


ψ-Irreducibility:
        ψ(y) > 0 =⇒ P X(t) reaches y X(0) = x > 0 all x

        ψ(y) > 0 =⇒ R(x, y) > 0 all x

Small functions and measures: For a function s and probability ν,
                R(x, y) ≥ s(x)ν(y),     x, y ∈ X
Resolvent dominates rank-one matrix,                       R=
                                                                    ∞
                                                                        e−t P t dt
                                                                0
                     R ≥s⊗ν
ψ-Irreducibility justi es assumption: s(x) > 0 for all x
                         and WLOG,     ν = δx ∗ ,   where ψ(x∗ ) > 0
α           µ
Example: MM1 Queue


R(x, y) > 0 for all x and y     (irreducible in usual sense)



Conclusion:
                   R(x, y ) ≥ s(x)ν(y )

                where s(x) := R(x, 0)
                              ν := δ0
2              2
                                                         σW = 0         σW = 1




Example: O-U Model
                                                       dX(t) = AX(t) dt + B dW (t)




R(0, . ) Gaussian
         Full rank if and only if (A, B) is controllable.

Conclusion: Under controllability, for any m, there is ε s.t.,


               R(x, A) ≥ s(x)ν(A)        all x and A


           where      s(x) = ε I    x ≤m
                     ν(A) uniform on         x ≤m
Potential Matrix

                              ∞
Potential matrix: G(x, y) =         (R − s ⊗ ν)n (x, y)
                              n=0



                   G = [I − (R − s ⊗ ν)]−1
Representation of π

                              ∞
Potential matrix: G(x, y) =         (R − s ⊗ ν)n (x, y)
                              n=0




                     π ∝ νG
                                            νG (y) =         ν(x)G(x, y)
                                                       x∈X
Representation of h

                              ∞
Potential matrix: G(x, y) =         (R − s ⊗ ν)n (x, y)
                              n=0




               h = RG c
                      ˜                 + constant

               c(x) = c(x) − η
               ˜
                                            G˜ (y) =
                                             c               G(x, y)˜(y)
                                                                    c
                 η=         π(x)c(x)                   y∈X
                      y∈X
Representation of h

                               ∞
Potential matrix: G(x, y) =         (R − s ⊗ ν)n (x, y)
                              n=0




                h = RG c
                       ˜                + constant

                c(x) = c(x) − η
                ˜
                                            G˜ (y) =
                                             c               G(x, y)˜(y)
                                                                    c
                  η=         π(x)c(x)                  y∈X
                       y∈X


If sum converges, then Poisson’s equation is solved:
                       c(x) + Dh (x) = η
III
Lyapunov Theory
              P n (x, · ) − π   f   →0




                                         sup Ex [SτC (f )] < ∞
                                         C
  π(f ) < ∞




          ∆V (x) ≤ −f (x) + bIC (x)
Lyapunov Functions



            DV ≤ −g + bs
Lyapunov Functions



               DV ≤ −g + bs

General assumptions: V : X → (0,∞)
                      g : X → [1, ∞)
                      b < ∞, s small
                             e.g., s (x) = IC (x), C nite
Lyapunov Bounds on G                   DV ≤ −g + bs


Resolvent equation gives RV − V ≤ −Rg + bRs
Lyapunov Bounds on G                   DV ≤ −g + bs


Resolvent equation gives RV − V ≤ −Rg + bRs
Since s⊗ν is non-negative,
  −[I − (R − s ⊗ ν)]V ≤ RV − V ≤ −Rg + bRs

          G−1
Lyapunov Bounds on G                       DV ≤ −g + bs


Resolvent equation gives RV − V ≤ −Rg + bRs
Since s⊗ν is non-negative,
  −[I − (R − s ⊗ ν)]V ≤ RV − V ≤ −Rg + bRs

           G−1

More positivity,    V ≥ GRg − bGRs

Some algebra,      GR = G(R − s ⊗ ν) + (Gs) ⊗ ν ≥ G − I
                   Gs ≤ 1
Lyapunov Bounds on G                       DV ≤ −g + bs


Resolvent equation gives RV − V ≤ −Rg + bRs
Since s⊗ν is non-negative,
  −[I − (R − s ⊗ ν)]V ≤ RV − V ≤ −Rg + bRs

           G−1

More positivity,    V ≥ GRg − bGRs

Some algebra,      GR = G(R − s ⊗ ν) + (Gs) ⊗ ν ≥ G − I
                   Gs ≤ 1

General bound:       GRg     ≤ V + 2b
                      Gg     ≤ V + g + 2b
Existence of π                              DV ≤ −g + bs


Condition (V2)    DV ≤ −1 + bs


Representation: π ∝ νG

Bound:   GRg ≤ V + 2b =⇒ G(x, X) ≤ V (x) + 2b

Conclusion: π exists as a probability measure on X
Existence of moments            DV ≤ −g + bs


Condition (V3)   DV ≤ −g + bs


Representation: π ∝ νG

Bound:   Gg ≤ V + g + 2b
Existence of moments                            DV ≤ −g + bs


Condition (V3)    DV ≤ −g + bs


Representation: π ∝ νG

Bound:   Gg ≤ V + g + 2b

Conclusion: π exists as a probability measure on X
                 and the steady-state mean is nite,

                       π(g) :=         π(x)g(x) ≤ b
                                 x∈X
α                      α   µ
Example: MM1 Queue ρ =
                       µ


Linear Lyapunov function, V (x) = x
                ∞
     DV (x) =         Q(x, y)y
                y=0

             = α(x + 1) + µ(x − 1) − (α + µ)x

             = −(µ − α)          x>0



Conclusion: (V2) holds if and only if ρ < 1
α                   α           µ
Example: MM1 Queue ρ =
                       µ


QuadraticLyapunov function, V (x) = x 2
               ∞
    DV (x) =         Q(x, y)y 2
               y=0

            = α(x + 1)2 + µ(x − 1)2 − (α + µ)x2
            = α(x2 + 2x + 1) + µ(x2 − 2x + 1)2 − (α + µ)x2
            = −2(µ − α)x + α + µ


Conclusion: (V3) holds, g(x) = 1 + x
                 if and only if ρ < 1
2           2
                                                         σW = 0      σW = 1




Example: O-U Model
                                                    dX(t) = AX(t) dt + B dW (t)



h quadratic,      h(x) = 1 xT P x
                         2               h (x) = P x
                                         2
                                             h (x) = P


               Dh (x) = 1 xT (P A + AT P )x + B T P B
                        2


Suppose that P > 0 solves the Lyapunov equation,

                         P A + AT P = -I
2           2
                                                         σW = 0      σW = 1




Example: O-U Model
                                                    dX(t) = AX(t) dt + B dW (t)



h quadratic,      h(x) = 1 xT P x
                         2               h (x) = P x
                                         2
                                             h (x) = P


               Dh (x) = 1 xT (P A + AT P )x + B T P B
                        2


Suppose that P > 0 solves the Lyapunov equation,

                         P A + AT P = -I

Then (V3) follows from the identity,
                         2      2       2
      Dh (x) = − 1 x
                 2           + σX ,    σX = B T P B
2              2
                                                            σW = 0         σW = 1




Example: O-U Model
                                                          dX(t) = AX(t) dt + B dW (t)



The function h(x) = 1 xT P x solves Poisson’s equation,
                    2


                                              1       2
            Dh = −g + η              g(x) =   2   x
                                            2
                                       η = σX


Suppose that P > 0 solves the Lyapunov equation,

                       P A + AT P = -I

Then (V3) follows from the identity,
                        2      2         2
      Dh (x) = − 1 x
                 2          + σX ,      σX = B T P B
Poisson’s Equation                        DV ≤ −g + bs


Condition (V3)    DV ≤ −g + bs


Representation:   h = RG c
                         ˜   + constant     c(x) = c(x) − η
                                            ˜

Bound: RGg ≤ V + 2b =⇒ RGg (x) ≤ V (x) + 2b
Poisson’s Equation                           DV ≤ −g + bs


Condition (V3)      DV ≤ −g + bs


Representation:    h = RG c
                          ˜   + constant         c(x) = c(x) − η
                                                 ˜

Bound: RGg ≤ V + 2b =⇒ RGg (x) ≤ V (x) + 2b


Conclusion: If c is bounded by g, then h is bounded,

                  h(x) ≤ V (x) + 2b
α                         α     µ
Example: MM1 Queue ρ =
                       µ


Poisson’s equation with g (x) = x

                       Dh = −g + η


We have (V3) with V a quadratic function of x:

Recall, with h (x) = x 2

           Dh (x) = −2(µ − α)x + α + µ           x>0
α                α   µ
Example: MM1 Queue ρ =
                       µ


Poisson’s equation with g (x) = x

                      Dh = −g + η
Solved with

                   x2 + x          ρ
          h(x) = 1
                 2 µ−α         η=
                                  1−ρ
IV
Conclusions
P t (x, · ) − π   f   →0




                                                                                            sup Ex [SτC (f )] < ∞
                                                                                            C
Final words




                                                     π(f ) < ∞
                                                             DV (x) ≤ −f (x) + bIC (x)




Just as in linear systems theory, Lyapunov functions
provide a characterization of system properties, as well
as a practical verification tool
P t (x, · ) − π   f   →0




                                                                                             sup Ex [SτC (f )] < ∞
                                                                                             C
 Final words




                                                      π(f ) < ∞
                                                              DV (x) ≤ −f (x) + bIC (x)




 Just as in linear systems theory, Lyapunov functions
 provide a characterization of system properties, as well
 as a practical verification tool

 Much is left out of this survey - in particular,

• Converse theory
• Limit theory
• Approximation techniques to construct Lyapunov functions
          or approximations to value functions
• Application to controlled Markov processes, and
          approximate dynamic programming
References
 [1,4] ψ-Irreducible foundations
 [2,11,12,13] Mean- eld models, ODE models, and Lyapunov functions
 [1,4,5,9,10] Operator-theoretic methods.                             See also appendix of [2]
 [3,6,7,10] Generators and continuous time models

[1] S. P. Meyn and R. L. Tweedie. Markov chains and stochastic    [9] I. Kontoyiannis and S. P. Meyn. Spectral theory and limit
    stability. Cambridge University Press, Cambridge, second           theorems for geometrically ergodic Markov processes. Ann.
    edition, 2009. Published in the Cambridge Mathematical             Appl. Probab., 13:304–362, 2003. Presented at the INFORMS
    Library.                                                           Applied Probability Conference, NYC, July, 2001.
[2] S. P. Meyn. Control Techniques for Complex Networks. Cam-     [10] I. Kontoyiannis and S. P. Meyn. Large deviations asymptotics
    bridge University Press, Cambridge, 2007. Pre-publication          and the spectral theory of multiplicatively regular Markov
    edition online: http://black.csl.uiuc.edu/˜meyn.                   processes. Electron. J. Probab., 10(3):61–123 (electronic),
[3] S. N. Ethier and T. G. Kurtz. Markov Processes : Charac-           2005.
    terization and Convergence. John Wiley & Sons, New York,      [11] W. Chen, D. Huang, A. Kulkarni, J. Unnikrishnan, Q. Zhu,
    1986.                                                              P. Mehta, S. Meyn, and A. Wierman. Approximate dynamic
[4] E. Nummelin. General Irreducible Markov Chains and Non-            programming using fluid and diffusion approximations with
    negative Operators. Cambridge University Press, Cambridge,         applications to power management. Accepted for inclusion in
    1984.                                                              the 48th IEEE Conference on Decision and Control, December
[5] S. P. Meyn and R. L. Tweedie. Generalized resolvents               16-18 2009.
    and Harris recurrence of Markov processes. Contemporary       [12] P. Mehta and S. Meyn. Q-learning and Pontryagin’s Minimum
    Mathematics, 149:227–250, 1993.                                    Principle. Accepted for inclusion in the 48th IEEE Conference
[6] S. P. Meyn and R. L. Tweedie. Stability of Markovian               on Decision and Control, December 16-18 2009.
    processes III: Foster-Lyapunov criteria for continuous time   [13] G. Fort, S. Meyn, E. Moulines, and P. Priouret. ODE
    processes. Adv. Appl. Probab., 25:518–548, 1993.                   methods for skip-free Markov chain stability with applications
[7] D. Down, S. P. Meyn, and R. L. Tweedie. Exponential                to MCMC. Ann. Appl. Probab., 18(2):664–707, 2008.
    and uniform ergodicity of Markov processes. Ann. Probab.,
    23(4):1671–1691, 1995.
[8] P. W. Glynn and S. P. Meyn. A Liapounov bound for solutions
    of the Poisson equation. Ann. Probab., 24(2):916–931, 1996.

                                                                   See also earlier seminal work by Hordijk, Tweedie, ... full references in [1].

Weitere ähnliche Inhalte

Was ist angesagt?

Auc silver spring
Auc silver springAuc silver spring
Auc silver spring
Vivian Stg
 
Runtime Analysis of Population-based Evolutionary Algorithms
Runtime Analysis of Population-based Evolutionary AlgorithmsRuntime Analysis of Population-based Evolutionary Algorithms
Runtime Analysis of Population-based Evolutionary Algorithms
PK Lehre
 
Bayesian Methods for Machine Learning
Bayesian Methods for Machine LearningBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning
butest
 
11.[104 111]analytical solution for telegraph equation by modified of sumudu ...
11.[104 111]analytical solution for telegraph equation by modified of sumudu ...11.[104 111]analytical solution for telegraph equation by modified of sumudu ...
11.[104 111]analytical solution for telegraph equation by modified of sumudu ...
Alexander Decker
 
Sampling strategies for Sequential Monte Carlo (SMC) methods
Sampling strategies for Sequential Monte Carlo (SMC) methodsSampling strategies for Sequential Monte Carlo (SMC) methods
Sampling strategies for Sequential Monte Carlo (SMC) methods
Stephane Senecal
 

Was ist angesagt? (19)

Auc silver spring
Auc silver springAuc silver spring
Auc silver spring
 
Mesh Processing Course : Geodesics
Mesh Processing Course : GeodesicsMesh Processing Course : Geodesics
Mesh Processing Course : Geodesics
 
Runtime Analysis of Population-based Evolutionary Algorithms
Runtime Analysis of Population-based Evolutionary AlgorithmsRuntime Analysis of Population-based Evolutionary Algorithms
Runtime Analysis of Population-based Evolutionary Algorithms
 
Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)
Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)
Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)
 
Slides euria-1
Slides euria-1Slides euria-1
Slides euria-1
 
Bayesian Methods for Machine Learning
Bayesian Methods for Machine LearningBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning
 
Chapter 5 (maths 3)
Chapter 5 (maths 3)Chapter 5 (maths 3)
Chapter 5 (maths 3)
 
Low Complexity Regularization of Inverse Problems - Course #3 Proximal Splitt...
Low Complexity Regularization of Inverse Problems - Course #3 Proximal Splitt...Low Complexity Regularization of Inverse Problems - Course #3 Proximal Splitt...
Low Complexity Regularization of Inverse Problems - Course #3 Proximal Splitt...
 
Rousseau
RousseauRousseau
Rousseau
 
1
11
1
 
YSC 2013
YSC 2013YSC 2013
YSC 2013
 
Lesson 26: The Fundamental Theorem of Calculus (slides)
Lesson 26: The Fundamental Theorem of Calculus (slides)Lesson 26: The Fundamental Theorem of Calculus (slides)
Lesson 26: The Fundamental Theorem of Calculus (slides)
 
11.[104 111]analytical solution for telegraph equation by modified of sumudu ...
11.[104 111]analytical solution for telegraph equation by modified of sumudu ...11.[104 111]analytical solution for telegraph equation by modified of sumudu ...
11.[104 111]analytical solution for telegraph equation by modified of sumudu ...
 
Sampling strategies for Sequential Monte Carlo (SMC) methods
Sampling strategies for Sequential Monte Carlo (SMC) methodsSampling strategies for Sequential Monte Carlo (SMC) methods
Sampling strategies for Sequential Monte Carlo (SMC) methods
 
Beam theory
Beam theoryBeam theory
Beam theory
 
Signal Processing Course : Convex Optimization
Signal Processing Course : Convex OptimizationSignal Processing Course : Convex Optimization
Signal Processing Course : Convex Optimization
 
Slides euria-2
Slides euria-2Slides euria-2
Slides euria-2
 
Prml
PrmlPrml
Prml
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...
 

Andere mochten auch

2012 Tutorial: Markets for Differentiated Electric Power Products
2012 Tutorial:  Markets for Differentiated Electric Power Products2012 Tutorial:  Markets for Differentiated Electric Power Products
2012 Tutorial: Markets for Differentiated Electric Power Products
Sean Meyn
 
Control Synthesis by Sum of Squares Optimization
Control Synthesis by Sum of Squares OptimizationControl Synthesis by Sum of Squares Optimization
Control Synthesis by Sum of Squares Optimization
Behzad Samadi
 

Andere mochten auch (9)

Lyapunov theory
Lyapunov theoryLyapunov theory
Lyapunov theory
 
2012 Tutorial: Markets for Differentiated Electric Power Products
2012 Tutorial:  Markets for Differentiated Electric Power Products2012 Tutorial:  Markets for Differentiated Electric Power Products
2012 Tutorial: Markets for Differentiated Electric Power Products
 
Lyapunov stability
Lyapunov stability Lyapunov stability
Lyapunov stability
 
Control Synthesis by Sum of Squares Optimization
Control Synthesis by Sum of Squares OptimizationControl Synthesis by Sum of Squares Optimization
Control Synthesis by Sum of Squares Optimization
 
Lyapunov stability
Lyapunov stabilityLyapunov stability
Lyapunov stability
 
sistem non linier Lyapunov Stability
sistem non linier Lyapunov Stabilitysistem non linier Lyapunov Stability
sistem non linier Lyapunov Stability
 
Lyapunov Stability Rizki Adi Nugroho [1410501075]
Lyapunov Stability   Rizki Adi Nugroho [1410501075]Lyapunov Stability   Rizki Adi Nugroho [1410501075]
Lyapunov Stability Rizki Adi Nugroho [1410501075]
 
Tugas non linear lyapunov stability
Tugas non linear lyapunov stabilityTugas non linear lyapunov stability
Tugas non linear lyapunov stability
 
Lyapunov stability
Lyapunov stability Lyapunov stability
Lyapunov stability
 

Ähnlich wie Markov Tutorial CDC Shanghai 2009

A numerical method to solve fractional Fredholm-Volterra integro-differential...
A numerical method to solve fractional Fredholm-Volterra integro-differential...A numerical method to solve fractional Fredholm-Volterra integro-differential...
A numerical method to solve fractional Fredholm-Volterra integro-differential...
OctavianPostavaru
 
Parameterized curves in r^3
Parameterized curves in r^3Parameterized curves in r^3
Parameterized curves in r^3
Tarun Gehlot
 
Case Study (All)
Case Study (All)Case Study (All)
Case Study (All)
gudeyi
 
Stochastic Control and Information Theoretic Dualities (Complete Version)
Stochastic Control and Information Theoretic Dualities (Complete Version)Stochastic Control and Information Theoretic Dualities (Complete Version)
Stochastic Control and Information Theoretic Dualities (Complete Version)
Haruki Nishimura
 
Adaptive dynamic programming for control
Adaptive dynamic programming for controlAdaptive dynamic programming for control
Adaptive dynamic programming for control
Springer
 
2D1431 Machine Learning
2D1431 Machine Learning2D1431 Machine Learning
2D1431 Machine Learning
butest
 
Emat 213 study guide
Emat 213 study guideEmat 213 study guide
Emat 213 study guide
akabaka12
 
On problem-of-parameters-identification-of-dynamic-object
On problem-of-parameters-identification-of-dynamic-objectOn problem-of-parameters-identification-of-dynamic-object
On problem-of-parameters-identification-of-dynamic-object
Cemal Ardil
 
12.5. vector valued functions
12.5. vector valued functions12.5. vector valued functions
12.5. vector valued functions
math267
 

Ähnlich wie Markov Tutorial CDC Shanghai 2009 (20)

2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...
2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...
2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...
 
Levitan Centenary Conference Talk, June 27 2014
Levitan Centenary Conference Talk, June 27 2014Levitan Centenary Conference Talk, June 27 2014
Levitan Centenary Conference Talk, June 27 2014
 
Comparison Theorems for SDEs
Comparison Theorems for SDEs Comparison Theorems for SDEs
Comparison Theorems for SDEs
 
A numerical method to solve fractional Fredholm-Volterra integro-differential...
A numerical method to solve fractional Fredholm-Volterra integro-differential...A numerical method to solve fractional Fredholm-Volterra integro-differential...
A numerical method to solve fractional Fredholm-Volterra integro-differential...
 
Real Time Code Generation for Nonlinear Model Predictive Control
Real Time Code Generation for Nonlinear Model Predictive ControlReal Time Code Generation for Nonlinear Model Predictive Control
Real Time Code Generation for Nonlinear Model Predictive Control
 
Parameterized curves in r^3
Parameterized curves in r^3Parameterized curves in r^3
Parameterized curves in r^3
 
Case Study (All)
Case Study (All)Case Study (All)
Case Study (All)
 
Assignment6
Assignment6Assignment6
Assignment6
 
cswiercz-general-presentation
cswiercz-general-presentationcswiercz-general-presentation
cswiercz-general-presentation
 
Stochastic Control and Information Theoretic Dualities (Complete Version)
Stochastic Control and Information Theoretic Dualities (Complete Version)Stochastic Control and Information Theoretic Dualities (Complete Version)
Stochastic Control and Information Theoretic Dualities (Complete Version)
 
residue
residueresidue
residue
 
Adaptive dynamic programming for control
Adaptive dynamic programming for controlAdaptive dynamic programming for control
Adaptive dynamic programming for control
 
Chapter3 laplace
Chapter3 laplaceChapter3 laplace
Chapter3 laplace
 
2D1431 Machine Learning
2D1431 Machine Learning2D1431 Machine Learning
2D1431 Machine Learning
 
Signal Processing Homework Help
Signal Processing Homework HelpSignal Processing Homework Help
Signal Processing Homework Help
 
Emat 213 study guide
Emat 213 study guideEmat 213 study guide
Emat 213 study guide
 
Geometric and viscosity solutions for the Cauchy problem of first order
Geometric and viscosity solutions for the Cauchy problem of first orderGeometric and viscosity solutions for the Cauchy problem of first order
Geometric and viscosity solutions for the Cauchy problem of first order
 
Cash Settled Interest Rate Swap Futures
Cash Settled Interest Rate Swap FuturesCash Settled Interest Rate Swap Futures
Cash Settled Interest Rate Swap Futures
 
On problem-of-parameters-identification-of-dynamic-object
On problem-of-parameters-identification-of-dynamic-objectOn problem-of-parameters-identification-of-dynamic-object
On problem-of-parameters-identification-of-dynamic-object
 
12.5. vector valued functions
12.5. vector valued functions12.5. vector valued functions
12.5. vector valued functions
 

Mehr von Sean Meyn

Quasi-Stochastic Approximation: Algorithm Design Principles with Applications...
Quasi-Stochastic Approximation: Algorithm Design Principles with Applications...Quasi-Stochastic Approximation: Algorithm Design Principles with Applications...
Quasi-Stochastic Approximation: Algorithm Design Principles with Applications...
Sean Meyn
 
State Space Collapse in Resource Allocation for Demand Dispatch - May 2019
State Space Collapse in Resource Allocation for Demand Dispatch - May 2019State Space Collapse in Resource Allocation for Demand Dispatch - May 2019
State Space Collapse in Resource Allocation for Demand Dispatch - May 2019
Sean Meyn
 
Distributed Randomized Control for Ancillary Service to the Power Grid
Distributed Randomized Control for Ancillary Service to the Power GridDistributed Randomized Control for Ancillary Service to the Power Grid
Distributed Randomized Control for Ancillary Service to the Power Grid
Sean Meyn
 

Mehr von Sean Meyn (20)

Quasi-Stochastic Approximation: Algorithm Design Principles with Applications...
Quasi-Stochastic Approximation: Algorithm Design Principles with Applications...Quasi-Stochastic Approximation: Algorithm Design Principles with Applications...
Quasi-Stochastic Approximation: Algorithm Design Principles with Applications...
 
DeepLearn2022 1. Goals & AlgorithmDesign.pdf
DeepLearn2022 1. Goals & AlgorithmDesign.pdfDeepLearn2022 1. Goals & AlgorithmDesign.pdf
DeepLearn2022 1. Goals & AlgorithmDesign.pdf
 
DeepLearn2022 3. TD and Q Learning
DeepLearn2022 3. TD and Q LearningDeepLearn2022 3. TD and Q Learning
DeepLearn2022 3. TD and Q Learning
 
DeepLearn2022 2. Variance Matters
DeepLearn2022  2. Variance MattersDeepLearn2022  2. Variance Matters
DeepLearn2022 2. Variance Matters
 
Smart Grid Tutorial - January 2019
Smart Grid Tutorial - January 2019Smart Grid Tutorial - January 2019
Smart Grid Tutorial - January 2019
 
State Space Collapse in Resource Allocation for Demand Dispatch - May 2019
State Space Collapse in Resource Allocation for Demand Dispatch - May 2019State Space Collapse in Resource Allocation for Demand Dispatch - May 2019
State Space Collapse in Resource Allocation for Demand Dispatch - May 2019
 
Irrational Agents and the Power Grid
Irrational Agents and the Power GridIrrational Agents and the Power Grid
Irrational Agents and the Power Grid
 
Zap Q-Learning - ISMP 2018
Zap Q-Learning - ISMP 2018Zap Q-Learning - ISMP 2018
Zap Q-Learning - ISMP 2018
 
Introducing Zap Q-Learning
Introducing Zap Q-Learning   Introducing Zap Q-Learning
Introducing Zap Q-Learning
 
Reinforcement Learning: Hidden Theory and New Super-Fast Algorithms
Reinforcement Learning: Hidden Theory and New Super-Fast AlgorithmsReinforcement Learning: Hidden Theory and New Super-Fast Algorithms
Reinforcement Learning: Hidden Theory and New Super-Fast Algorithms
 
State estimation and Mean-Field Control with application to demand dispatch
State estimation and Mean-Field Control with application to demand dispatchState estimation and Mean-Field Control with application to demand dispatch
State estimation and Mean-Field Control with application to demand dispatch
 
Demand-Side Flexibility for Reliable Ancillary Services
Demand-Side Flexibility for Reliable Ancillary ServicesDemand-Side Flexibility for Reliable Ancillary Services
Demand-Side Flexibility for Reliable Ancillary Services
 
Spectral Decomposition of Demand-Side Flexibility for Reliable Ancillary Serv...
Spectral Decomposition of Demand-Side Flexibility for Reliable Ancillary Serv...Spectral Decomposition of Demand-Side Flexibility for Reliable Ancillary Serv...
Spectral Decomposition of Demand-Side Flexibility for Reliable Ancillary Serv...
 
Demand-Side Flexibility for Reliable Ancillary Services in a Smart Grid: Elim...
Demand-Side Flexibility for Reliable Ancillary Services in a Smart Grid: Elim...Demand-Side Flexibility for Reliable Ancillary Services in a Smart Grid: Elim...
Demand-Side Flexibility for Reliable Ancillary Services in a Smart Grid: Elim...
 
Why Do We Ignore Risk in Power Economics?
Why Do We Ignore Risk in Power Economics?Why Do We Ignore Risk in Power Economics?
Why Do We Ignore Risk in Power Economics?
 
Distributed Randomized Control for Ancillary Service to the Power Grid
Distributed Randomized Control for Ancillary Service to the Power GridDistributed Randomized Control for Ancillary Service to the Power Grid
Distributed Randomized Control for Ancillary Service to the Power Grid
 
Ancillary service to the grid from deferrable loads: the case for intelligent...
Ancillary service to the grid from deferrable loads: the case for intelligent...Ancillary service to the grid from deferrable loads: the case for intelligent...
Ancillary service to the grid from deferrable loads: the case for intelligent...
 
Control Techniques for Complex Systems
Control Techniques for Complex SystemsControl Techniques for Complex Systems
Control Techniques for Complex Systems
 
Tutorial for Energy Systems Week - Cambridge 2010
Tutorial for Energy Systems Week - Cambridge 2010Tutorial for Energy Systems Week - Cambridge 2010
Tutorial for Energy Systems Week - Cambridge 2010
 
Panel Lecture for Energy Systems Week
Panel Lecture for Energy Systems WeekPanel Lecture for Energy Systems Week
Panel Lecture for Energy Systems Week
 

Kürzlich hochgeladen

The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptx
heathfieldcps1
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
ciinovamais
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptx
heathfieldcps1
 
Making and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdfMaking and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdf
Chris Hunter
 
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
kauryashika82
 

Kürzlich hochgeladen (20)

The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptx
 
Measures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeMeasures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and Mode
 
Role Of Transgenic Animal In Target Validation-1.pptx
Role Of Transgenic Animal In Target Validation-1.pptxRole Of Transgenic Animal In Target Validation-1.pptx
Role Of Transgenic Animal In Target Validation-1.pptx
 
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptxINDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The Basics
 
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
 
Class 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfClass 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdf
 
Python Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxPython Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docx
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
 
Asian American Pacific Islander Month DDSD 2024.pptx
Asian American Pacific Islander Month DDSD 2024.pptxAsian American Pacific Islander Month DDSD 2024.pptx
Asian American Pacific Islander Month DDSD 2024.pptx
 
This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.
 
Z Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphZ Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot Graph
 
Unit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptxUnit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptx
 
Application orientated numerical on hev.ppt
Application orientated numerical on hev.pptApplication orientated numerical on hev.ppt
Application orientated numerical on hev.ppt
 
psychiatric nursing HISTORY COLLECTION .docx
psychiatric  nursing HISTORY  COLLECTION  .docxpsychiatric  nursing HISTORY  COLLECTION  .docx
psychiatric nursing HISTORY COLLECTION .docx
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptx
 
Unit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptxUnit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptx
 
How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17
 
Making and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdfMaking and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdf
 
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
 

Markov Tutorial CDC Shanghai 2009

  • 1. Lyapunov functions, value functions, and performance bounds Sean Meyn Department of Electrical and Computer Engineering University of Illinois and the Coordinated Science Laboratory Joint work with R. Tweedie, I. Kontoyiannis, and P. Mehta Supported in part by NSF (ECS 05 23620, and prior funding), and AFOSR
  • 2. Objectives Nonlinear state space model ≡ (controlled) Markov process, state process X Typical form: dX(t) = f (X(t), U (t)) dt + σ(X(t), U (t)) dW (t) noise control
  • 3. Objectives Nonlinear state space model ≡ (controlled) Markov process, state process X Typical form: dX(t) = f (X(t), U (t)) dt + σ(X(t), U (t)) dW (t) noise control Questions: For a given feedback law, • Is the state process stable? • Is the average cost finite? E[c(X(t), U (t))] • Can we solve the DP equations? min c(x, u) + Du h∗ (x) = η ∗ u • Can we approximate the average cost η∗? The value function h∗ ?
  • 4. Outline Markov Models P t (x, · ) − π f →0 sup Ex [SτC (f )] < ∞ C π(f ) < ∞ Representations Lyapunov Theory DV (x) ≤ −f (x) + bIC (x) Conclusions
  • 6. Notation Markov chain: X = X(t) : t ≥ 0 Countable state space, X Transition semigroup, P t (x, y) = P X(s + t) = y X(s) = x , x, y ∈ X
  • 7. Notation: Generators & Resolvents Markov chain: X = X(t) : t ≥ 0 Countable state space, X Transition semigroup, P t (x, y) = P X(s + t) = y X(s) = x , x, y ∈ X Generator: For some domain of functions h, 1 Dh (x) = lim E[h(X(s + t)) − h(X(s)) X(s) = x] t→0 t 1 t = lim (P h (x) − h(x)) t→0 t
  • 8. Notation: Generators & Resolvents Generator: For some domain of functions h, 1 Dh (x) = lim E[h(X(s + t)) − h(X(s)) X(s) = x] t→0 t 1 t = lim (P h (x) − h(x)) t→0 t Rate matrix: Dh (x) = Q(x, y)h(y) P t = eQt y
  • 9. α µ Example: MM1 Queue  x + 1 Prob εα  Sample paths: X(t + ε) ≈ x − 1 Prob εµ   x Prob 1 − ε(α + µ) Rate matrix:   −α α 0 0 0 0 ···  µ −α − µ  α 0 0 0 · · ·   0  µ −α − µ α 0 0 · · ·   Q= 0 0 µ −α − µ α 0 · · ·   0  0 0 µ −α − µ α · · ·   0  0 0 0 µ −α − µ · · ·  . . . . . . . . . . . . . . . . . .
  • 10. 2 2 σW = 0 σW = 1 Example: O-U Model Sample paths: dX(t) = AX(t) dt + B dW (t) A n × n, Bn × 1, W standard BM 2 Generator: Dh (x) = (Ax)T h (x) + B T h (x)B
  • 11. 2 2 σW = 0 σW = 1 Example: O-U Model Sample paths: dX(t) = AX(t) dt + B dW (t) A n × n, Bn × 1, W standard BM 2 Generator: Dh (x) = (Ax)T h (x) + B T h (x)B h quadratic, h(x) = 1 xT P x 2 h (x) = P x 2 h (x) = P Dh (x) = 1 xT (P A + AT P )x + B T P B 2
  • 12. Notation: Generators & Resolvents Generator: For some domain of functions h, 1 Dh (x) = lim E[h(X(s + t)) − h(X(s)) X(s) = x] t→0 t 1 t = lim (P h (x) − h(x)) t→0 t Rate matrix: Dh (x) = Q(x, y)h(y) P t = eQt y Resolvent: ∞ Rα = e−αt P t 0
  • 13. Notation: Generators & Resolvents Generator: For some domain of functions h, 1 Dh (x) = lim E[h(X(s + t)) − h(X(s)) X(s) = x] t→0 t 1 t = lim (P h (x) − h(x)) t→0 t Rate matrix: Dh (x) = Q(x, y)h(y) P t = eQt y Resolvent: Resolvent equations: ∞ Rα = e−αt P t Rα = [ Iα − Q]−1 0 QRα = Rα Q = αRα − I
  • 14. Notation: Generators & Resolvents Motivation: Dynamic programming. For a cost function c, hα (x) = Rα c (x) = Rα (x, y)c(y) y∈X ∞ = eαt E[c(X(t)) X(0) = x] dt 0 Discounted-cost value function
  • 15. Notation: Generators & Resolvents Motivation: Dynamic programming. For a cost function c, hα (x) = Rα c (x) = Rα (x, y)c(y) y∈X ∞ = eαt E[c(X(t)) X(0) = x] dt 0 Discounted-cost value function Resolvent equation = dynamic programming equation, c + Dhα = αhα
  • 16. Notation: Steady State Distribution Invariant (probability) measure π: X is stationary. In particular, X(t) ∼ π, t≥0
  • 17. Notation: Steady State Distribution Invariant (probability) measure π: X is stationary. In particular, X(t) ∼ π, t≥0 Characterizations: π(x)P t (x, y) = π(y) x∈X α π(x)Rα (x, y) = π(y), α > 0 x∈X π(x)Q(x, y) = 0 x∈X y ∈X
  • 18. Notation: Relative Value Function Invariant measure π, cost function c , steady-state mean η Relative value function: ∞ h(x) = E[c(X(t)) − η X(0) = x] dt 0
  • 19. Notation: Relative Value Function Invariant measure π, cost function c , steady-state mean η Relative value function: ∞ h(x) = E[c(X(t)) − η X(0) = x] dt 0 Solution to Poisson’s equation (average-cost DP equation): c + Dh = η
  • 20. II Representations π ∝ ν[I − (R − s ⊗ ν)]−1 h = [I − (R − s ⊗ ν)]−1 c ˜
  • 21. Irreducibility ψ-Irreducibility: ψ(y) > 0 =⇒ P X(t) reaches y X(0) = x > 0 all x ψ(y) > 0 =⇒ R(x, y) > 0 all x
  • 22. Small Functions and Small Measures ψ-Irreducibility: ψ(y) > 0 =⇒ P X(t) reaches y X(0) = x > 0 all x ψ(y) > 0 =⇒ R(x, y) > 0 all x Small functions and measures: For a function s and probability ν, R(x, y) ≥ s(x)ν(y), x, y ∈ X ∞ R= e−t P t dt 0
  • 23. Small Functions and Small Measures ψ-Irreducibility: ψ(y) > 0 =⇒ P X(t) reaches y X(0) = x > 0 all x ψ(y) > 0 =⇒ R(x, y) > 0 all x Small functions and measures: For a function s and probability ν, R(x, y) ≥ s(x)ν(y), x, y ∈ X Resolvent dominates rank-one matrix, R= ∞ e−t P t dt 0 R ≥s⊗ν
  • 24. Small Functions and Small Measures ψ-Irreducibility: ψ(y) > 0 =⇒ P X(t) reaches y X(0) = x > 0 all x ψ(y) > 0 =⇒ R(x, y) > 0 all x Small functions and measures: For a function s and probability ν, R(x, y) ≥ s(x)ν(y), x, y ∈ X Resolvent dominates rank-one matrix, R= ∞ e−t P t dt 0 R ≥s⊗ν ψ-Irreducibility justi es assumption: s(x) > 0 for all x and WLOG, ν = δx ∗ , where ψ(x∗ ) > 0
  • 25. α µ Example: MM1 Queue R(x, y) > 0 for all x and y (irreducible in usual sense) Conclusion: R(x, y ) ≥ s(x)ν(y ) where s(x) := R(x, 0) ν := δ0
  • 26. 2 2 σW = 0 σW = 1 Example: O-U Model dX(t) = AX(t) dt + B dW (t) R(0, . ) Gaussian Full rank if and only if (A, B) is controllable. Conclusion: Under controllability, for any m, there is ε s.t., R(x, A) ≥ s(x)ν(A) all x and A where s(x) = ε I x ≤m ν(A) uniform on x ≤m
  • 27. Potential Matrix ∞ Potential matrix: G(x, y) = (R − s ⊗ ν)n (x, y) n=0 G = [I − (R − s ⊗ ν)]−1
  • 28. Representation of π ∞ Potential matrix: G(x, y) = (R − s ⊗ ν)n (x, y) n=0 π ∝ νG νG (y) = ν(x)G(x, y) x∈X
  • 29. Representation of h ∞ Potential matrix: G(x, y) = (R − s ⊗ ν)n (x, y) n=0 h = RG c ˜ + constant c(x) = c(x) − η ˜ G˜ (y) = c G(x, y)˜(y) c η= π(x)c(x) y∈X y∈X
  • 30. Representation of h ∞ Potential matrix: G(x, y) = (R − s ⊗ ν)n (x, y) n=0 h = RG c ˜ + constant c(x) = c(x) − η ˜ G˜ (y) = c G(x, y)˜(y) c η= π(x)c(x) y∈X y∈X If sum converges, then Poisson’s equation is solved: c(x) + Dh (x) = η
  • 31. III Lyapunov Theory P n (x, · ) − π f →0 sup Ex [SτC (f )] < ∞ C π(f ) < ∞ ∆V (x) ≤ −f (x) + bIC (x)
  • 32. Lyapunov Functions DV ≤ −g + bs
  • 33. Lyapunov Functions DV ≤ −g + bs General assumptions: V : X → (0,∞) g : X → [1, ∞) b < ∞, s small e.g., s (x) = IC (x), C nite
  • 34. Lyapunov Bounds on G DV ≤ −g + bs Resolvent equation gives RV − V ≤ −Rg + bRs
  • 35. Lyapunov Bounds on G DV ≤ −g + bs Resolvent equation gives RV − V ≤ −Rg + bRs Since s⊗ν is non-negative, −[I − (R − s ⊗ ν)]V ≤ RV − V ≤ −Rg + bRs G−1
  • 36. Lyapunov Bounds on G DV ≤ −g + bs Resolvent equation gives RV − V ≤ −Rg + bRs Since s⊗ν is non-negative, −[I − (R − s ⊗ ν)]V ≤ RV − V ≤ −Rg + bRs G−1 More positivity, V ≥ GRg − bGRs Some algebra, GR = G(R − s ⊗ ν) + (Gs) ⊗ ν ≥ G − I Gs ≤ 1
  • 37. Lyapunov Bounds on G DV ≤ −g + bs Resolvent equation gives RV − V ≤ −Rg + bRs Since s⊗ν is non-negative, −[I − (R − s ⊗ ν)]V ≤ RV − V ≤ −Rg + bRs G−1 More positivity, V ≥ GRg − bGRs Some algebra, GR = G(R − s ⊗ ν) + (Gs) ⊗ ν ≥ G − I Gs ≤ 1 General bound: GRg ≤ V + 2b Gg ≤ V + g + 2b
  • 38. Existence of π DV ≤ −g + bs Condition (V2) DV ≤ −1 + bs Representation: π ∝ νG Bound: GRg ≤ V + 2b =⇒ G(x, X) ≤ V (x) + 2b Conclusion: π exists as a probability measure on X
  • 39. Existence of moments DV ≤ −g + bs Condition (V3) DV ≤ −g + bs Representation: π ∝ νG Bound: Gg ≤ V + g + 2b
  • 40. Existence of moments DV ≤ −g + bs Condition (V3) DV ≤ −g + bs Representation: π ∝ νG Bound: Gg ≤ V + g + 2b Conclusion: π exists as a probability measure on X and the steady-state mean is nite, π(g) := π(x)g(x) ≤ b x∈X
  • 41. α α µ Example: MM1 Queue ρ = µ Linear Lyapunov function, V (x) = x ∞ DV (x) = Q(x, y)y y=0 = α(x + 1) + µ(x − 1) − (α + µ)x = −(µ − α) x>0 Conclusion: (V2) holds if and only if ρ < 1
  • 42. α α µ Example: MM1 Queue ρ = µ QuadraticLyapunov function, V (x) = x 2 ∞ DV (x) = Q(x, y)y 2 y=0 = α(x + 1)2 + µ(x − 1)2 − (α + µ)x2 = α(x2 + 2x + 1) + µ(x2 − 2x + 1)2 − (α + µ)x2 = −2(µ − α)x + α + µ Conclusion: (V3) holds, g(x) = 1 + x if and only if ρ < 1
  • 43. 2 2 σW = 0 σW = 1 Example: O-U Model dX(t) = AX(t) dt + B dW (t) h quadratic, h(x) = 1 xT P x 2 h (x) = P x 2 h (x) = P Dh (x) = 1 xT (P A + AT P )x + B T P B 2 Suppose that P > 0 solves the Lyapunov equation, P A + AT P = -I
  • 44. 2 2 σW = 0 σW = 1 Example: O-U Model dX(t) = AX(t) dt + B dW (t) h quadratic, h(x) = 1 xT P x 2 h (x) = P x 2 h (x) = P Dh (x) = 1 xT (P A + AT P )x + B T P B 2 Suppose that P > 0 solves the Lyapunov equation, P A + AT P = -I Then (V3) follows from the identity, 2 2 2 Dh (x) = − 1 x 2 + σX , σX = B T P B
  • 45. 2 2 σW = 0 σW = 1 Example: O-U Model dX(t) = AX(t) dt + B dW (t) The function h(x) = 1 xT P x solves Poisson’s equation, 2 1 2 Dh = −g + η g(x) = 2 x 2 η = σX Suppose that P > 0 solves the Lyapunov equation, P A + AT P = -I Then (V3) follows from the identity, 2 2 2 Dh (x) = − 1 x 2 + σX , σX = B T P B
  • 46. Poisson’s Equation DV ≤ −g + bs Condition (V3) DV ≤ −g + bs Representation: h = RG c ˜ + constant c(x) = c(x) − η ˜ Bound: RGg ≤ V + 2b =⇒ RGg (x) ≤ V (x) + 2b
  • 47. Poisson’s Equation DV ≤ −g + bs Condition (V3) DV ≤ −g + bs Representation: h = RG c ˜ + constant c(x) = c(x) − η ˜ Bound: RGg ≤ V + 2b =⇒ RGg (x) ≤ V (x) + 2b Conclusion: If c is bounded by g, then h is bounded, h(x) ≤ V (x) + 2b
  • 48. α α µ Example: MM1 Queue ρ = µ Poisson’s equation with g (x) = x Dh = −g + η We have (V3) with V a quadratic function of x: Recall, with h (x) = x 2 Dh (x) = −2(µ − α)x + α + µ x>0
  • 49. α α µ Example: MM1 Queue ρ = µ Poisson’s equation with g (x) = x Dh = −g + η Solved with x2 + x ρ h(x) = 1 2 µ−α η= 1−ρ
  • 51. P t (x, · ) − π f →0 sup Ex [SτC (f )] < ∞ C Final words π(f ) < ∞ DV (x) ≤ −f (x) + bIC (x) Just as in linear systems theory, Lyapunov functions provide a characterization of system properties, as well as a practical verification tool
  • 52. P t (x, · ) − π f →0 sup Ex [SτC (f )] < ∞ C Final words π(f ) < ∞ DV (x) ≤ −f (x) + bIC (x) Just as in linear systems theory, Lyapunov functions provide a characterization of system properties, as well as a practical verification tool Much is left out of this survey - in particular, • Converse theory • Limit theory • Approximation techniques to construct Lyapunov functions or approximations to value functions • Application to controlled Markov processes, and approximate dynamic programming
  • 53. References [1,4] ψ-Irreducible foundations [2,11,12,13] Mean- eld models, ODE models, and Lyapunov functions [1,4,5,9,10] Operator-theoretic methods. See also appendix of [2] [3,6,7,10] Generators and continuous time models [1] S. P. Meyn and R. L. Tweedie. Markov chains and stochastic [9] I. Kontoyiannis and S. P. Meyn. Spectral theory and limit stability. Cambridge University Press, Cambridge, second theorems for geometrically ergodic Markov processes. Ann. edition, 2009. Published in the Cambridge Mathematical Appl. Probab., 13:304–362, 2003. Presented at the INFORMS Library. Applied Probability Conference, NYC, July, 2001. [2] S. P. Meyn. Control Techniques for Complex Networks. Cam- [10] I. Kontoyiannis and S. P. Meyn. Large deviations asymptotics bridge University Press, Cambridge, 2007. Pre-publication and the spectral theory of multiplicatively regular Markov edition online: http://black.csl.uiuc.edu/˜meyn. processes. Electron. J. Probab., 10(3):61–123 (electronic), [3] S. N. Ethier and T. G. Kurtz. Markov Processes : Charac- 2005. terization and Convergence. John Wiley & Sons, New York, [11] W. Chen, D. Huang, A. Kulkarni, J. Unnikrishnan, Q. Zhu, 1986. P. Mehta, S. Meyn, and A. Wierman. Approximate dynamic [4] E. Nummelin. General Irreducible Markov Chains and Non- programming using fluid and diffusion approximations with negative Operators. Cambridge University Press, Cambridge, applications to power management. Accepted for inclusion in 1984. the 48th IEEE Conference on Decision and Control, December [5] S. P. Meyn and R. L. Tweedie. Generalized resolvents 16-18 2009. and Harris recurrence of Markov processes. Contemporary [12] P. Mehta and S. Meyn. Q-learning and Pontryagin’s Minimum Mathematics, 149:227–250, 1993. Principle. Accepted for inclusion in the 48th IEEE Conference [6] S. P. Meyn and R. L. Tweedie. Stability of Markovian on Decision and Control, December 16-18 2009. processes III: Foster-Lyapunov criteria for continuous time [13] G. Fort, S. Meyn, E. Moulines, and P. Priouret. ODE processes. Adv. Appl. Probab., 25:518–548, 1993. methods for skip-free Markov chain stability with applications [7] D. Down, S. P. Meyn, and R. L. Tweedie. Exponential to MCMC. Ann. Appl. Probab., 18(2):664–707, 2008. and uniform ergodicity of Markov processes. Ann. Probab., 23(4):1671–1691, 1995. [8] P. W. Glynn and S. P. Meyn. A Liapounov bound for solutions of the Poisson equation. Ann. Probab., 24(2):916–931, 1996. See also earlier seminal work by Hordijk, Tweedie, ... full references in [1].