Documentation

Probability.MDP.Histories

structure MDPs.MDP :

Markov decision process

Instances For
    inductive MDPs.Hist (M : MDP) :

    Represents a history. The state is type ℕ and action is type ℕ.

    Instances For
      instance MDPs.instCoeFinSHist (M : MDP) :
      Coe (Fin M.S) (Hist M)

      Coerces a state to a history

      Equations
      @[reducible]
      def MDPs.Hist.length (M : MDP) :
      Hist M

      The length of the history corresponds to the zero-based step of the decision

      Equations
      Instances For
        @[reducible, inline]
        abbrev MDPs.HistNE (m : MDP) :

        Nonempty histories

        Equations
        Instances For
          def MDPs.Hist.last (M : MDP) :
          Hist MFin M.S

          Returns the last state of the history

          Equations
          Instances For
            def MDPs.Hist.append (M : MDP) (h : Hist M) (as : Fin M.A × Fin M.S) :

            Appends the state and action to the history -

            Equations
            Instances For
              def MDPs.Hist.one (M : MDP) (s₀ : Fin M.S) (a : Fin M.A) (s : Fin M.S) :
              Equations
              Instances For
                def MDPs.Hist.prefix (M : MDP) (k : ) (h : Hist M) :

                Return the prefix of hist of length k

                Equations
                Instances For