• Aucun résultat trouvé

[PDF] Top 20 Strong Uniform Value in Gambling Houses and Partially Observable Markov Decision Processes

Has 10000 "Strong Uniform Value in Gambling Houses and Partially Observable Markov Decision Processes" found on our website. Below are the top 20 most common "Strong Uniform Value in Gambling Houses and Partially Observable Markov Decision Processes".

Strong Uniform Value in Gambling Houses and Partially Observable Markov Decision Processes

Strong Uniform Value in Gambling Houses and Partially Observable Markov Decision Processes

... space and action sets are finite, Blackwell [6] has proved the existence of a pure strategy that is optimal for every discount factor close to 0, and one can deduce that the strong uniform ... Voir le document complet

25

Strong Uniform Value in Gambling Houses and Partially Observable Markov Decision Processes

Strong Uniform Value in Gambling Houses and Partially Observable Markov Decision Processes

... space and action sets are finite, Blackwell [6] has proved the existence of a pure strategy that is optimal for every discount factor close to 0, and one can deduce that the strong uniform ... Voir le document complet

26

Pathwise uniform value in gambling houses and Partially Observable Markov Decision Processes

Pathwise uniform value in gambling houses and Partially Observable Markov Decision Processes

... times. In the n-stage problem ...) and (v λ ) converge when n goes to infinity and λ goes to 0, and whether the two limits ...asymptotic value. The asymptotic value represents ... Voir le document complet

25

Efficient Policies for Stationary Possibilistic Markov  Decision Processes

Efficient Policies for Stationary Possibilistic Markov Decision Processes

... lexicographic value iteration VS Unbounded lexicographic value iteration and the possibility degree of the other one is uniformly fired in ...time and (ii) Pairwise success rate: ... Voir le document complet

12

Lightweight Verification of Markov Decision Processes with Rewards

Lightweight Verification of Markov Decision Processes with Rewards

... As a plausible variant of [22], it may seem reasonable to use the bounded properties of SMC to restrict the length of traces in the Kearns algorithm, rather than doing this implicitly with discounted rewards. The ... Voir le document complet

16

Subgeometric rates of convergence of f-ergodic strong Markov processes

Subgeometric rates of convergence of f-ergodic strong Markov processes

... subexponential in the tails. This model is particu- larly useful in Markov Chain Monte Carlo ...mial in the tails [11]; (c) we also give a toy hypoelliptic example, namely a stochastic damping ... Voir le document complet

39

Constructivist Anticipatory Learning Mechanism (CALM): Dealing with Partially Deterministic and Partially Observable Environments

Constructivist Anticipatory Learning Mechanism (CALM): Dealing with Partially Deterministic and Partially Observable Environments

... experiments in simple scenarios showing that the agent converges to the expected behavior, constructing correct knowledge to represent the environment deterministic regularities, as well as the regularities of its ... Voir le document complet

9

Planning in Markov Decision Processes with Gap-Dependent Sample Complexity

Planning in Markov Decision Processes with Gap-Dependent Sample Complexity

... rithms in a fixed-budget ...planning in Markov Decision Processes, that combines tools from best arm identification and optimistic planning and exploits tight confidence ... Voir le document complet

25

Efficient Policies for Stationary Possibilistic Markov  Decision Processes

Efficient Policies for Stationary Possibilistic Markov Decision Processes

... lexicographic value iteration VS Unbounded lexicographic value iteration and the possibility degree of the other one is uniformly fired in ...time and (ii) Pairwise success rate: ... Voir le document complet

11

Near Optimal Exploration-Exploitation in Non-Communicating Markov Decision Processes

Near Optimal Exploration-Exploitation in Non-Communicating Markov Decision Processes

... term in the regret bound. All the efforts we put in this direction were ...[18] and Theocharous et al. [19] proposed posterior sampling algorithms and proved bounds on the expected Bayesian ... Voir le document complet

28

Algorithmic aspects of mean–variance optimization in Markov decision processes

Algorithmic aspects of mean–variance optimization in Markov decision processes

... singleton and because the sum or convex hull of finitely many polyhedra is a ...polyhedron in terms of its finitely many extreme points. In the worst case, this translates into an exponential time ... Voir le document complet

26

Smart Sampling for Lightweight Verification of Markov Decision Processes

Smart Sampling for Lightweight Verification of Markov Decision Processes

... SMC and is iterated until an example is found or sufficient attempts have been ...not in general converge to the true maximum (the number of state-actions does not actually indicate scheduler probability), ... Voir le document complet

14

Aggregating Optimistic Planning Trees for Solving Markov Decision Processes

Aggregating Optimistic Planning Trees for Solving Markov Decision Processes

... UCT, and FSSS on the inverted pendulum benchmark problem, showing the sum of discounted rewards for simulations of 50 time ...budgets. In the cases of ASOP, UCT, and FSSS, the budget is in ... Voir le document complet

9

DetH*: Approximate Hierarchical Solution of Large Markov Decision Processes

DetH*: Approximate Hierarchical Solution of Large Markov Decision Processes

... [Sanner and McAllester, 2005; Sanner et ...ical and focused topological value iteration [Dai and Gold- smith, 2007; Dai et ...ours in that it de- composes a large MDP based on the ... Voir le document complet

9

A Learning Design Recommendation System Based on Markov Decision Processes

A Learning Design Recommendation System Based on Markov Decision Processes

... easy and in the end did not meet the essential requirement: assist teachers ...teachers in choosing a course structure according to a defined instructional design approach ...IMS-LD and adding ... Voir le document complet

9

Collision Avoidance for Unmanned Aircraft using Markov Decision Processes

Collision Avoidance for Unmanned Aircraft using Markov Decision Processes

... can in fact generate non-trivial (if not superior) collision avoidance strategies that can compete with hand-crafted ...field-of-view and lack of horizontal localization ability, provides us a good example ... Voir le document complet

23

Planning in Partially Observable Domains with Fuzzy Epistemic States and Probabilistic Dynamics

Planning in Partially Observable Domains with Fuzzy Epistemic States and Probabilistic Dynamics

... described in a purely qualitative possi- bilistic ...reward and possibility degrees, ...proposed in these ...detailed in Section ...current and next beliefs, or epistemic states, are ... Voir le document complet

15

The steady-state control problem for Markov decision processes

The steady-state control problem for Markov decision processes

... Conclusion In this paper, we have defined the steady-state control problem for MDP, and shown that this question is decidable for (ergodic) MDP in polynomial time, and for labeled MDP ... Voir le document complet

17

Constrained Markov Decision Processes with Total Expected Cost Criteria

Constrained Markov Decision Processes with Total Expected Cost Criteria

... Theorem 2. Choose any initial distribution β and policy u. Then either C(β, u) = ∞ or there exists a stationary policy w such that C(β, w) ≤ C(β, u), Proof. This is an extension of thm 8.1 in p.100 of [1]. ... Voir le document complet

3

Limits of Multi-Discounted Markov Decision Processes

Limits of Multi-Discounted Markov Decision Processes

... [17] and was also considered in [5, ...MDPs and one–player Shapley game and this correspondence pre- serves expected payoff of ...MDP and the corresponding one–player Shapley game have ... Voir le document complet

13

Show all 10000 documents...