نتایج جستجو برای: partially s
تعداد نتایج: 828349 فیلتر نتایج به سال:
In this paper, we focus on a “local property” of permutations: value-peak. A permutation σ has a value-peak σ(i) if σ(i − 1) < σ(i) > σ(i + 1) for some i ∈ [2, n − 1]. Define V P (σ) as the set of value-peaks of the permutation σ. For any S ⊆ [3, n], define V Pn(S) such that V P (σ) = S. Let Pn = {S | V Pn(S) 6= ∅}. we make the set Pn into a poset Pn by defining S T if S ⊆ T as sets. We prove t...
In this paper we investigate the behavior of credit yield spreads in an equilibrium framework in which the risk-free interest rate and yields on risky debt are jointly and endogenously determined. The model is one of a partially observable pure exchange economy in which debt is a contingent claim on the cash flow. We find that an increase in risk aversion or an increase in uncertainty about gro...
Integrating diagnosis and repair is particularly crucial when gaining sufficient information to discriminate between several candidate diagnoses requires carrying out some repair actions. A typical case is supply restoration in a faulty power distribution system. This problem, which is a major concern for electricity distributors, features partial observability, and stochastic repair actions wh...
106 AI MAGAZINE RTS games — such as StarCraft by Blizzard Entertainment and Command and Conquer by Electronic Arts — are popular video games that can be described as real-time war simulations in which players delegate units under their command to gather resources, build structures, combat and support units, scout opponent locations, and attack. The winner of an RTS game usually is the player or...
An Abstraction-Based Approach to Interleaving Planning and Execution in Partially-Observable Domains
ing Actions The DRIPS planner primarily uses two types of abstraction: interaction-abstraction and sequential abstraction. In this section, we briefly describe each. For a formal theory and detailed discussion of action abstraction that includes both these types of abstraction,
RÉSUMÉ. Cet article se focalise sur des approches formelles pour la construction de systèmes multi-agents. Ce travail a cherché à proposer des apprentissages décentralisés pour construire les comportements d’agents sociaux. Cet article propose un formalisme original, l’interacDEC-POMDP inspiré des modèles markoviens au sein duquel les agents peuvent interagir directement et localement entre eux...
We consider reinforcement learning in partially observable domains where the agent can query an expert for demonstrations. Our nonparametric Bayesian approach combines model knowledge, inferred from expert information and independent exploration, with policy knowledge inferred from expert trajectories. We introduce priors that bias the agent towards models with both simple representations and s...
It is commonly stated that reinforcement learning (RL) algorithms require more samples to learn than humans. In this work, we investigate this claim using two standard problems from the RL literature. We compare the performance of human subjects to RL techniques. We find that context—the meaningfulness of the observations—plays a significant role in the rate of human RL. Moreover, without conte...
In this paper we formalize a local strategy for plan repair in a Multi Agent Plan, where agents exhibit a collaborative behavior to reach a common global goal in a partially observable environment. The recovery strategy is based on a plan failure analysis where two main results are inferred: the agent diagnosis (which explains the action failure in terms of faults in agent functionalities) and ...
We consider the problem of synthesizing an agent protocol satisfying LTL specifications for multiple, partiallyobservable environments. We present a sound and complete procedure for solving the synthesis problem in this setting and show it is computationally optimal from a theoretical complexity standpoint. While this produces perfect-recall, hence unbounded, strategies we show how to transform...
نمودار تعداد نتایج جستجو در هر سال
با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید