نتایج جستجو برای: markov reward models

تعداد نتایج: 981365  

Behnam Zarpak, Rahman Farnoosh,

  Stochastic models such as mixture models, graphical models, Markov random fields and hidden Markov models have key role in probabilistic data analysis. In this paper, we have learned Gaussian mixture model to the pixels of an image. The parameters of the model have estimated by EM-algorithm.   In addition pixel labeling corresponded to each pixel of true image is made by Bayes rule. In fact, ...

2007
Yujuan Bao İlker N. Bozkurt Tuǧrul Dayar Xiaobai Sun Kishor S. Trivedi

This contribution proposes a decompositional iterative method for the steadystate analysis of Kronecker structured Markov chains [1]. The Markovian system, which is formed by a composition of subsystems using the Kronecker sum operator for local transitions and the Kronecker product operator for synchronized transitions, is assumed to have irreducible subsystem matrices associated with local tr...

2014
Martin Schwarick

This thesis investigates the efficient analysis, especially the model checking, of bounded stochastic Petri nets (SPNs) which can be augmented with reward structures. An SPN induces a continuous-time Markov chain (CTMC). A reward structure associates a reward to each state of the CTMC and defines a Markov reward model (MRM). The Continuous Stochastic Reward Logic (CSRL) permits to define sophis...

2016
Kamyar Azizzadenesheli Alessandro Lazaric Anima Anandkumar

We propose a new reinforcement learning algorithm for partially observable Markov decision processes (POMDP) based on spectral decomposition methods. While spectral methods have been previously employed for consistent learning of (passive) latent variable models such as hidden Markov models, POMDPs are more challenging since the learner interacts with the environment and possibly changes the fu...

Journal: :CoRR 2017
Kamyar Azizzadenesheli Alessandro Lazaric Anima Anandkumar

We propose a new reinforcement learning algorithm for partially observable Markov decision processes (POMDP) based on spectral decomposition methods. While spectral methods have been previously employed for consistent learning of (passive) latent variable models such as hidden Markov models, POMDPs are more challenging since the learner interacts with the environment and possibly changes the fu...

2005
Yaxin Liu Sven Koenig

Decision-theoretic planning with risk-sensitive planning objectives is important for building autonomous agents or decision-support systems for real-world applications. However, this line of research has been largely ignored in the artificial intelligence and operations research communities since planning with risk-sensitive planning objectives is more complicated than planning with risk-neutra...

Journal: :Expert Syst. Appl. 2015
Saisakul Chernbumroong Shuang Cang Hongnian Yu

In the multi-sensor activity recognition domain, the input space is often large and contains irrelevant and overlapped features. It is important to perform feature selection in order to select the smallest number of features which can describe the outputs. This paper proposes a new feature selection algorithms using the maximal relevance and maximal complementary criteria (MRMC) based on neural...

2007
Jinchun Xia

Performance is a crucial factor in software system. For many software development groups, it remains easier to wait until a system has been built before evaluating its performance. Among the three most used techniques, measurement, simulation and analytic modeling , measurement is believed to be the most accurate method. But it is only feasible after system is implemented. Most organizations re...

ژورنال: پژوهش در پزشکی 2013
اشرفی حافظ, اصغر, الماسی, اسحاق, امینی, یاسمن, ساروخانی, دیانا, سایه میری, روح الله, سایه میری, کورش, علی مقدم, کامران,

Abstract Background: Semi-Markov multi-state models are very important to describe regression and progression in chronic diseases and cancers. Purpose of this research was to determine the prognostic factors for survival after acute leukemia using multi-state models. Materials and Methods: In this descriptive longitudinal research, a total of 507 acute leukemia patients (206 acute lymphocyt...

2013
Karel Sladký

This contribution is devoted to risk-sensitive and risk-neutral optimality in Markov decision chains. Since the traditional optimality criteria (e.g. discounted or average rewards) cannot reflect the variability-risk features of the problem, and using the mean variance selection rules that stem from the classical work of Markowitz present some technical difficulties, we are interested in expect...

نمودار تعداد نتایج جستجو در هر سال

با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید