首页    期刊浏览 2024年12月04日 星期三
登录注册

文章基本信息

  • 标题:DNN Tree Search for Bayesian Reinforcement Learning to Machine Intelligence
  • 本地全文:下载
  • 作者:Anil Kumar Yadav ; Ajay Kumar Sachan
  • 期刊名称:International Journal of Soft Computing & Engineering
  • 电子版ISSN:2231-2307
  • 出版年度:2014
  • 卷号:4
  • 期号:4
  • 页码:8-11
  • 出版社:International Journal of Soft Computing & Engineering
  • 摘要:Bayesian model-based reinforcement learning can be formulated as a partially observable Markova decision process (POMDP) to provide a principled framework for optimally balancing exploitation and exploration. Then, a POMDP solver can be used to solve the problem. If the prior distribution over the environment’s dynamics is a product of dirichlet distributions, the POMDP’s optimal value function can be represented using a set of multivariate polynomials. Unfortunately, the size of the polynomials grows exponentially with the problem horizon [3]. During machine learning agent required lots of training inputs of execution cycle. Due to this situation look up table contain huge amount of data base. In this paper, we observe the use of dynamic neural network tree search (DNNTS) algorithm for large POMDPs, to solve the Bayesian reinforcement learning problem. The keen idea of DNN tree search is to train agent and act as a NN classifier to help agent for taking self decision without prior knowledge of the system during data learning .We will show that such an algorithm successfully searches for a near-optimal policy and achieve goal. Experiments show that the used DNN methods improve performance of Bayesian reinforcement learning in the context of training episodes, reward and discount rate
  • 关键词:Bayesian reinforcement learning; machine learning;DNN tree search; POMDP
国家哲学社会科学文献中心版权所有