期刊名称:International Journal of Computer Trends and Technology
电子版ISSN:2231-2803
出版年度:2016
卷号:35
期号:2
页码:60-66
DOI:10.14445/22312803/IJCTT-V35P111
出版社:Seventh Sense Research Group
摘要:The output of the system is a sequence of actions in some applications. There is no such measure as the best action in any inbetween state; an action is excellent if it is part of a good policy. A single action is not important; the policy is important that is the sequence of correct actions to reach the goal. In such a case, machine learning program should be able to assess the goodness of policies and learn from past good action sequences to be able to generate a policy. A multiagent environment is one in which there is more than one agent, where they interact with one another, and further, where there are restrictions on that environment such that agents may not at any given time know everything about the world that other agents know. Two features of multiagent learning which establish its study as a separate field from ordinary machine learning. Parallelism, scalability, simpler construction and cost effectiveness are main characteristics of multiagent systems. Multiagent learning model is given in this paper. Two multiagent learning algorithms i. e. Strategy Sharing & Joint Rewards algorithm are implemented. In Strategy Sharing algorithm simple averaging of Q tables is taken. Each Qlearning agent learns from all of its teammates by taking the average of Qtables. Joint reward learning algorithm combines the Q learning with the idea of joint rewards. Paper shows result and performance comparison of the two multiagent learning algorithms.