首页 | 官方网站   微博 | 高级检索  
     

基于ART2的Q学习算法研究
引用本文:姚明海,瞿心昱,李佳鹤,顾勤龙,汤丽平.基于ART2的Q学习算法研究[J].控制与决策,2011,26(2):227-232.
作者姓名:姚明海  瞿心昱  李佳鹤  顾勤龙  汤丽平
作者单位:浙江工业大学,信息工程学院,杭州,310023
基金项目:国家自然科学基金项目,浙江省自然科学基金项目
摘    要:为了解决Q学习应用于连续状态空间的智能系统所面临的"维数灾难"问题,提出一种基于ART2的Q学习算法.通过引入ART2神经网络,让Q学习Agent针对任务学习一个适当的增量式的状态空间模式聚类,使Agent无需任何先验知识,即可在未知环境中进行行为决策和状态空间模式聚类两层在线学习,通过与环境交互来不断改进控制策略,从而提高学习精度.仿真实验表明,使用ARTQL算法的移动机器人能通过与环境交互学习来不断提高导航性能.

关 键 词:Q学习  ART2  增量式学习  两层在线学习  移动机器人导航
收稿时间:2009/11/25 0:00:00
修稿时间:2010/6/18 0:00:00

Study on Q-learning Algorithm Based on ART2
TAO Meng-Hai,JI Xin-Yu,LI Jia-He,GU Qi-Long,SHANG Li-Beng.Study on Q-learning Algorithm Based on ART2[J].Control and Decision,2011,26(2):227-232.
Authors:TAO Meng-Hai  JI Xin-Yu  LI Jia-He  GU Qi-Long  SHANG Li-Beng
Affiliation:(College of Information Engineering,Zhejiang University of Technology,Hangzhou 310023,China.)
Abstract:

In order to solve the problem of dimension disaster which may be produced by applying Q-learing to intelligent
system of continuous state-space, this paper proposes a  Q-learning algorithm based on ART 2 and gives the specific steps. Through introducing the ART 2 neural network in the Q-learning algorithm, Q-learning Agent in view of the duty learns an appropriate incremental clustering of state-space model, so Agent can carry out decision-making and a two-tiers online learning of state-space model cluster in unknown environment without any priori knowledge. Through the interaction with the environment unceasingly alternately to improve the control strategies, the learning accuracy is increased. Finally, the mobile robot navigation simulation experiments show that, using the ARTQL algorithm, motion robot can improve its navigation performance continuously by interactive learning with the environment.

Keywords:ART2
本文献已被 CNKI 万方数据 等数据库收录!
点击此处可从《控制与决策》浏览原始摘要信息
点击此处可从《控制与决策》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号