|
[1]C. J. C. H. Watkins , and P. Dayan, “Technical note: Q-Learning, ” Machine Learning, 8(3-4): pp. 279-292, 1992. [2]Araujo A. F. R., Braga, A. P. S., “Reward-penalty reinforcement learning scheme for planning and reactive behavior ,” IEEE International Conference on Systems, Man, and Cybernetics, Vol. 2, pp.1485-1490, 1998. [3]Junfei Qiao, Zhanjun Hou, Xiaogang Ruan, “Q-learning Based on Neural Network in Learning Action Selection of Mobile Robot,” IEEE International Conference on Automation and Logistics, pp. 263 - 267, 2007. [4]Bing-Qiang Huang, Guang-Yi Cao, Min Guo, “Reinforcement Learning Neural Network to the Problem of Autonomous Mobile Robot Obstacle Avoidance,” International Conference on Machine Learning and Cybernetics, Vol. 1, pp. 85-89, 2005. [5]Caihong Li, Jingyuan Zhang, Yibin Li, “Application of Artificial Neural Network Based on Q-learning for Mobile Robot Path Planning,” IEEE International Conference on Information Acquisition, pp.978 - 982, 2006. [6]Minato T., Asada M., “Environmental change adaptation for mobile robot navigation,” IEEE/RSJ International Conference on Intelligent Robots and Systems , Vol. 3, pp.1859 -1864, 1998. [7]C. F. Touzet, “Q-learning for robot,” in M. A. Arbib, editor, Handbook of Brain Theory and Neural Networks, pp. 934-937, 2003. [8]R. S. Sutton, “Dyna, an integrated architecture for learning, planning and reacting,” Working Notes of the 1991 AAAI Spring Symposium on Integrated Intelligent Architectures and SIGART Bulletin 2, pp. 160-163, 1991. [9]Leslie Pack Kaebling, Michael L. Littman, and Andrew W. Moore. “Reinforcement learning:A survey,” Journal of Artificial Intelligence Research 4, pp.237-285, May 1996. [10]Peng J., Williams R. J., “Efficient learning and planning within the Dyna framework,” IEEE International Conference on Neural Networks, vol. 1, pp168 - 174, 1993. [11]Hoang-huu VIET, Sang-hyeok AN, Tae-choong CHUNG, “Extended Dyna-Q Algorithm for Path Planning of Mobile Robots,” Journal of Measurement Science and Instrumentation, Vol. 2, No. 3, pp. 283-287, 2011. [12]R. S. Sutton and A. G. Barto, Reinforcement Learning An Introduction, Cambridge, Mass., MIT Press, 1998. [13]Tom M. Mitchell, Machine Learning, Chapter 13, MCGRAW HILL Publisher , 1997. [14]Cameron B. Browne, Edward Powley, Daniel Whitehouse, Simon M. Lucas, Peter I. Cowling, Philipp Rohlfshagen, Stephen Tavener, Diego Perez, Spyridon Samothrakis, Simon Colton, “A Survey of Monte Carlo Tree Search Methods,” IEEE Transactions on Computational Intelligence and AI in Games, Vol. 4, pp. 1-43, 2012. [15]Jason Sanders and Edward Kandrot, CUDA by Example: An Introduction to General-Purpose GPU Programming, Addison-Wesley, 2010. [16]張舒、褚豔利、趙開勇、張鈺勃, GPU 高效能運算之CUDA, 碁峰資訊, 2011 [17]NVIDIA CUDA Library Documentation, http://developer.download.nvidia.com/ compute/cuda/4_1/rel/toolkit/docs/online/index.html
|