|
[1]R. S. Sutton and A. G. Barto, Reinforcement learning: An introduction. MIT press Cambridge, vol. 1, no. 1, 1998. [2]J. Li and Y. Li, “Dynamic analysis and PID control for a quadrotor,” 2011 IEEE Int’l Conf. Mechatronics and Automation (ICMA), pp. 573-578, 2011. [3]Salih, A.L. Moghavvemi, M. Mohamed, H.A.F. Gaeid, K.S. , "Modelling and PID controller design for a quadrotor unmanned air vehicle," Automation Quality and Testing Robotics (AQTR), 2010 IEEE International Conference on , vol.1, no., pp.1-5, 28-30 May 2010. [4]Jun Li, Yuntang Li, "Dynamic analysis and PID control for a quadrotor", Mechatronics and Automation (ICMA) 2011 International Conference on, pp. 573-578, 2011, ISSN 2152-7431. [5]S. Bouabdallah, A. Noth, and R. Siegwart, "PID vs LQ control techniques applied to an indoor micro quadrotor," IEEE/RSJ Int'l Conf. Intelligent Robots and Systems (IROS 2004), vol. 3, pp. 2451-2456, 2004. [6]T. Madani and A. Benallegue, “Backstepping control for a quadrotor helicopter,” IEEE/RSJ Int’l Conf. Intelligent Robots and Systems (IROS 2006, pp. 3255-260, 2006. [7]C. Nicol, C. J. B. Macnab, and A. Ramirez-Serrano, “Robust adaptive control of aquadrotor helicopter,” Mechatronics, vol. 21, no. 6, pp. 927–938, 2011. [8]D. Lee, H. J. Kim, and S. Sastry, “Feedback linearization vs. adaptive sliding mode control for a quadrotor helicopter,” Int. J. Control, Autom., Syst., vol. 7, no. 3, pp. 419–428, 2009. [9]C. Diao, B. Xian, Q. Yin, W. Zeng, H. Li, and Y. Yang, “A nonlinear adaptive control approach for quadrotor UAVs,” 2011 IEEE 8th Asian Control Conference (ASCC), pp. 223-228, 2011. [10]S. Bouabdallah, P. Murrieri, and R. Siegwart, “Design and control of an indoor micro quadrotor,” IEEE Int’l Conf. Robotics and Automation (ICRA’04), vol. 5, pp. 4393-4398, 2004. [11]S. Grzonka, et al. “A Fully Autonomous Indoor Quadrotor”. IEEE Transactions on Robotics 28(1): 90-100, 2012 [12]S. Ross, N. Melik-Barkhudarov, K. S. Shankar, A. Wendel, D. Dey, J. A. Bagnell and M. Hebert, “Learning Monocular Reactive UAV Control in Cluttered Natural Environments,” Proceedings of IEEE International Conference on Robotics and Automation (ICRA2013), pp. 1765-1772, 2013. [13]A. Cherian, J. Andersh, V. Morellas and B. Mettler, “Autonomous Altitude Estimation of aUAV using a Single Onboard Camera,” Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 3900-3905, 2009. [14]H. Kimura and S. Kobayashi: “An analysis of actor-critic algorithms using eligibility traces: Reinforcement learning with imperfect value functions,” Japanese Society for Aritificial Intelligence, 2000. [15]R. S. Sutton, “Learning to Predict by The methods of Temporal Differences”, Machine Learning 3: 9-44, 1988 [16]A. G. Barto, R. S. Sutton, and C. W. Anderson, “Neuronlike adaptive elements that can solve difficult learning control problems,” IEEE Transactions on Systems, Man, and Cybernetics, vol. 13, no. 5, pp. 834–846, 1983. [17]Hamilton, W.R. 1969. Elements of quaternions. New York: Chelsea. [18]C. Y. Huang, “Compliance Locomotion Control for a Quadruped Robot with Damper and Spring Coefficients Assigned by Reinforcement Learning”, Toyota Technological Institute Master’s Thesis, 2017 [19]H. Kimura, T. Yamashita, S. Kobayashi, “Reinforcement learning of walking behavior for a four-legged robot”, In Proc. of 40th IEEE Conference on Decision and Control, pp 411-416, 2001 [20]E. Jucker, “Equations fondamentales des micromoteurs courant continu avec rotor sans fer”. Bulletin technique Portescap, La Chaud-de-Fonds, 1974. [21]T. Hester, M. Quinlan, P. Stone, “Generalized model learning for Reinforcement Learning on a humanoid robot”, IEEE International Conference on Robotics and Automation (ICRA), 2010. [22]M. A. K. Jaradat, M. Al-Rousan , L. Quadan, “Reinforcement based mobile robot navigation in dynamic environment”, Robotics and Computer-Integrated Manufacturing, vol 27, pp 135-149, 2011 [23]P. Fankhauser, M. Hutter, C. Gehring, M. Bloesch, M. A. Hoepflinger, R. Siegwart, “Reinforcement learning of single legged locomotion”, IEEE International Conference on Intelligent Robots and Systems (IROS), Nov. 2013, pp. 188–193.
|