|
[1]I. Goodfellow, Y. Bengio, and A. Courville, Deep learning. MIT press, 2016. [2]R. Yamashita, M. Nishio, R. K. G. Do, and K. Togashi, "Convolutional neural networks: an overview and application in radiology," Insights into Imaging, vol. 9, no. 4, pp. 611-629, 2018/08/01 2018, doi: 10.1007/s13244-018-0639-9. [3]savyakhosla. "CNN | Introduction to Pooling Layer." https://www.geeksforgeeks.org/cnn-introduction-to-pooling-layer/ (accessed 2022). [4]R. S. Sutton and A. G. Barto, Reinforcement Learning: An Introduction. A Bradford Book, 2018. [5]M. Naeem, S. T. H. Rizvi, and A. Coronato, "A Gentle Introduction to Reinforcement Learning and its Application in Different Fields," IEEE Access, vol. 8, pp. 209320-209344, 2020, doi: 10.1109/ACCESS.2020.3038605. [6]D. Silver et al., "Mastering the game of Go with deep neural networks and tree search," Nature, vol. 529, no. 7587, pp. 484-489, 2016/01/01 2016, doi: 10.1038/nature16961. [7]C. Watkins and P. Dayan, "Technical Note: Q-Learning," Machine Learning, vol. 8, pp. 279-292, 05/01 1992, doi: 10.1007/BF00992698. [8]R. Sutton, D. McAllester, S. Singh, and Y. Mansour, "Policy Gradient Methods for Reinforcement Learning with Function Approximation," Adv. Neural Inf. Process. Syst, vol. 12, 02/16 2000. [9]V. Mnih et al., "Playing Atari with Deep Reinforcement Learning," p. arXiv:1312.5602. [Online]. Available: https://ui.adsabs.harvard.edu/abs/2013arXiv1312.5602M [10]J. Schulman, S. Levine, P. Abbeel, M. Jordan, and P. Moritz, "Trust region policy optimization," in International conference on machine learning, 2015: PMLR, pp. 1889-1897. [11]N. Heess et al., "Emergence of locomotion behaviours in rich environments," arXiv preprint arXiv:1707.02286, 2017. [12]C. C.-Y. Hsu, C. Mendler-Dünner, and M. Hardt, "Revisiting design choices in proximal policy optimization," arXiv preprint arXiv:2009.10897, 2020. [13]T. Zhang, M. Xiao, Y.-b. Zou, J.-d. Xiao, and S.-y. Chen, "Robotic curved surface tracking with a neural network for angle identification and constant force control based on reinforcement learning," International Journal of Precision Engineering and Manufacturing, vol. 21, no. 5, pp. 869-882, 2020. [14]R. M. Associates. "Rhinoceros 3D." https://www.rhino3d.com/features/ (accessed 2022). [15]"瀝青混凝土路面施工及檢驗基準 ", 2020/6. Accessed: 2022. [16]陳建旭、劉韋廷、廖敏志、王慶雄、蔡益智、林和志, "分析鋪面坑洞產生原因與建議維護方法," 2011/10. Accessed: 2022. [17]U. Engine. "Unreal Engine." https://docs.unrealengine.com/4.26/zh-CN/ (accessed 2022). [18]Microsoft. "Aerial Informatics and Robotics Platform (AirSIm)." https://microsoft.github.io/AirSim/ (accessed 2022). [19]A. Raffin, A. Hill, A. Gleave, A. Kanervisto, M. Ernestus, and N. Dormann, "Stable-baselines3: Reliable reinforcement learning implementations," Journal of Machine Learning Research, 2021.
|