|
[1]D.E. Broadbent, “Perception and communication,” Elsevier, pp. 75-80, 2013. [2]A.L. Yarbus, “Eye movements during perception of complex objects,” Eye movements and vision, pp.171-211, 1967. [3]C.W. Kao, Y.J. Huang, K.C. Fan, H.H. Chen, W.C. Chung, B.J. Hwan and C.H. Hsieh, “The integrated gaze, web and object tracking techniques for the web-based e-learning platform,” IEEE International Conference on Teaching, Assessment and Learning for Engineering (TALE), pp. 720—724, 2013. [4]Y. Sugano, Y. Matsushita and Y. Sato, “Appearance-Based Gaze Estimation Using Visual Saliency,” IEEE Transactions on Pattern Analysis and Machine Intelligence, pp. 329-341, 2013. [5]S.V. Sheela and P.A. Vijaya, “An Appearance based Method for Eye Gaze Tracking,” Journal of Computer Science, pp.1194-1203, 2011. [6]H.C. Lu, G.L. Fang, C. Wang and Y.W. Chen, “A novel method for gaze tracking by local pattern model and support vector regressor,” Signal Processing, pp. 1290-1299, 2010. [7]C.W. Kao, B.J. Hwang, C.W. Yang, K.C. Fan and C.P. Huang, “A Novel with Low Complexity Gaze Point Estimation Algorithm,” Proceedings of the International MultiConference of Engineers and Computer Scientists., vol. 1, 2012. [8]K. Nayan,“EyeX Controller: SteelSeries and Tobii Technology team up to take on Kinect with eye-tracking controller,” SIGGRAPH ASIA 2016 Courses, pp. 17, 2016. [9]K. Krafka, A. Khosla, P. Kellnhofer, H. Kannan, S. Bhandarkar, W. Matusik and A. Torralba, “Eye tracking for everyone,” Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2176-2184, 2016. [10]E. Wood, T. Baltrusaitis, X. Zhang, Y. Sugano, P. Robinson and A. Bulling, “Rendering of eyes for eye-shape registration and gaze estimation,” Proceedings of the IEEE International Conference on Computer Vision, pp. 3756-3764, 2015. [11]R. Jafari and D. Ziou, “Gaze estimation using Kinect/PTZ camera,” Robotic and Sensors Environments (ROSE), 2012. [12]K. A. Funes Mora and J. M. Odobez, “Gaze estimation from multimodal kinect data,” IEEE Conference in Computer Vision and Pattern Recognition, Workshop on Gesture Recognition(CVPRW), pp. 25-30, 2012. [13]J. Markoff, “How many computers to identify a cat? 16,000,” New York Times, pp. 06-25, 2012. [14]S. Draghici, “A neural network based artificial vision system for licence plate recognition,” International Journal of Neural Systems, pp. 113-126, 1997. [15]C. Meng and X. Zhao, “Webcam-Based Eye Movement Analysis Using CNN,” IEEE Access 5, pp. 19581-19587, 2017. [16]S. Hoppe and A. Bulling, “End-to-end eye movement detection using convolutional neural networks,” arXiv preprint arXiv:1609.02452, 2016. [17]C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke and A. Rabinovich, “Going deeper with convolutions,” Proceedings of the IEEE conference on computer vision and pattern recognition, 2015. [18]K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” arXiv preprint arXiv:1409.1556, 2014. [19]F. Chollet, “Keras,” Keras, pp. 153-157, 2015. [20]T. Tran and T. Pham, “A Bayesian Data Augmentation Approach for Learning Deep Models,” Advances in Neural Information Processing Systems, pp. 2794-2803, 2017. [21]H. Furukawa, “Deep Learning for Target Classification from SAR Imagery: Data Augmentation and Translation Invariance,” IEICE Technical Report, pp. 13-17, 2017. [22]K. Nguyen, T. Le, V. Nguyen, T. Nguyen, and D. Phung, “Multiple Kernel Learning with Data Augmentation,” Asian Conference on Machine Learning, pp. 49-64, 2016. [23]D. Komura, and S. Ishikawa, “Machine learning methods for histopathological image analysis,” arXiv preprint arXiv:1709.00786, 2017. [24]R. Kohavi, “A study of cross-validation and bootstrap for accuracy estimation and model selection,” Ijcai., vol. 14. no. 2, pp.1137-1145, 1995. [25]B. F. Klare, B. Klein, E. Taborsky, A. Blanton, J. Cheney, K. Allen, P. Grother, A. Mah and A. K. Jain, “Pushing the frontiers of unconstrained face detection and recognition: Iarpa janus benchmark a,” Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1931-1939, 2015. [26]D. E. King, “Dlib-ml: A machine learning toolkit,” Journal of Machine Learning Research, pp. 1755-1758, 2009 [27]A. Jana, “Kinect for windows SDK programming guide,” Packt Publishing Ltd, 2012. [28]K. K. Shieh and D. S. Lee, “Preferred viewing distance and screen angle of electronic paper displays,” Applied Ergonomics, pp. 601-608, 2007. [29]D. Birant and A. Kut, “ST-DBSCAN: An algorithm for clustering spatial–temporal data,” Data & Knowledge Engineering, vol. 60. no.1, pp. 208-221, 2007. [30]K. He, X. Zhang, S. Ren and J. Sun, “Identity mappings in deep residual networks,” European conference on computer vision, pp. 630-645, 2016.
|