|
[1] A. A. Zakharov and A. E. Barinov, “An algorithm for 3D-object reconstruction from video using stereo correspondences,” Journal of Pattern Recognition and Image Analysis, vol. 25, no.1, 2015, pp. 117-121. [2] D. Kosmopoulos and S. Chatzis, “Robust visual behavior recognition: A framework based on holistic representations and multicamera information fusion,” IEEE Signal Processing Magazine, vol. 27, no. 5, 2010, pp. 34-45. [3] C. Orrite, M. Rodriguez, E. Herrero, G. Rogez, and S.A. Velastin, “Automatic segmentation and recognition of human actions in monocular sequences,” in Proc. IEEE International Conference on Pattern Recognition (ICPR), 2014 , pp. 4218-4223. [4] S. Samanta and B. Chanda, “Space-time facet model for human activity classification,” IEEE Transactions On Multimedia, vol. 16, no. 6, 2014, pp. 1525-1535. [5] A. Jayabalan1, H. Karunakaran1, S. Murlidharan1, and T. Shizume, “Dynamic action recognition: A convolutional neural network model for temportally organized joint location data,” arXiv:1612.06703 [cs.CV]. [6] S. Karaman, L. Seidenari, and A. Del Bimbo, “Fast saliency based pooling of Fisher encoded dense trajectories,” in ECCV THUMOS Workshop, vol. 1, 2014, p. 6. [7] Y. Gang and J. Yuan, "Fast action proposals for human action detection and search," in Procs. of the IEEE Conference on Computer Vision and Pattern Recognition, 2015, pp. 1302-1311. [8] G. Gkioxari and J. Malik, “Finding action tubes,” in Procs. of the IEEE Conference on Computer Vision and Pattern Recognition, 2015, pp. 759-768. [9] M. Shugao, L. Sigal, and S. Sclaroff, "Learning activity progression in lstms for activity detection and early detection," in Procs. of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 1942-1950. [10] S.-C. Cheng, J.-Y. Su, K.-F. Hsiao, Habib F. Rashvand, “Latent semantic learning with time-series cross correlation analysis for video scene detection and classification,” Multimedia Tools Appl., vol. 75, issue 20, 2016, pp. 12919-12940. [11] M. Hasan and A. K. Roy-Chowdhury, “A continuous learning framework for activity recognition using deep hybrid feature models,” IEEE Trans. Multimedia, vol. 17, no. 11, 2015, pp. 1909-1922. [12] Li-Qun Xu and Yongmin Li, “Video Classification Uing Spatial-Temporal Features And PCA,” International Conference on Multimedia and Expo, vol.3,2003,pp485-8 [13] Wei-Hao Lin and Alexander Hauptmann, “News Video Classification Using SVM-based Multimodal Classifiers and Combination Strategies.” the tenth ACM international conference on Multimedia ,2002,pp. 323-326 [14] Zhipeng Liu, Xiujuan Chai, Zhuang Liu, Xilin Chen, “Continuous Gesture Recognition with Hand-oriented Spatiotemporal Feature.” , IEEE International Conference on Computer Vision Workshops ,2017,pp.3056-3064 [15] Pavlo Molchanov, Xiaodong Yang, Shalini Gupta, Kihwan Kim, Stephen Tyree and Jan Kautz, ”Online Detection and Classification of Dynamic Hand Gestures with Recurrent 3D Convolutional Neural Networks”, The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 4207-4215 [16] Qian Xie , Oussama Remil, Yanwen Guo , Meng Wang , Senior Member, IEEE, Mingqiang Wei , and Jun Wang, “Object Detection and Tracking Under Occlusion for Object-Level RGB-D Video Segmentation.” IEEE Transactions on Multimedia,2017,pp.580-592 [17] Hong Liu,Jun Wang,Xiangdong Wang and Yueliang Qian, “Efficient Multi-scale Plane Extraction Based RGBD Video Segmentation.” International Conference on Multimedia Modeling,2017,pp.614-625 [18] Yeong-Seok Kim, Jong-Chul Yoon and In-Kwon Lee, “Real-time human segmentation from RGB-D video sequence based on adaptive geodesic distance computation.” Multimedia Tools and Applications • November 2017,pp.1-13 [19] Max Schwarz, Anton Milan, Arul Selvam Periyasamy, , , Sven Behnke, Max Schwarz, Anton Milan, Arul Selvam Periyasamy, Sven Behnke, “RGB-D object detection and semantic segmentation for autonomous manipulation in clutter.” The International Journal of Robotics Research,2017 [20] A. Jain, Amir R. Zamir, S. Savarese, and A. Saxena, “ Structural-RNN: Deep Learning on Spatio-Temporal Graphs,” n Proc. Intl. Conf. CVPR, 2016 [21] Bingbing Ni Gang Wang and Pierre Moulin, “RGBD-HuDaAct: A Color-Depth Video Database For Human Daily Activity Recognition”, IEEE ICCV Workshops, 2011 [22] Hema Swetha Koppula, Rudhir Gupta and Ashutosh Saxena, ”Learning human activities and object affordances from RGB-D videos,” The International Journal of Robotics Research ,2013,pp. 951–970 [23] Urbano Miguel Nunes,Diego R. Faria and Paulo Peixoto, “A human activity recognition framework using max-min features and key poses with differential evolution random forests classifier.” Pattern Recognition Letters,Volume 99, 2017, pp. 21-31 [24] Saeed Ghodsi, Hoda Mohammadzade and Erfan Korki, “Simultaneous Joint and Object Trajectory Templates for Human Activity Recognition from 3-D Data.” Cornell University, Computer Vision and Pattern Recognition (cs.CV),2017 [25] AhmadJalalaYeon-HoKimaYong-JoongKimaShaharyarKamalbDaijinKima, “Robust human activity recognition from depth video using spatiotemporal multi-fused features.” Pattern Recognition,Volume 61,2017,pp. 295-308 [26] B. Zhang, L. Wang, Z. Wang, Y. Qiao, and H. Wang, ”Real-time action recognition with enhanced motion vector CNNs,” in Proc. Intl. Conf. CVPR, 2016. [27] J. Y.-H. Ng, M. Hausknecht, S. Vijayanarasimhan, O. Vinyals, R. Monga, and G. Toderici. “Beyond short snippets: Deep networks for video classification,” in Proc. CVPR,2015, pp. 4694–4702. [28] S. Ma, L. Sigal, and S. Scalroff, “Learning activity progression in LSTMs for activity detection and early detection,” in Proc. Intl. Conf. CVPR, 2016. [29] M. Sabokrou, M. Fayyaz, and M. Fathy, “Deep-Cascade: Cascading 3D Deep Neural Networks for Fast Anomaly Detection and Localization in Crowded Scenes,” IEEE Transactions on Image Processing, Vol. 26, Issue: 4, pp. 1992-2004, April 2017 [30] F. Murtaza, M. H. Yousaf, and S.A. Velastin, “Multi-view human action recognition using histograms of oriented gradients (HOG) description of motion history images (MHIs),” in Proc. 13th IEEE International Conference on Frontiers of Information Technology (FIT), 2015, pp. 297-302. [31] C.-Y. Hsieh, S.-C. Cheng, C.-C Chang, and C.-L. Lin, “Automatic liver segmentation from CT images using latent semantic indexing,” in Proc. IEEE MMSP, 2015,pp. 1-6 [32] W.-K. Huang, C.-H. Chung, S.-C. Cheng, and J.-W. Hsieh, “Fast cube-based video shot retrieval using 3D moment-preserving technique,” in Proc. IEEE Conf. Image Processing, 2009. [33] H. Hoffmann, “Kernel PCA for novelty detection,” Pattern Recognition, vol. 40, 2007, pp. 863-874. [34] M. Barnachon, S. Bouakaz, B. Boufama, and E. Guillou, “Ongoing human action recognition with motion capture,” Pattern Recognition, vol. 47, no. 1, 2014, pp. 238–247. [35] http://pr.cs.cornell.edu/humanactivities/data.php [36] Cornell Activity Datasets: CAD-60 & CAD-120, http://pr.cs.cornell.edu/humanactivities/data.php, accessed online March 7 2017. [37] H. S. Koppula and A. Saxena, “Learning spatio-temporal structure from RGB-D videos for human activity detection and anticipation,” in Proc. International Conference on Machine Learning (ICML), 2013. [38] Y. Jiang and A. Saxena, “Low-dimensional Modeling of Humans in Environment Context for Activity Anticipation,” Robotics: Science and Systems (RSS), 2014.
|