|
[1] J. Sturm, N. Engelhard, F. Endres, W. Burgard, and D. Cremers, “A benchmark for the evaluation of RGB-D SLAM systems,” in Proceedings of the International Conference on Intelligent Robot Systems (IROS), Oct. 2012. [2] M. P. Martorell, A. Maki, S. Martull, Y. Ohkawa, and K. Fukui, “Towards a simulation driven stereo vision system,” in Proceedings of the 21st International Conference on Pattern Recognition (ICPR), Nov 2012, pp. 1038–1042. [3] A. Richtsfeld, T. Mörwald, J. Prankl, M. Zillich, and M. Vincze, “Segmentation of unknown objects in indoor environments,” in IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Oct 2012, pp. 4791–4796. [4] B. Oehler, J. Stueckler, J. Welle, D. Schulz, and S. Behnke, “Efficient multiresolution plane segmentation of 3d point clouds,” in International Conference on Intelligent Robotics and Applications (ICIRA). Springer, 2011, pp. 145–156. [5] A. Anand, H. S. Koppula, T. Joachims, and A. Saxena, “Contextually guided semantic labeling and search for three-dimensional point clouds,” The International Journal of Robotics Research, vol. 32, no. 1, pp. 19–34, 2013. [6] H. S. Koppula, A. Anand, T. Joachims, and A. Saxena, “Semantic labeling of 3D point clouds for indoor scenes,” in Proceedings of the 24th International Conference on Neural Information Processing Systems (NIPS). USA: Curran Associates Inc., 2011, pp. 244–252. [Online]. Available: http://dl.acm.org/citation.cfm?id=2986459.2986487 [7] A. Smolic, “3D video and free viewpoint video - from capture to display,” Pattern Recognition, vol. 44, no. 9, pp. 1958 – 1968, 2011, computer Analysis of Images and Patterns. [8] K. Muller, P. Merkle, and T. Wiegand, “3-D video representation using depth maps,” Proceedings of the IEEE, vol. 99, no. 4, pp. 643–656, Apr. 2011. [9] S.-F. Tsai, C.-C. Cheng, C.-T. Li, and L.-G. Chen, “A real-time 1080p 2D-to-3D video conversion system,” IEEE Transactions on Consumer Electronics, vol. 57, no. 2, pp. 915–922, May 2011. [10] J. Han, L. Shao, D. Xu, and J. Shotton, “Enhanced computer vision with Microsoft Kinect sensor: A review,” IEEE Transactions on Cybernetics, vol. 43, no. 5, pp. 1318–1334, Oct. 2013. [11] K. Venkataraman, D. Lelescu, J. Duparré, A. McMahon, G. Molina, P. Chatterjee, R. Mullis, and S. Nayar, “PiCam: An ultra-thin high performance monolithic camera array,” ACM Trans. Graph., vol. 32, no. 6, pp. 166:1–166:13, Nov. 2013. [12] G. Sullivan, J. Ohm, W.-J. Han, and T. Wiegand, “Overview of the high efficiency video coding (HEVC) standard,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 22, no. 12, pp. 1649–1668, Dec. 2012. [13] A. Vetro, T. Wiegand, and G. Sullivan, “Overview of the stereo and multiview video coding extensions of the H.264/MPEG-4 AVC standard,” Proceedings of the IEEE, vol. 99, no. 4, pp. 626–642, Apr. 2011. [14] C. Fehn, “Depth-image-based rendering (DIBR), compression, and transmission for a new approach on 3D-TV,” in Proc. SPIE 5291, Stereoscopic Displays and Virtual Reality Systems XI, 2004. [15] C. Wheatstone, “Contributions to the physiology of vision.–part the first. on some remarkable, and hitherto unobserved, phenomena of binocular vision,” Philosophical transactions of the Royal Society of London, pp. 371–394, 1838. [16] R. I. Hartley and A. Zisserman, Multiple View Geometry in Computer Vision, 2nd ed. Cambridge University Press, ISBN: 0521540518, 2004. [17] N. Snavely, S. M. Seitz, and R. Szeliski, “Photo tourism: Exploring photo collections in 3D,” ACM Trans. Graph., vol. 25, no. 3, pp. 835–846, Jul. 2006. [18] K.-C. Wei, Y.-L. Huang, and S.-Y. Chien, “Point-based model construction for free-viewpoint TV,” in IEEE International Conference on Consumer Electronics - Berlin (ICCE-Berlin), Sept. 2013, pp. 220–221. [19] R. A. Newcombe, S. Izadi, O. Hilliges, D. Molyneaux, D. Kim, A. J. Davison, P. Kohi, J. Shotton, S. Hodges, and A. Fitzgibbon, “Kinectfusion: Real-time dense surface mapping and tracking,” in Proceedings of the 10th IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Oct 2011, pp. 127–136. [20] S. Izadi, D. Kim, O. Hilliges, D. Molyneaux, R. Newcombe, P. Kohli, J. Shotton, S. Hodges, D. Freeman, A. Davison, and A. Fitzgibbon, “Kinectfusion: Real-time 3D reconstruction and interaction using a moving depth camera,” in Proceedings of the 24th Annual ACM Symposium on User Interface Software and Technology (UIST), 2011, pp. 559–568. [21] J. Shen and S.-C. Cheung, “Layer depth denoising and completion for structured-light RGB-D cameras,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2013, pp. 1187–1194. [22] D. Miao, J. Fu, Y. Lu, S. Li, and C. W. Chen, “Texture-assisted Kinect depth inpainting,” in Proceedings of IEEE International Symposium on Circuits and Systems (ISCAS), May 2012, pp. 604–607. [23] L. Chen, H. Lin, and S. Li, “Depth image enhancement for Kinect using region growing and bilateral filter,” in Proceedings of the 21st International Conference on Pattern Recognition (ICPR), Nov 2012, pp. 3070–3073. [24] J. Liu, X. Gong, and J. Liu, “Guided inpainting and filtering for Kinect depth maps,” in Proceedings of the 21st International Conference on Pattern Recognition (ICPR), Nov 2012, pp. 2055–2058. [25] A. Telea, “An image inpainting technique based on the fast marching method,” Journal of Graphics Tools, vol. 9, no. 1, pp. 23–34, 2004. [26] P. K. Nathan Silberman, Derek Hoiem and R. Fergus, “Indoor segmentation and support inference from RGBD images,” in European Conference on Computer Vision (ECCV), 2012. [27] A. Boev, D. Hollosi, A. Gotchev, and K. Egiazarian, “Classification and simulation of stereoscopic artifacts in mobile 3DTV content,” in Proc. SPIE 7237, Stereoscopic Displays and Applications XX, 2009. [28] G. Zhang, J. Jia, T. T. Wong, and H. Bao, “Consistent depth maps recovery from a video sequence,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 31, no. 6, pp. 974–988, June 2009. [29] C. H. Hung, L. Xu, and J. Jia, “Consistent binocular depth and scene flow with chained temporal profiles,” International Journal of Computer Vision, pp. 1–22, Augest 2012. [30] S. P. Lipshitz, R. A.Wannamaker, and J. Vanderkooy, “Quantization and dither: A theoretical survey,” Journal of the Audio Engineering Society, vol. 40, no. 5, pp. 355–375, May 1992. [31] S. Boyd and L. Vandenberghe, Convex Optimization. Cambridge University Press, 2004. [32] Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image quality assessment: From error visibility to structural similarity,” IEEE Transactions on Image Processing, vol. 13, no. 4, pp. 600–612, April 2004. [33] R. Schnabel, R. Wahl, and R. Klein, “Efficient RANSAC for point-cloud shape detection,” in Computer graphics forum, vol. 26, no. 2. Wiley Online Library, 2007, pp. 214–226. [34] D. Holz and S. Behnke, Fast Range Image Segmentation and Smoothing Using Approximate Surface Reconstruction and Region Growing. Berlin, Heidelberg: Springer Berlin Heidelberg, 2013, pp. 61–73. [Online]. Available: http://dx.doi.org/10.1007/978-3-642-33932-5_7 [35] S. C. Stein, M. Schoeler, J. Papon, and F. Wörgötter, “Object partitioning using local convexity,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2014, pp. 304–311. [36] J. Papon, A. Abramov, M. Schoeler, and F. Wörgötter, “Voxel cloud connectivity segmentation - supervoxels for point clouds,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2013, pp. 2027–2034. [37] D. G. Lowe, “Distinctive image features from scale-invariant keypoints,” Int. J. Comput. Vision, vol. 60, no. 2, pp. 91–110, Nov. 2004. [Online]. Available: http://dx.doi.org/10.1023/B:VISI.0000029664.99615.94 [38] P. Henry, M. Krainin, E. Herbst, X. Ren, and D. Fox, “RGB-D mapping: Using Kinect-style depth cameras for dense 3D modeling of indoor environments,” Int. J. Rob. Res., vol. 31, no. 5, pp. 647–663, Apr. 2012. [Online]. Available: http://dx.doi.org/10.1177/0278364911434148 [39] M. Keller, D. Lefloch, M. Lambers, S. Izadi, T. Weyrich, and A. Kolb, “Real-time 3D reconstruction in dynamic scenes using point-based fusion,” in International Conference on 3D Vision (3DV), June 2013, pp. 1–8. [40] O. van Kaick, H. Zhang, G. Hamarneh, and D. Cohen-Or, “A survey on shape correspondence,” Computer Graphics Forum, vol. 30, no. 6, pp. 1681–1707, 2011. [41] B. K. Horn, “Closed-form solution of absolute orientation using unit quaternions,” JOSA A, vol. 4, no. 4, pp. 629–642, 1987. [42] R. Diaz, S. Hallman, and C. Fowlkes, “Detecting dynamic objects with multi-view background subtraction,” in IEEE International Conference on Computer Vision (ICCV), Dec. 2013, pp. 273–280. [43] M. Domañski, T. Grajek, K. Klimaszewski, M. Kurc, O. Stankiewicz, J. Stankowski, and K. Wegner, “Poznan multiview video test sequences and camera parameters,” in ISO/IEC JTC1/SC29/WG11 MPEG 2009/M17050, Xian, China, Oct. 2009. [44] F. Zhou, H.-L. Duh, and M. Billinghurst, “Trends in augmented reality tracking, interaction and display: A review of ten years of ISMAR,” in IEEE/ACM International Symposium on Mixed and Augmented Reality (ISMAR), Sept 2008, pp. 193–202. [45] P.-C. Wu, Y.-H. Tsai, and S.-Y. Chien, “Stable pose tracking from a planar target with an analytical motion model in real-time applications,” in IEEE International Workshop on Multimedia Signal Processing (MMSP), Sept 2014, pp. 1–6. [46] H. Kato and M. Billinghurst, “Marker tracking and HMD calibration for a video-based augmented reality conferencing system,” in IEEE/ACM International Workshop on Augmented Reality (IWAR), 1999, pp. 85–94. [47] G. Simon, A. Fitzgibbon, and A. Zisserman, “Markerless tracking using planar structures in the scene,” in IEEE and ACM International Symposium on Augmented Reality (ISAR 2000), oct 2000, pp. 120–128. [48] T. Lee and T. Hollerer, “Handy AR: Markerless inspection of augmented reality objects using fingertip tracking,” in IEEE International Symposium on Wearable Computers, Oct 2007, pp. 83–90. [49] K. Kim, N.-y. Park, and W. Woo, “Vision-based all-in-one solution for augmented reality and its storytelling applications,” The Visual Computer, vol. 30, no. 4, pp. 417–429, 2014. [50] A. O. Ulusoy and J. L. Mundy, “Image-based 4-d Reconstruction Using 3-d Change Detection,” in European Conference on Computer Vision (ECCV), ser. Lecture Notes in Computer Science, D. Fleet, T. Pajdla, B. Schiele, and T. Tuytelaars, Eds. Springer International Publishing, Sep. 2014, pp. 31–45. [51] Y.-L. Huang, K.-C.Wei, and S.-Y. Chien, “3D background modeling in multi-view RGB-D video,” in Proceedings of the ACM International Conference on Multimedia (Short Paper), ser. MM ’15. ACM, 2015. [52] S. Y. Chien,W. K. Chan, Y. H. Tseng, and H. Y. Chen, “Video object segmentation and tracking framework with improved threshold decision and diffusion distance,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 23, no. 6, pp. 921–934, June 2013. [53] C. L. Zitnick, S. B. Kang, M. Uyttendaele, S. Winder, and R. Szeliski, “High-quality video view interpolation using a layered representation,” ACM Trans. Graph., vol. 23, no. 3, pp. 600–608, Aug. 2004. [54] M. Bertalmio, A. L. Bertozzi, and G. Sapiro, “Navier-stokes, fluid dynamics, and image and video inpainting,” in Computer Vision and Pattern Recognition (CVPR), 2001. [55] K.-J. Oh, S. Yea, and Y.-S. Ho, “Hole filling method using depth based in-painting for view synthesis in free viewpoint television and 3-D video,” in Picture Coding Symposium, May 2009. [56] I. Daribo and H. Saito, “A novel inpainting-based layered depth video for 3DTV,” IEEE Transactions on Broadcasting, vol. 57, no. 2, pp. 533 – 541, June 2011. [57] K. Müller, A. Smolic, K. Dix, P. Merkle, P. Kauff, and T. Wiegand, “View synthesis for advanced 3D video systems,” EURASIP Journal on Image and Video Processing, vol. 2008, pp. 1–11, 2008. [58] Y. Mori, N. Fukushima, T. Yendo, T. Fujii, and M. Tanimoto, “View generation with 3D warping using depth information for FTV,” Journal of Image Communication, vol. 24, no. 1-2, pp. 65–72, January 2009. [59] D. Berjon, A. Hornung, F. Moran, and A. Smolic, “Evaluation of backward mapping DIBR for FVV applications,” in IEEE International Conference on Multimedia and Expo (ICME), July 2011. [60] S. J. Gortler, R. Grzeszczuk, R. Szeliski, and M. F. Cohen, “The lumigraph,” in SIGGRAPH, New York, NY, USA, 1996. [61] R. Marroquim, M. Kraus, and P. R. Cavalcanti, “Efficient point-based rendering using image reconstruction,” in Proceedings Symposium on Point-Based Graphics, 2007, pp. 101–108. [62] View Synthesis Software Manual, ISO/IEC JTC1/SC29/WG11, MPEG, September 2009, release 3.5. [63] S. P. McKee and D. G. Taylor, “The precision of binocular and monocular depth judgments in natural settings,” Journal of vision, vol. 10, no. 10, pp. 5–5, 2010. [64] Z. Wu, S. Song, A. Khosla, F. Yu, L. Zhang, X. Tang, and J. Xiao, “3D shapenets: A deep representation for volumetric shapes,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015, pp. 1912–1920. [65] I. Biederman, “Recognition-by-components: a theory of human image understanding.” Psychological review, vol. 94, no. 2, p. 115, 1987.
|