跳到主要內容

臺灣博碩士論文加值系統

(216.73.216.30) 您好!臺灣時間:2025/11/28 19:15
字體大小: 字級放大   字級縮小   預設字形  
回查詢結果 :::

詳目顯示

: 
twitterline
研究生:林威憲
研究生(外文):Wei-Xian Lin
論文名稱:基於物件之虛擬視訊會議之視訊處理
論文名稱(外文):Video Processing for an Object-Based Virtual Video Conferencing
指導教授:陳永昌陳永昌引用關係
指導教授(外文):Yung-Chang Chen
學位類別:碩士
校院名稱:國立清華大學
系所名稱:電機工程學系
學門:工程學門
學類:電資工程學類
論文種類:學術論文
論文出版年:2000
畢業學年度:88
語文別:英文
論文頁數:60
中文關鍵詞:物件虛擬視訊會議視訊處理
外文關鍵詞:Object-BasedVirtualVideo ConferencingVideo Processing
相關次數:
  • 被引用被引用:1
  • 點閱點閱:501
  • 評分評分:
  • 下載下載:13
  • 收藏至我的研究室書目清單書目收藏:0
在一般的多點視訊會議中,參加人員為眾人焦點所在,而非背景。並且運動量大的參加者會受到較多的注意。因此,我們提出一種基於物件特性將被限制之位元率分配給各物件的方法。我們希望讓複雜度高、運動量大的物件分配到較大的位元率,使之擁有較高的畫質。反之,複雜度較低、運動量較小的物件分配到較少的位元率,使之畫質降低。由於複雜度高、運動量大的物件會受較高的注意力,而複雜度較低、運動量較小的物件受較小注意力,因此整個畫面的視覺效果便能提升。而此方法根據ITU-T TMN8的位元率與失真模型推導出位元率與失真的最佳化之物件位元率分配,其中考慮了物件複雜度。另外,我們引入了物件運動量及物件大小的因素推導出聯合的位元率分配方法。實驗的結果證明了使用此方法之後,較引人興趣之物件的畫質被提昇,同時較不引人興趣之物件的畫質被降低。
本論文中,我們提出兩種基於物件之多點虛擬視訊會議環境以提供更真實的會議,分別為二維虛擬場景會議與三維虛擬環境會議。在二維虛擬場景會議中,物件被改變大小,並放置在預先設計的二維虛擬場景。而在三維虛擬環境會議中,物件根據其相對位置被組合放置在三維虛擬環境。本論文中,我們並提出一物件分割的程序來將物件由視訊影像中分離出來。其中,我們使用背景相減、形態學運算子、行列掃描等方法。
In a general multipoint video conferencing, the participants focus on the other participants rather than the background, and they will pay more attention on the active conferees. Therefore, we propose a bit-allocation method which may distribute the constrained bit-rates to each object based on its characteristics. From the rate and distortion model proposed in ITU-T TMN8, we derive an object-based R-D optimized bit allocation equation, which only considers spatial activity. Then we introduce the factors of temporal activity and object size into the bit allocation equation to derive the joint bit-allocation method. Simulation results show that the image quality of interest is improved while the image quality of less interest is degraded. In this thesis, two multipoint virtual conferencing environments are proposed to provide a more realistic conference, the conferencing with two-dimensional virtual scene and the conferencing with three-dimensional virtual environment. In the conferencing with two-dimensional virtual scene, the objects are resized and composed onto a pre-designed virtual scene. In the conferencing with three-dimensional virtual environment, the objects are manipulated onto a three-dimensional environment based on their specific locations. An object-segmentation procedure is also proposed in this thesis to discriminate the objects from the video sequence based on the background subtraction, morphological operations, and row-column scan method.
Chapter 1 Introduction2
1.1 Multipoint Video Conference2
1.2 Motivation2
1.3 Thesis Organization2
Chapter 2 H.263 Based Video Transcoding for Multipoint Video Conference2
2.1 Overview of H.263 Video Coding Standard2
2.2 Video Transcoding Architectures2
2.3 TMN-8 Rate and Distortion Model2
2.3.1 . Rate Model2
2.3.2 . Distortion Model2
2.3.3 Optimized Quantization Parameter2
Chapter 3 Object-based Video Transcoding with Chroma-Keying2
3.1 Object-based Video Processing Using Chroma-Keying2
3.2 R-D Optimized Bit Allocation for Video Objects2
3.3 Simulation Results2
Chapter 4 Object-based Virtual Video Conferencing Environments2
4.1 Overview of Proposed Object-based Virtual Video Conferencing Environments2
4.1.1 . Virtual video conferencing with two-dimensional scene2
4.1.2 . Virtual video conferencing in three-dimensional environment2
4.2 Object Segmentation Based on Background Subtraction Method2
4.3 Construction of three-dimensional Environment Using Open GL2
Chapter 5 Conclusions2
Bibliography2
[1]ITU-T Recommendation H.261, “Video codec for audiovisual services at p x 64 kbits/s,” Mar. 1993.
[2]ITU-T Recommendation H.263, “Video coding for low bit-rate communication,” Mar. 1997.
[3]ISO/IEC 13818-2 “Generic coding of moving pictures and associated audio”. (MPEG-2), Part 2: Video, Nov. 1993.
[4]ISO/IEC JTC1/SC29/WG11 “Coding of moving pictures and associated audio MPEG98/W2194”. (MPEG-4), Mar. 1998.
[5]M.-T. Sun, T.-D. Wu, and J.-N. Hwang, “Dynamic bit allocation in video combining for multipoint video conferencing,“ IEEE Trans. Circuit and Systems., vol. 45, no. 5, pp. 644-648, May. 1998.
[6]T.-D. Wu, “Rate conversion by transcoding for video composition in multipoint control unit, “ Ph.D. Dissertation, Dept. Elec. Eng., Univ. Washington, Seattle, WA, Jun. 1999.
[7]C.-W. Lin, “Video transcoding techniques for multipoint video conferencing,” Ph.D. Dissertation, Dept. Elec. Eng., Univ. Tsing Hua, Hsinchu, Taiwan, ROC, Jan. 2000.
[8]ITU-T/SG15, “Video codec test model, TMN8,” Portland, June 1997.
[9]Ming-Ting Sun and I-Ming Pao, “Multipoint Video Conferencing,” Visual Communication and Image Processing, Marcel Dekker, C.W. Chen and Y.Q. Zhang ed., 1997.
[10]J. Youn, M.-T. Sun, and C.-W. Lin “Motion Estimation for High-Performance Transcoders,” IEEE Trans. Consumer Electronics, vol. 44, pp. 649-658, Aug. 1998.
[11]Chia-Wen Lin, Jeongnam Youn, Yung-Chang Chen, and Ming-Ting Sun, "A Study on Video Transcoder Architectures," IEEE Int. Symp. Consumer Electronics, Oct. 1998, Taipei, Taiwan.
[12]A. Eleftheriadis and D. Anastassiou, “Constrained and General Dynamic Rate Shaping of Compressed Digital Video,” ICIP ’95, 1995.
[13]G. Keesman et al., “Transcoding of MPEG Bitstream,” Signal Proc.. Image Commun., pp. 481-500, 1996.
[14]P. A. A. Assuncao and M. Ghanbari, “A Frequency Domain Video Transcoder for Dynamic Bit Rate Reduction of MPEG-2 Bit Streams,” IEEE Trans. Circuits Syst. Video Technol., Vol. 8, No. 8, pp. 953-567, Dec. 1998.
[15]Jeongnam Youn, Ming-Ting. Sun and Chia-Wen Lin, “Adaptive Motion Vector Refinement for High Performance Transcoding,” IEEE Trans. On Multimedia, vol 1, No. 1, pp. 30-40 Mar. 1999.
[16]K. Ramchandran, A. Orterga, and M, Vetterli, “Bit Allocation for Dependent Quantization with Applications to Multiresolution and MPEG Video Voders,” IEEE Trans. Image Processing, vol. 3, pp. 533-545. Sep. 1994.
[17]L.-J. Lin, A. Ortega. and C.-C J. Kuo, “Rate Control Using Spline-Interpolated Rate-Distortion Characteristics,” in Proc.VCIP, Orlando, FL, Mar. 1996, pp. 111-122.
[18]ITU-T/SG15, “Video codec test model, TMN7,” Nice, Feb. 1997.
[19]T. Chiang and Y.-Q. Zhang, “A new rate control scheme using quadratic rate distortion model,” IEEE Trans. Circuits Syst. Video Technol., vol. 7, pp. 246-250, Feb. 1997.
[20]B. Tao, H. A. Peterson, and B. W. Dickinson, “A rate-quantization model for MPEG encoders,” in Proc. ICIP, Santa Barbara, CA, vol. Oct. 1997, pp.338-341.
[21]J. Ribas-Corbera and Shawmin Lei, “Rate Control in DCT Video Coding for Low-Delay Communications,” IEEE Trans. Circuits Syst. Video Technol., vol 9, No. 1, pp. 172-185, Feb. 1999.
[22]M. E. Lukacs and G. D. Boyer, and M. Mills, “The personal presence system experimental research prototype,” IEEE Int. Conf. Comm., vol. 2, pp. 1112-1116, Jun. 1996.
[23]M. E. Lukacs and G. D. Boyer, “A universal broadband multipoint teleconferencing service for the 21 century,” IEEE Comm. Magazine, vol. 33, no. 11, pp.36-43, Nov.
1995.
[24]T. Chen, C. T. Swain, and B. G. Hsakell, “ Coding of subregions for content-based scalable video,” IEEE Trans. Circuits Syst. Video Technol., vol. 7, no. 1, pp. 256-260, Feb. 1997.
[25]B.-R. Wen, “Region-oriented video coding with chroma-key-based shape coding,” Master Thesis, Dept. Electro. Eng., National Chiao Tung Univ., Hsinchu, Taiwan, Jun. 1997.
[26]G. Keesman, “Multi-program video compression using joint bit-rate control,” Philips Journal of Research, vol, 50, pp. 21-45, 1996.
[27]W.-C. Gu and D. W. Lin, “Joint rate-distortion coding of multiple videos,” IEEE Trans. Consumer Eletron., vol. 45, pp. 159-164, Feb. 1999.
[28]L. Wang and A. Vincent, “Bit allocation and constraints for joint coding of multiple video programs,” IEEE Trans. Circuits Syst. Video Technol., vol. 9, no. 6, pp. 949-959, Sep. 1999.
[29]J. I. Ronda et al., “Rate control and bit allocation for MPEG-4,” IEEE Trans. Circuits Syst. Video Technol., vol. 9, no. 8, pp. 1243-1258, Dec. 1999.
[30]C.-W. Lin, “Video transcoding techniques for multipoint video conferencing,” Ph.D. Dissertation, Dept. Elec. Eng., Univ. Tsing Hua, Hsinchu, Taiwan, ROC, Jan. 2000.
[31]I. Haritaoglu, D. Harwood, and L.S. Davis. W4: Who? When? Where? What? A real-time system for detecting and tracking people. In Proc. The third IEEE Int’l Conf. Automatic Face and Gesture Recognition (Nara, Japan), pages 222-227. IEEE Computer Society Press, Los Alamitos, Calif., 1998.
[32]N. Friedman and S. Russell. Image segmentation in video sequences: A probabilistic approach. In Proc. 13th Conf. Uncetainty in Artificial Intelligence. Morgan Kaufmann, 1997.
[33]J. Davis and A. Bobick. “The representation and recognition of action using temporal templates.” In Proc. The Computer Vision and Pattern Recognition, 1997.
[34]ITU-T Recommendation G.732.1, “Dual rate speech coders for multimedia communication transmitting at 5.3 & 6.3 kb/s,” 1996.
[35]ITU-T Recommendation H.323, “Visual telephone systems and terminal equipment for local area networks which provide a non-guaranteed quality of service”
[36]ITU-T Recommendation H.324, “Terminal for Low Bit Rate Multimedia Communication”.
QRCODE
 
 
 
 
 
                                                                                                                                                                                                                                                                                                                                                                                                               
第一頁 上一頁 下一頁 最後一頁 top