跳到主要內容

臺灣博碩士論文加值系統

(44.221.73.157) 您好!臺灣時間:2024/06/20 12:01
字體大小: 字級放大   字級縮小   預設字形  
回查詢結果 :::

詳目顯示

我願授權國圖
: 
twitterline
研究生:邱國權
研究生(外文):Kuo-Chuan Chiu
論文名稱:基於臉部表情圖片之唱歌人像動畫系統
論文名稱(外文):A Singing-Face Animation System Based on Facial Images with Emotional Expressions
指導教授:鄭士康
指導教授(外文):Shyh-Kang Jeng
學位類別:碩士
校院名稱:國立臺灣大學
系所名稱:電機工程學研究所
學門:工程學門
學類:電資工程學類
論文種類:學術論文
論文出版年:2006
畢業學年度:94
語文別:中文
論文頁數:60
中文關鍵詞:人臉動畫圖形漸變合成情感合成
外文關鍵詞:Facial AnimationImage MorphingExpressive Synthesis
相關次數:
  • 被引用被引用:0
  • 點閱點閱:358
  • 評分評分:
  • 下載下載:0
  • 收藏至我的研究室書目清單書目收藏:1
本論文實作完成一套人臉歌唱動畫系統。此系統呈現出人物跟隨著背景音樂來唱歌,並且同時能有臉部表情的變化。我們使用多張臉部表情圖片作為表情資料庫,來合成出臉部的情緒表現。採用的表情圖為Paul Ekman博士所提出的六大基本情緒:快樂﹙happy﹚、憤怒﹙anger﹚、害怕﹙fear﹚、驚訝﹙surprise﹚、哀傷﹙sad﹚、厭惡﹙disgust﹚。歌唱時的發聲嘴形則藉由文字轉換語音系統,Microsoft Speech SDK,來找出相對應於歌詞的嘴形。接著利用文章中所敘述的方法來將表情圖與發聲嘴形做合成,即完成動畫中關鍵影格﹙key-frame﹚的製作。關鍵影格與關鍵影格之間的連續畫面,使用圖形漸形合成﹙image morphing﹚的技術來產生。將這些畫面串連起來,就是栩栩如生的人臉唱歌動畫。此系統除了能降低製造擬真歌唱動畫的成本,在娛樂性的應用方面也相當具有效果。
In this thesis, we implemented a singing-face animation system. In this system, the character sings a song with background music and can change his/her facial expressions simultaneously. We use several expression images to create facial emotional expressions. The images in emotional database are with six basic emotions: happy、anger、fear、surprise、sad and disgust, proposed by Dr. Paul Ekman. The visemes mapping to lyrics are obtained by a text-to-speech system, Microsoft speech SDK. Then, the approach proposed in this article is adopted to synthesize an expression image and a viseme. After that, generating a key-frame of the animation is finished. Using image morphing technique can generate the frames between two key-frames. These clips form a mimic singing-face animation. The proposed system not only lowers the cost of producing a realistic singing-face animation, but also quite powerful for recreational applications.
摘要 i
Abstract ii
目錄 iii
圖目錄 iv
表目錄 vi
第一章 緒論 1
1.1 引言 1
1.2 研究背景 2
1.3 系統概述 4
1.4 章節概要 5
第二章 圖形漸變合成 6
2.1 圖形漸變合成概述 6
2.2 場域圖形漸變合成法 10
2.3 徑向基底函數 15
2.4 多圖形漸變合成 17
第三章 臉部模型 23
3.1 臉部表情資料庫 23
3.2 發聲嘴形 24
3.3 人臉圖片合成 27
3.4 情緒與發聲嘴形合成 33
第四章 系統實作 44
4.1 系統整體架構 44
4.2 表情圖片產生 44
4.3 表情與嘴形合成 51
4.4 動畫的生成 53
第五章 結論 56
參考文獻 58
[1] T. Beier, S. Neely, ”Feature-based image metamorphosis,” Proceedings of the 19th annual conference on Computer graphics and interactive techniques, pp. 35-42, July 1992.
[2] D. Ruprecht, H. Muller, “Image warping with scattered data interpolation,” IEEE Computer Graphics and Applications, vol. 15, issue 2, pp. 37- 43, March 1995.
[3] S. Y. Lee , K. Y. Chwa , S. Y. Shin, “Image metamorphosis using snakes and free-form deformations,” Proceedings of the 22nd annual conference on Computer graphics and interactive techniques, pp. 439-448, September 1995
[4] V Zanella, O. Fuentes, “An Approach to Automatic Morphing of Face Images in Frontal View,” Proceedings of Mexican International Conference on Artificial Intelligence (MICAI), Lecture Notes in Artificial Intelligence 2972, pp. 679-687, 26-30 April 2004.
[5] S. Karungaru, M. Fukumi, N. Akamatsu, “Morphing face images using automatically specified features,” Proceedings of the 46th IEEE International Midwest Symposium on vol. 2, pp. 741-744, 27-30 December 2003.
[6] T.D. Bui, M. Poel, D. Heylen, A. Nijholt, “Automatic Face Morphing for Transferring Facial Animation,” Proceedings of the 6th IASTED International Conference on Computers, Graphics, and Imaging, pp.19-24, 13-16 August 2003.
[7] S. Lee, G. Wolberg, S. Y. Shin, “Polymorph: morphing among multiple images,” IEEE Computer Graphics and Applications, vol.18, issue 1, pp. 58 – 71, January 1998.
[8] G. Wolberg, “Image Morphing: A Survey,” The Visual Computer, vol. 14, pp. 360-372, 1998.
[9] T. Ezzat, T. Poggio, “Visual speech synthesis by morphing visemes,” International Journal of Computer Vision,, vol. 38, no.1, pp. 45-57, 2000.
[10] J. D. Edge, S. Maddock, “Image-based talking heads using radial basis functions,” Proceedings of IEEE Theory and Practice of Computer Graphics, pp. 74–80, June 2003.
[11] S. A. King, R. E. Parent, “Lip synchronization for song,” Proceedings of IEEE Computer Animation, pp.233-239, 19-21 June 2002.
[12] S. A. King, R. E. Parent, “Animating song,” Journal of Visualization and Computer Animation, vol. 15, no.1, pp. 53-61, March 2004.
[13] E. M. Caldognetto, P. Cosi, C. Drioli, G. Tisato, F. Cavicchio, "Coproduction of speech and emotions: visual and acoustic modifications of some phonetic labial targets," Proceedings of AVSP, Audio Visual Speech Processing, ISCA Workshop, St Jorioz, France, pp. 209-214, 4-7 September 2003.
[14] E. Bevacqua, M. Mancini, C. Pelachaud, “Speaking with Emotions,” Proceedings of the AISB Symposium on Motion, Emotion and Cognition, 2004
[15] I. C. Lin, J. S. Yeh, M. Ouhyoung, "Realistic 3D Facial Animation Parameters from Mirror-Reflected Multi-View Video," in IEEE Computer Animation 2001 Conference Proceedings, pp. 2-11, 2001.
[16] I. C. Lin, J. S. Yeh, M. Ouhyoung, "Extracting 3D facial animation parameters from multiview video clips," IEEE Computer Graphics and Applications, vol. 22, no. 6, pp. 72-80, 2002.
[17] I. C. Lin, M. Ouhyoung, "Mirror MoCap: Automatic and Efficient Capture of Dense 3D Facial Motion Parameters from Video," The Visual Computer, vol. 12, no. 6, pp. 355-372, 2005
[18] S. Kshirsagar, S. Garchery, G. Sannier, N. M. Thalmann, “Synthetic faces: Analysis and applications,” International Journal of Imaging Systems and Technology, Vol. 13, no. 1, pp. 65–73, 2003.
[19] S. Kshirsagar, T. Molet, and N. M. Thalmann, “Principal components of expressive speech animation,” Computer Graphics International, pp. 38-46, July 2001.
[20] W. S. Lee, M Escher, G. Sannier; N. Magnenat-Thalmann, ”MPEG-4 compatible faces from orthogonal photos,” Proceedings of IEEE Computer Animation, pp. 186 – 194, 26-29 May 1999.
[21] Y. Cao, W. C. Tien, P. Faloutsos, F. H. Pighin, ”Expressive speech-driven facial animation,” ACM Transactions on Graphics, vol. 24, issue 4, pp. 1283-1302, October 2005.
[22] Z. Deng, M. Bulut, U. Neumann, S. S. Narayanan, “Automatic
dynamic expression synthesis for speech animation,” Proceedings of IEEE 17th International Conference on Computer Animation and Social Agents (CASA), pp.267-274, July 2004.
[23] T.D. Bui, D. Heylen, A. Nijholt, “Combination of facial movements on a 3D talking head,”. Proceedings of Computer Graphics International, pp. 284-290, 2004.
[24] T.D. Bui, D. Heylen, A. Nijholt, M. Poel, “On combining the facial movements of a talking head,” Proceedings of Measuring Behavior. 5th International Conference on Methods and Techniques in Behavioral Research, pp.19-22, 2005.
[25] I. C. Lin, C. F. Huang, J. C. Wu, M. Ouhyoung, "A Low Bit-rate Web-enabled Synthetic Head with Speech-driven Facial Animation," Proceedings of Workshop on Computer Animation and Simulation, pp. 29-40, 2000.
[26] M.Eck, “Interpolation Methods for Reconstruction of 3D Surfaces from Sequences of Planar Slices,” CAD und Computergraphik, vol. 13, no. 5, pp. 109-120, February 1991.
[27] P. Ekman, W. V. Friesen, Unmasking the face: a guide to recognizing emotions facial cues, Prentice-Hall, 1975.
[28] http://www.mmk.ei.tum.de/~waf/fgnet/feedtum.html
[29] S.A. King, R.E. Parent, and B.L. Olsafsky. “An anatomically-based 3D parametric lip model to support facial animation and synchronized speech,” Proceedings of Deform, pp. 7-19, November 2000.
[30] C. Pelachaud, ”Emotion Expressiveness Embedded in Representation
Languages for ECAs,” IUT of Montreuil - University of Paris 8, 28 November 2003.
[31] M.M. Cohen, D.W. Massaro, “Modeling coarticulation in synthetic visual speech”, In N. M. Thalmann, and D. Thalmann, editors, Model and Technique in Computer Animation, pp. 139-156, Springer-Verlag, Tokyo, 1993.
[32] Z. Liu, Y. Shan, Z. Zhang, “Expressive Expression Mapping with Ratio Images,” Computer Graphics, Annual Conference Series, ACM SIGGRAPH, pp. 271-276, August 2001.
[33] P. H. Tu, I. C. Lin, J. S. Yeh, R. H. Liang, M. Ouhyoung, "Surface Detail Capturing for Realistic Facial Animation," Journal of Computer Science and Technology, vol. 19, no. 4, pp. 618-625, 2004.
[34] Q. Zhang, Z. Liu, B. Guo, H. Shum, “Geometry-driven photorealistic facial
expression synthesis,” Proceedings of ACM SIGGRAPH/Eurographics Symposium on Computer animation, 26-27 July 2003.
[35] M. Brand, “Voice puppetry,” In Computer Graphics, Annual Conference Series, ACM SIGGRAPH, pp. 22-28, August 1999.
[36] C. Bregler, M. Covell, M. Slaney, “Video rewrite: Driving visual speech with audio,” In Computer Graphics, ACM SIGGRAPH, pp. 353-360, August 1997.
[37] T.Ezzat, G.Geiger, T. Poggio, “Trainable videorealistic speech animation,” In Computer Graphics Annual Conference Series, ACM SIGGRAPH, pp. 388-398, August 2002.
QRCODE
 
 
 
 
 
                                                                                                                                                                                                                                                                                                                                                                                                               
第一頁 上一頁 下一頁 最後一頁 top