跳到主要內容

臺灣博碩士論文加值系統

(44.222.134.250) 您好!臺灣時間:2024/10/08 03:56
字體大小: 字級放大   字級縮小   預設字形  
回查詢結果 :::

詳目顯示

: 
twitterline
研究生:李珮嘉
研究生(外文):Pei-jia Li
論文名稱:連續中文語音之情緒辨識
論文名稱(外文):Emotion Recognition from Continuous Mandarin Speech Signal
指導教授:包蒼龍包蒼龍引用關係
指導教授(外文):Tsang-Long Pao
學位類別:碩士
校院名稱:大同大學
系所名稱:資訊工程學系(所)
學門:工程學門
學類:電資工程學類
論文種類:學術論文
論文出版年:2006
畢業學年度:94
語文別:英文
論文頁數:45
中文關鍵詞:連續中文語音情緒辨識
外文關鍵詞:Emotion Recognition
相關次數:
  • 被引用被引用:1
  • 點閱點閱:333
  • 評分評分:
  • 下載下載:53
  • 收藏至我的研究室書目清單書目收藏:2
語言是溝通的橋樑,在社會中扮演著重要的角色。因此如何正確的表達情感是人際溝通中重要的一環。在本篇論文裡,我們嘗試去實作一個連續中文語音之情緒辨識的系統。在實驗過程中,選擇了梅爾頻率倒頻譜係數 (MFCC) 和線性預測倒頻譜係數 (LPCC) 為特徵參數值。 情緒資料庫包含了5 種情緒,分別是生氣,快樂,悲傷,厭煩,和一般。此外,利用端點偵測來對連續語音進行切割。分類器則是使用權重式離散型最近鄰居演算法。 在這研究過程中,連續的句子由短時距句子所組成。這些合成句子的辨識率為83%。
Language is a mean of communication and speech plays an important role in the society. So how to express emotion correctly is an important aspect in communication. In this thesis, emotion recognition from continuous Mandarin speech signal is implemented. In the experiment, Mel-Frequency Cepstral Coefficients (MFCC) and Linear Prediction Cepstral Coefficients (LPCC) are selected as the features used in the recognition. Five emotions are investigated, including anger, happiness, sadness, boredom, and neutral.
Endpoint detection is tried to segment the continuous speech. Weighted discrete K-nearest neighbor method is chosen as the classifier. In this research, the continuous sentences are composed by several short sentences with known emotion. The average recognition accuracy is 83% for these sentences.
CHAPTER 1 INTRODUCTION
1.1 Introduction
1.2 Motivation and Objective
1.3 Thesis Organization
CHAPTER 2 BACKGROUND
2.1 Emotion Category
2.2 Related Works on Emotional Speech Recognition
CHAPTER 3 CONTINUOUS SPEECH SEGMENTATION AND EMOTION RECOGNITION
3.1 Preprocessing
3.2 Continuous Speech Segmentation
3.3 Speech Database
3.4 Emotion Recognition
CHAPTER 4 EXPERIMENTAL RESULT
4.1 Experimental Environment
4.2 Experimental Results with Uniform Segmentation
4.3 Experimental Results of Segmentation with Silence
4.4 Experimental Results Using Segmentation with Endpoint Dtection
4.5 Recognition of Continuous Speech
CHAPTER 5 CONCLUSION AND FUTURE WORK
REFERENCE
[1]S. Furui, Digital Speech Processing, Synthesis, and Recognition, Marcel Dekker Inc, February 10, 1989
[2]N. Sebe, I. Cohen, T. Gevers, T.S. Huang, “Multimodal Approaches for Emotion Recognition: A Survey,” Proceedings of SPIE, Vol. 5670, pp. 56-67, January 2005
[3]馮觀富, 情緒心理學, 心理出版社, 2005
[4]Encyclopedia Britannica Online, http://www.britannica.com/
[5]D. Morrison, R. Wang, L.C. De Silva, W.L. Xu, “Real-time Spoken Affect Classification and its Application in Call-Centers,” Proceedings of the Third International Conference on Information Technology and Applications, Vol. 1, pp. 483-487, July 2005
[6]L. Vidrascu, L. Devillers, “Annotation and Detection of Blended Emotions in Real Human-Human Dialogs Recorded in a Call Center,” IEEE International Conference on Multimedia and Expo, pp. 944 – 947, July 2005
[7]C. Breazeal, “Emotive qualities in robot speech,” IEEE/RSJ International Conference on Intelligent Robots and Systems, Vol. 3, pp.1388- 394, 2001
[8]http://www.ai.mit.edu/projects/humanoid-robotics-group/index.html
[9]B. Schuller, G. Rigoll, M. Lang, “Speech Emotion Recognition Combining Acoustic Features and Linguistic Information in a Hybrid Support Vector Machine-Belief Network Architecture,” Proceedings of IEEE International Conference on Acoustics, Speech, and Signal Processing, Vol. 1, pp. 577-80, May 2004
[10]Y.M. Chen, “Investigating and Finding Meaningful Use Scenarios for Emotion-Aware Technologies,” 2006, http://www.iis.sinica.edu.tw/~kevinc/
[11]A. Ortony, T.J. Turner, “What's Basic about Basic Emotions,” Psychological Review, pp. 315-331, 1990
[12]D. Canamero, J. Fredslund, “I Show You How I Like You: Human-Robot Interaction through Emotional Expression and Tactile Stimulation,” http://www.daimi.au.dk/~chili/feelix/feelix.html, May 30, 2006
[13]http://changingminds.org/explanations/emotions/basic%20emotions.htm, May 30, 2006
[14]R. Cowie, E. Douglas-Cowie, N. Tsapatsoulis, G. Votsis, S. Kollias, W. Fellenz, J.G. Taylor, “Emotion Recognition in Human-Computer Interaction,” IEEE Signal Processing Magazine, Vol.18(1), pp.32 – 80, Jan 2001
[15]R. Tato, R. Santos, R. Kompe, J.M. Pardo, “Emotional Space Improves Emotion Recognition,” ICSLP, pp. 2029-2032, 2002
[16]J.H. Yeh, Emotion Recognition from Mandarin Speech Signals, Master Thesis, Tatung University, 2004
[17]J. Liscombe, J. Venditti, J. Hirschberg, “Classifying subject ratings of emotional speech using acoustic features,” Proceedings of EuroSpeech, Geneva, Switzerland ISCA Archive 8th European Conference on Speech Communication and Technology Geneva, Switzerland, pp. 725-728, September, 2003
[18]R. Cowie and E. Douglas-Cowie, “Automatic statistical analysis of the signal and prosodic signs of emotion in speech,” Proc. 4th Int. Conf. Spoken Language Processing, pp. 1989-1992, 1996
[19]F. Dellaert, T. Polzin, A. Waibel, “Recognizing Emotion in Speech,” ICSLP Proceedings of Fourth International Conference on Spoken Language, Vol. 3, pp. 1970-1973, Oct. 1996
[20]M.W. Bhatti, Y. Wang, L. Guan, “A neural network approach for human emotion recognition in speech,” Proceedings of the International Symposium on Circuits and Systems, Vol. 2, pp. 181-184, May 2004
[21]D. Ververidis, C. Kotropoulos, I. Pitas, “Automatic emotional speech classification,” IEEE International Conference on Acoustics, Speech, and Signal Processing, Vol. 1, pp. 593-596, May 2004
[22]Z.J. Chuang, C.H. Wu, “Emotion recognition using acoustic features and textual content,” IEEE International Conference on Multimedia and Expo, Vol. 1, pp. 53-56, June 2004
[23]C.M. Lee, S.S. Narayanan, “Toward Detecting Emotions in Spoken Dialogs,” IEEE Transactions on Speech and Audio Processing, VOL. 13, pp. 293-303, MARCH 2005
[24]S. Davis, P. Mermelstein, “Comparison of Parametric Representations for Monosyllabic Word recognition in Continuously Spoken Sentences,” IEEE Transactions on Acoustics, Speech, and Signal Processing, Vol. 28, pp. 357-366, Aug 1980
[25]T.L. Nwe, S.W. Foo, L.C. De Silva, “Detection of Stress and Emotion in Speech Using Traditional and FFT Based Log Energy Features,” Proceedings of the Joint Conference of the Fourth International Conference on Information, Communications and Signal Processing, Vol. 3, pp. 1619-1623, Dec. 2003
[26]D.N. Jiang, L.H. Cai , “Speech emotion classification with the combination of statistic features and temporal features,” IEEE International Conference on Multimedia and Expo, Vol. 3, pp. 1967-1970, June 2004
[27]J.J. Lu, Construction and Testing of a Mandarin Emotional Speech Database and Its Application, Master Thesis, Tatung University, 2004
[28]Y.H. Chang, Emotion Recognition and Evaluation of Mandarin Speech Using Weighted D-KNN Classification, Master Thesis, Tatung University, 2005
[29]O. Segawa, K. Takeda, F. Itakura, “Continuous speech recognition without end-point detection,” IEEE International Conference on Acoustics, Speech, and Signal Processing, Vol. 1, pp.245-248, May 2001
[30]V.K. Prasad, T. Nagarajan, H.A Murthy, “Continuous speech recognition using automatically segmented data at syllabic units,” 2002 6th International Conference on Signal Processing, Vol. 1, pp.235-238, Aug. 2002
[31]V.A. Petrushin, “Emotion in Speech: Recognition and Application to Call Centers,” Proceedings of the Conference on Artificial Neural Networks in Engineering, pp. 7-10, Nov. 1999
[32]L. Lu, D. Liu, H.J. Zhang “Automatic Mood Detection and Tracking of Music Audio Signals,” IEEE Transactions on Audio, Speech and Language Processing, Vol. 14, pp. 5-18, Jan. 2006
[33]A. M. Kondoz, Digital Speech: Coding for Low Bit Rate Communication Systems, John Wiley & Sons, 1994
[34]R. Gutierrez-Osuna, “Pattern Analysis for Machine Olfaction: A Review,” IEEE Sensors Journal, Vol. 2, pp. 189-202, June 2002
[35]T.L. Pao, Y.T. Chen, J.J. Lu and J.H. Yeh, “The Construction and Testing of a Mandarin Emotional Speech Database,” Proceeding of ROCLING XVI, pp. 355-363, Sep. 2004
[36]王小川, 語音訊號處理, 全華科技圖書, 2004
QRCODE
 
 
 
 
 
                                                                                                                                                                                                                                                                                                                                                                                                               
第一頁 上一頁 下一頁 最後一頁 top