|
[1]D. Pye, “Content-Based Methods for the Management of Digital Music,” in Proc. IEEE Conf. Acoustics, Speech, Signal Processing (ICASSP, pp. 2437–2440, 2000. [2]Shih-Chuan Chiu and Man-Kwan Shan, “Computer Music Composition Based on Discovered Music Patterns,” in Proc. IEEE Conference on Systems, Man, Cybernetics, Taipei, Taiwan, 2006. [3]Tao Li and Mitsunori Ogihara, “Toward Intelligent Music Information Retrieval,” IEEE Transactions on Multimedia, vol. 8, no. 3, pp. 564–574, June 2006. [4]C. McKay, Automatic genre classification of MIDI recordings, Master’s thesis, McGill University, Canada, 2004. [5]M.F. McKinney, and J. Breebaart. “Features for audio and music classification”, in Proc. International Symposium on Music Information Retrieval. 2003. [6]Bozena Kostek, “Musical instrument classification and duet analysis employing music information retrieval techniques,” in Proc. of the IEEE, vol. 92, issue 4, pp. 721–729, Apr 2004. [7]J. Saunders, “Real time discrimination of broadcast speech/music,” in Proc. IEEE Conf. Acoustics, Speech, Signal Processing (ICASSP), pp. 993–996, 1996. [8]E. Scheirer and M. Slaney, “Construction and evaluation of a robust multifeature speech/music discriminator,” in Proc. IEEE Conf. Acoustics, Speech, Signal Processing (ICASSP), pp. 1331–1334, 1997. [9]D. Kimber and L.Wilcox, “Acoustic segmentation for audio browsers,” in Proc. Interface Conf., Sydney, Australia, July 1996. [10]A. L. Berenzweig and D. P. Ellis, “Locating singing voice segments within musical signals,” in Proc. Int. Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), Mohonk, NY, pp. 119–123, 2001. [11]G. Tzanetakis and P. Cook, “Musical Genre Classification of Audio Signals,” IEEE Transactions on Speech and Audio Processing, vol. 10, no. 5, July 2002. [12]C. Xu and N. C. Maddage, “Automatic Music Classification and Summarization,” IEEE Transactions on Speech and Audio Processing, vol. 13, no. 3, May 2005. [13]E. Wold, T. Blum, D. Keislar, and J. Wheaton, “Content-based classification, search, and retrieval of audio,” IEEE Multimedia, vol. 3, no. 2, 1996. [14]J. Foote, “Content-based retrieval of music and audio,” Multimed. Storage Archiv. Syst. , pp. 138–147, 1997. [15]G. Li and A. Khokar, “Content-based indexing and retrieval of audio data using wavelets,” in Proc. IEEE Conf. Multimedia Expo , pp. 885–888, 2000. [16]S. Li, “Content-based classification and retrieval of audio using the nearest feature line method,” IEEE Trans. Speech Audio Processing, vol. 8, pp. 619–625, Sept. 2000. [17]X. Shao, C. Xu. M. Kankanhalli, “Unsupervised classification of musical genre using hidden Markov model,” in IEEE Int. Conf. of Multimedia Expo (ICME), pp. 2023–2026, 2004. [18]F. Mörchen, A. Ultsch, M. Nöcker, and C. Stamm, “Databionic visualization of music collections according to perceptual distance,” In Proc. International Symposium on Music Information Retrieval, pp. 396–403, 2005. [19]王小川,語音訊號處理,台北:全華科技圖書有限公司,2004。 [20]http://neural.cs.nthu.edu.tw/jang/books/audioSignalProcessing/ [21]A. Ramalingam and S. Krishnan, “Gaussian Mixture Modeling of Short-Time Fourier Transform Features for Audio Fingerprinting,” IEEE Transactions on Information Forensics and Security, 2006. [22]A. Dempster, N. Laird, and D. Rubin, “Maximum likelihood from incomplete data via the EM algorithm,” J. R. Stati. Soc., vol. 39, pp. 1–38, 1977. [23]L. Rabiner and B. H. Juang, Fundamentals of Speech Recognition, Prentice-Hall, 1993. [24]Training part 1 and Training part 2, “Audio Genre Classification, MIREX 2005”, http://www.musicir.org/mirex/2005/index.php/Audio_Genre_Classification, 2005. [25]A. Solomonoff, A. Mielke, M. Schmidt, and H. Gish, “Clustering speakers by their voices,” in Proc. IEEE Conf. Acoust., Speech, Signal Process. (ICASSP), pp. 757–760, 1998. [26]L. Hubert and P. Arabie, “Comparing partitions,” J. Classification, vol. 2, pp. 193–218, 1985. [27]D. A. Reynolds, E. Singer, B. A. Carson, G. C. O’Leary, J. J. McLaughlin, and M. A. Zissman, “Blind clustering of speech utterances based on speaker and language characteristics,” in Proc. Int. Conf. Spoken Lang. Process. (ICSLP), pp. 3193–3196, 1998. [28]M. Turk and A. Pentland, “Eigenfaces for Recognition,” Journal of Cognitive Neuroscience, vol. 3, pp. 71–86, 1991. [29]W. H. Tsai and H. M. Wang, “Speech utterance clustering based on the maximization of within-cluster homogeneity of speaker voice characteristics,” J. Acoust. Soc. Amer., vol. 120, no. 3, pp. 1631–1645, 2006. [30]Christopher D. Manning, Prabhakar Raghavan and Hinrich Schütze, Introduction to Information Retrieval, Cambridge University Press. 2008.
|