|
[1]王小川, “語音訊號處理,” 全華科技圖書, 2004. [2]S.Tarar, “Speech Analysis: Desktop Items Activation Using Dynamic Time Warping,” IEEE International Conference on Computer Science and Information Technology, 6, pp. 657 - 659, 2010. [3]T. Li, W. Xu, J. Pan, Y. Yan, “Improving Automatic Speech Recognizer of Voice Search using System Combination,” Fuzzy Systems and Knowledge Discovery, 4, pp. 477 - 480, 2009. [4]S. Boll, “Suppression of acoustic noise in speech using spectral subtraction,” IEEE Transactions on Acoustics Speech and Signal Processing, 27(2), pp. 113-120, 1979. [5]C. Plapous, C. Marro and P. Scalart, “Improved signal-to-noise ratio estimation for speech enhancement” IEEE Transactions on Acoustics Speech and Signal Processing, 14(6), pp. 2098-2108, 2006. [6]P. J. Moreno, B. Raj, and R. M. Stern, “A vector Taylor series approach for environment-independent speech recognition,” in Proceedings of IEEE International Conference on Acoustics Speech and Signal Processing, 2, pp. 733-736, 1996. [7]A. Davis, S. Y. Low, S. Nordholm, “A Subband Space Constrained Beamformer Incorporating VoiceActivity Detection,” in Proceedings IEEE International Conference on Acoustics Speech and Signal Processing, 3, pp. iii/65 - iii/68, 2005. [8]C. W. Hsu, L. S. Lee, “Higher Order Cepstral Moment Normalization for Improved Robust Speech Recognition,” IEEE Transactions on Audio Speech and Language Processing, 17(2), pp. 205 - 220, 2009. [9]H. Hermansky, N. Morgan, “RASTA processing of speech,” IEEE Transactions on Speech and Audio Processing, 2(4), pp. 578 - 589, 1994. [10]X. Huang, A. Acero and H. W. Hon, “Spoken language processing: a guide to theory, algorithm and system development,” Prentice Hall PTR, 2001. [11]R. Muralishankar,D. O'Shaughnessy, “A Comparative Analysis of Noise Robust Speech Features Extracted from All-pass based Warping with MFCC in a Noisy Phoneme Recognition,” The Third International Conference on Digital Telecommunications, pp. 180 - 185, 2008. [12]R. Schlider, I. Bezrukov, H. Wagner, H. Ney, “Gammatone features and feature combination for large vocabulary Speech recognition,” in Proceedings of IEEE International Conference on Acoustics Speech and Signal Processing, 4, pp. IV-649 - IV-652, 2007. [13]M. Zbancioc, M. Costin, “Using neural networks and LPCC to improve speech recognition,” in Proceedings of International Symposium on Signals Circuits and Systems, pp. 445 - 448, 2003. [14]Y. Peng,L. Mu,X. Kong, Z. Lin, L. Wang, “A Study On Echo Feature Extraction Based On The Modified Relative Spectra(RASTA) and Perception Linear Prediction(PLP) Auditory Model,” in Proceedings of IEEE International Conference on Intelligent Computing and Intelligent Systems, 2, pp. 657 - 661, 2010. [15]R. Muralishankar, A. Sangwan and D. O'Shaughnessy, “Warped discrete cosine transform cepstrum: a new feature for Speech processing,” in Proceedings of IEEE Workshop on Digital Object , pp.99 - 104, 2005. [16]S. Tiberewala and H. Hermansky, “Multiband and adaptation approaches to robust speech recognition,” in Proceedings of European Conference on Speech Communication and Technology, 25(1-3), pp. 2619-2622, 1997. [17]S. Yoshizawa, N. Hayasaka, N. Wada and Y. Miyanaga, “Cepstral gain normalization for noise robust speech recognition,” in Proceedings of IEEE International Conference on Acoustics Speech and Signal Processing, 1, pp. I-209-212, 2004. [18]F. Hilger and H. Ney, “Quantile based histogram equalization for noise robust large vocabulary speech recognition,” IEEE Transactions on Audio Speech and Language Processing, 14(3), pp. 845–854, 2006. [19]C. Kim, K. Kumar, and R. M. Stern, “Robust Speech Recognition using a Small Power Boosting Algorithm,” in Proceedings of IEEE Workshop on Automatic Speech Recognition & Understanding, pp. 243 - 248, 2009. [20]J. S. Lin, I. C. Liu, and J. W. Hung, “Small Power Reduction Technique in Noise-Robust Speech Recognition, ” submitted to IEICE transactions on Information and Systems, 2013 [21]H. G. Hirsch and D. Pearce, “The AURORA experimental framework for the performance evaluation of speech recognition systems under noisy conditions,” in Proceedings of the Automatic Speech Recognition: Challenges for the new Millenium, pp. 181-188, 2000. [22]S. Young, G. Evermann, M. Gales, T. Hain, D. Kershaw, X. Liu, G. Moore, J. Odell, D. Ollason, D. Povey, V. Valtchev and P. Woodland, “The HTK Book (for HTK Version 3.4),” Cambridge University Engineering Department, Cambridge, UK, 2006.
|