|
[1]全球人壽退休資訊,政府關鍵報告,https://www.transglobe.com.tw/transglobe-retireplan/content/10400。 [2]Amazon Echo, https://www.amazon.com/Amazon-Echo-Bluetooth-Speaker-with-WiFi-Alexa/dp/B00X4WHP5E . [3]Google Home, https://madeby.google.com/intl/en_us/home/ . [4]Tunstall Healthcare, http://www.tunstallhealthcare.com.au/ . [5]AT&T EverThere, http://www.goodhousekeeping.com/health-products/health-tracker-reviews/a30201/a-t-and-t-everthere/ . [6]蓋德科技股份有限公司,http://www.guidercare.com/。 [7]SecuFirst居家看護系列,數位無線家居影音監視器,http://www.secufirst.com.tw/products_detail.aspx?Pid=19。 [8]Amazon Echo Teardown - iFixit, https://www.ifixit.com/Teardown/Amazon+Echo+Teardown/33953 . [9]Google Home Teardown - iFixit, https://www.ifixit.com/Teardown/Google+Home+Teardown/72684 . [10]K. S. R. Murty, and B. Yegnanarayana, “Combining evidence from residual phase and MFCC features for speaker recognition,” IEEE Signal Processing Letters, vol. 13, no. 1, pp. 52-55, 2006. [11]S. Chu, S. Narayanan, and C. -C. J. Kuo, “Environmental sound recognition with time-frequency audio features,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 17, no. 6, pp. 1142-1158, 2009. [12]H. D. Tran, and H. Li, “Sound event recognition with probabilistic distance SVMs,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 19, no. 6, pp. 1556-1568, 2011. [13]Delay Sum Beamforming - The Lab Book Pages, http://www.labbookpages.co.uk/audio/beamforming/delaySum.html . [14]H. Xia, K. Yang, Y. Ma, Y. Wang, and Y. Liu, “Noise reduction method for acoustic sensor arrays in underwater noise,” IEEE Sensors Journal, vol. 16, no. 24, pp. 8972-8981, 2016. [15]X. Alameda-Pineda, and R. Horaud, “A geometric approach to sound source localization from time-delay estimates,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 22, no. 6, pp. 1082-1095, 2014. [16]X. Shi, B. D. O. Anderson, G. Mao, Z. Yang, J. Chen, and Z. Lin, “Robust localization using time difference of arrivals,” IEEE Signal Processing Letters, vol. 23, no. 10, pp. 1320-1324, 2016. [17]J. Benesty, J. Chen, and Y. Huang, “Time-delay estimation via linear interpolation and cross correlation,” IEEE Transactions on Speech and Audio Processing, vol. 12, no. 5, pp. 509-519, 2004. [18]J. Benesty, J. Chen, and Y. Huang, “Microphone array signal processing,” Springer Science & Business Media, vol. 1, 2008. ISBN: 978-3-540-78612-2. [19]L. Wang, T. -K. Hon, J. D. Reiss, and A. Cavallaro, “An iterative approach to source counting and localization using two distant microphones,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 24, no. 6, pp. 1079-1093, 2016. [20]C. H. Knapp, and G. C. Carter, “The generalized correlation method for estimation of time delay,” IEEE Transactions on Acoustic, Speech, and Signal Processing, vol. 24, no. 4, pp. 320-327, 1976. [21]D. Pavlidi, A. Griffin, M. Puigt, and A. Mouchtaris, “Real-time multiple sound source localization and counting using a circular microphone array,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 21, no. 10, pp. 2193-2206, 2013. [22]J. -S. Hu, and C. -H. Yang, “Estimation of sound source number and directions under a multisource reverberant environment,” EURASIP Journal on Advances in Signal Processing, 2010. [23]E. C. Cherry, “Some experiments on the recognition of speech, with one ear and with two ears,” The Journal of the Acoustical Society of America, vol. 25, no. 5, pp. 975-979, 1953. [24]A. Ozerov, E. Vincent, and F. Bimbot, “A general flexible framework for the handling of prior information of audio source separation,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 20, no. 4, pp. 1118-1133, 2012. [25]T. Otsuka, K. Ishiguro, T. Yoshioka, H. Sawada, and H. G. Okuno, “Multichannel sound source dereverberation and separation for arbitrary number of sources based on Bayesian nonparametrics,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 22, no. 12, pp. 2218-2232, 2014. [26]M. Castella, and E. Moreau, “New kurtosis optimization schemes for MISO equalization,” IEEE Transactions on Signal Processing, vol. 60, no. 3, pp. 1319-1330, 2012. [27]Z. Koldovský, and P. Tichavský, “Time-domain blind separation of audio sources on the basis of a complete ICA decomposition of an observation space,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 19, no. 2, pp. 406-416, 2011. [28]J. -T. Chien, and H. -L. Hsieh, “Convex divergence ICA for blind source separation,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 20, no. 1, pp. 302-313, 2012. [29]K. Rahbar, and J. P. Reilly, “A frequency domain method for blind source separation of convolutive audio mixtures,” IEEE Transactions on Speech and Audio Processing, vol. 13, no. 5, pp. 832-844, 2005. [30]H. Saruwatari, T. Kawamura, T. Nishikawa, A. Lee, and K. Shikano, “Blind source separation based on a fast-convergence algorithm combining ICA and beamforming,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 14, no. 2, pp. 666-678, 2006. [31]K. Matsuoka, “Minimal distortion principle for blind source separation,” Proceedings of the 41st SICE Annual Conference, vol. 4, pp. 2138-2143, 2002. [32]Y. Zhang, K. Cao, K. Wu, T. Yu, and N. Zhou, “Audio-visual underdetermined blind source separation algorithm based on Gaussian potential function,” China Communications, vol. 11, no. 6, pp. 71-80, 2014. [33]M. S. Khan, S. M. Naqvi, A. -u. -Rehnam, W. Wang, and J. Chambers, “Video-aided model-based source separation in real reverberant rooms,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 21, no. 9, pp. 1900-1912, 2013. [34]J. B. Allen, and D. A. Berkley, “Image method for efficiently simulating small‐room acoustics,” The Journal of the Acoustical Society of America, vol. 65, no. 4, pp. 943-950, 1979. [35]https://github.com/ehabets/RIR-Generator [36]Speechnotes, https://speechnotes.co/ . [37]C. Loader, “Local regression and likelihood,” Springer Science & Business Media, 2006. [38]J. Nikunen, and T. Virtanen, “Direction of arrival based spatial covariance model for blind sound source separation,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 22, no. 3, pp. 727-739, 2014. [39]A. M. Kondoz, “Digital speech: coding for low bit rate communication systems,” John Wiley & Sons, 2005. [40]Z. Zivkovic, and F. Van Der Heijden, “Efficient adaptive density estimation per image pixel for the task of background subtraction,” Pattern recognition letters, vol. 27, no. 7, pp. 773-780, 2006.
|