1.Akira, T., & Takenobu, T. (2001). Automatic disabbreviation by using context information. in Proceedings of the sixth natural language processing pacific rim symposium workshop on automatic paraphrasing:theories and applications, 21-28.
2.Berger, A., Pietra, S. D., & Pietra, V. D. (1996). A maximum entropy approach to natural language processing. Computational Linguistics, 22(1), 39-71.
3.Borthwick, A., Sterling, J., Agichtein, E., & Grishman, R. (1998). Exploiting diverse knowledge sources via maximum entropy in named entity recognition. Paper presented at the The Sixth Workshop on Very Large Corpora.
4.Brown, P., Pietra, S. D., Pietra, V. D., & Mercer, R. (1991). A statistical approach to sense disambiguation in machine translation. Paper presented at the DARPA Workshop on Speech and Natural Language.
5.Darroch, J. N., & Ratcliff, D. (1972). Generalized iterative scaling for log-linear models. The Annals of Mathematical Statistics, 43(5), 1470-1480.
6.Elmi, M. A., & Evens, M. (1998). Spelling correction using context.
7.Greiff, W. R., & Ponte, J. M. (2000). The maximum entropy approach and probabilistic ir models. ACM Transactions on Information Systems, 18, 246-287.
8.Jaynes, E. T. (1957). Information theory and statistical mechanics. Physical Review, 106(4), 620-630.
9.Kehler, A. (1997). Probabilistic coreference in information extraction. Paper presented at the Second Conference on Empirical Methods in Natural Language Processing.
10.Leah, L., Ogilvie, P., Price, A., & Tamilio, B. (2000). Acrophile:An automated acronym extractor and server. In Proceedings of the ACM digital libraries conference, 205-214.
11.LIN, H., & YUAN, C. F. (2002). Chinese part of speech tagging based on maximum entropy method. Paper presented at the First International Conference on Machine learning and Cybernetics,Beijing, Beijing.
12.Park, Y., & Byrd, R. J. (2001). Hybrid text mining for finding abbreviations and their definitions. In Proceedings of EMNP2001.
13.Pavlov, D. (2003). Sequence modeling with mixtures of conditional maximum entropy distributions. Paper presented at the Third IEEE International Conference on Data Mining(ICDM''03).
14.Pavlov, D., Popescul, A., Pennock, D., & Ungar, L. (2003). Mixtures of conditional maximum entropy models, Twentieth International Conference on Machine Learning(ICML-2003).
15.Pietra, V. D., Pietra, S. D., & Lafferty, J. (1995). Inducing features of random fields:Technical Report CMU-CS95-144,School of Computer Science,Carnegie-Mellon University.
16.Ratnaparkhi, A. (1996). A maximum entropy model for part-of-speech tagging. In conference on Empirical Methods in Natural Language Processing.
17.Ratnaparkhi, A. (1997). A simple introduction to maximum entropy models for natural language processing. Technical Report 97-08,Institute for Research in Cognitive Science,University of Pennsylvania.
18.Ratnaparkhi, A., JeffReynar, & Roulos, S. (1994). A maximum entropy model for prepositional phrase attachment. In Proceedings of the Human Language Technology Workshop(ARP,1994), 250-255.
19.Reynar, J. C., & Ratnaparkhi, A. (1997). A maximum entropy approach to identifying sentence boundaries. In Fifth Conference on Applied Natural Language Processing, 16-19.
20.Taghva, K., & Gilbreth, J. (1999). Recognizing acronyms and their definitions. International journal on document analysis and recognition(IJDAR), 191-198.
21.Terada, A., Tokunaga, T., & Tanaka, H. (2004). Automatic expansion of abbreviations by using context and character information. Information Processing and Management, 31-45.
22.Toole, J. (2000). A hybrid approach to the identification and expansion of abbreviations. In Proceedings of RIAO''2000, 1, 725-736.
23.賴育佐. (2003). 中文縮寫詞之機率統計模式. 國立暨南國際大學資訊工程學系碩士論文.