|
[1]D. M. Blei, A. Y. Ng, and M. I. Jordan, “Latent Dirichlet Allocation,” JMLR, vol. 3, pp. 993–1022, 2003. [2]D. P. Kingma and M. Welling, “Auto-Encoding Variational Bayes,” ICLR, 2014. [3]A. Srivastava and C. Sutton, “Autoencoding Variational Inference for Topic Models,” ICLR, 2017. [4]T. Hofmann and I. Computer, “Probabilistic latent semantic indexing,” in Proceedings of the 22nd annual international ACM SIGIR conference on Research and development in information retrieval, 1999, pp. 50–57. [5]D. Mimno, H. M. Wallach, E. Talley, M. Leenders, and A. McCallum, “Optimizing Semantic Coherence in Topic Models,” in Proceedings of the 2011 Conference on Empirical Methods in Natural Language Processing, no. 2, pp. 262–272, 2011. [6]R. Das, M. Zaheer, and C. Dyer, “Gaussian LDA for Topic Models with Word Embeddings,” in Proceedings ACL 2015, pp. 795–804, 2015. [7]D. Q. Nguyen, R. Billingsley, L. Du, and M. Johnson, “Improving topic models with latent feature word representations,” in Transactions of the Association for Computational Linguistics, vol. 3, pp. 299–313, 2015. [8]L. Niu, X. Dai, J. Zhang, and J. Chen, “Topic2Vec: Learning distributed representations of topics,” in Proceedings of 2015 International Conference on Asian Language Processing (IALP 2015), pp. 193–196, 2015 [9]C. E. Moody, “Mixing Dirichlet Topic Models and Word Embeddings to Make lda2vec,” 2016. [10]D. J. C. MacKay, “Choice of basis for Laplace approximation,” in Machine Learning, vol. 33, no. 1, pp. 77–86, 1998. [11]P. Henning, D. Stern, R. Herbrich, T. Graepel, and P. Hennig, “Kernel Topic Models,” Proc. 15th Int. Conf. Artif. Intell. Stat., vol. 22, pp. 511–519, 2012. [12]T. Mikolov, W.-T. Yih, and G. Zweig, “Linguistic regularities in continuous space word representations,” Proc. NAACL-HLT, no. June, pp. 746–751, 2013. [13]Pedregosa et al, “Scikit-learn: Machine Learning in Python,” JMLR, pp. 2825-2830, 2011. [14]C. Tan, D. Card, and N. A. Smith, “Friendships, Rivalries, and Trysts: Characterizing Relations between Ideas in Texts,” 2017. [15]McCallum, Andrew Kachites. “MALLET: A Machine Learning for Language Toolkit.” http://mallet.cs.umass.edu. 2002. [16]J. Pennington, R. Socher, and C. D. Manning, “GloVe: Global Vectors for Word Representation,” Proc. 2014 Conf. Empir. Methods Nat. Lang. Process., pp. 1532–1543, 2014. [17]D. Newman, J. Lau, K. Grieser, and T. Baldwin, “Automatic evaluation of topic coherence,” Hum. Lang. Technol. 2010 Annu. Conf. North Am. Chapter ACL, pp. 100–108, 2010. [18]J. H. Lau, D. Newman, and T. Baldwin, “Machine Reading Tea Leaves : Automatically Evaluating Topic Coherence and Topic Model Quality,” Proc. 14th Conf. Eur. Chapter Assoc. Comput. Linguist. (EACL 2014), pp. 530–539, 2014. [19]D. P. Kingma and J. B. Adam, “A method for stochastic optimization,” ICLR, 2015. [20]T. Mikolov, K. Chen, G. Corrado, and J. Dean, “Efficient Estimation of Word Representations in Vector Space,” in The Workshop Proceedings of the International Conference on Learning Representations (ICLR), 2013. [21]R. Salakhutdinov and G. E. Hinton, “Replicated softmax: An undirected topic model,” Adv. Neural Inf. Process. Syst. 22 - Proc. 2009 Conf., pp. 1607–1614, 2009. [22]G. E. Hinton, “Training products of experts by minimizing contrastive divergence,” Neural Comput., vol. 14, no. 8, pp. 1771–1800, 2002.
|