|
[1] S. Asur and B. A. Huberman, "Predicting the future with social media," in Proceedings of the 2010 IEEE/WIC/ACM International Conference on Web Intelligence and Intelligent Agent Technology-Volume 01, 2010. [2] D. Charles and R. Julie, Technical Analysis, 2006. [3] G. C. Cawley and N. L. C. Talbot, "On over-fitting in model selection and subsequent selection bias in performance evaluation," Journal of Machine Learning Research, vol. 11, pp. 2079-2107, 2010. [4] S. Nison, Beyond Candlesticks: New Japanese Charting Techniques Revealed, Wiley, 1994. [5] S. Nison, Japanese Candlestick Charting Techniques: A Contemporary Guide to the Ancient Investment Techniques of the Far East, New York Institute of Finance, 2001. [6] H. A. Latane and R. J. Rendleman Jr, "Standard deviations of stock price ratios implied in option prices," The Journal of Finance, vol. 31, pp. 369-381, 1976. [7] T. Kamo and C. Dagli, "Hybrid approach to the Japanese candlestick method for financial forecasting," Expert Systems with applications, vol. 36, pp. 5023-5030, 2009. [8] K. Martiny, "Unsupervised Discovery of Significant Candlestick Patterns for Forecasting Security Price Movements.," in KDIR, 2012. [9] E. Ahmadi, M. H. Abooie, M. Jasemi and Y. Z. Mehrjardi, "A nonlinear autoregressive model with exogenous variables neural network for stock market timing: The candlestick technical analysis," International Journal of Engineering, vol. 29, pp. 1717-1725, 2016. [10] E. Ahmadi, M. Jasemi, L. Monplaisir, M. A. Nabavi, A. Mahmoodi and P. A. Jam, "New efficient hybrid candlestick technical analysis model for stock market timing on the basis of the Support Vector Machine and Heuristic Algorithms of Imperialist Competition and Genetic," Expert Systems with Applications, vol. 94, pp. 21-31, 2018. [11] C.-F. Tsai and Z.-Y. Quan, "Stock prediction by searching for similarities in candlestick charts," ACM Transactions on Management Information Systems (TMIS), vol. 5, p. 9, 2014. [12] Z.-Y. Quan, "Stock prediction by searching similar candlestick charts," in Data Engineering Workshops (ICDEW), 2013 IEEE 29th International Conference on, 2013. [13] K.-i. Kamijo and T. Tanigawa, "Stock price pattern recognition-a recurrent neural network approach," in Neural Networks, 1990., 1990 IJCNN International Joint Conference on, 1990. [14] K. H. Lee and G. S. Jo, "Expert system for predicting stock market timing using a candlestick chart," Expert systems with applications, vol. 16, pp. 357-364, 1999. [15] J. T. Connor, R. D. Martin and L. E. Atlas, "Recurrent neural networks and robust time series prediction," IEEE transactions on neural networks, vol. 5, pp. 240-254, 1994. [16] G. Dorffner, "Neural networks for time series processing," in Neural network world, 1996. [17] P. J. Werbos and others, "Backpropagation through time: what it does and how to do it," Proceedings of the IEEE, vol. 78, pp. 1550-1560, 1990. [18] S. Hochreiter, "The vanishing gradient problem during learning recurrent neural nets and problem solutions," International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems, vol. 6, pp. 107-116, 1998. [19] K. Cho, B. Van Merriënboer, C. Gulcehre, D. Bahdanau, F. Bougares, H. Schwenk and Y. Bengio, "Learning phrase representations using RNN encoder-decoder for statistical machine translation," arXiv preprint arXiv:1406.1078, 2014. [20] X. Wang, W. Jiang and Z. Luo, "Combination of convolutional and recurrent neural network for sentiment analysis of short texts," in Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, 2016. [21] D. Tang, B. Qin and T. Liu, "Document modeling with gated recurrent neural network for sentiment classification," in Proceedings of the 2015 conference on empirical methods in natural language processing, 2015. [22] K. Tran, A. Bisazza and C. Monz, "Recurrent memory networks for language modeling," arXiv preprint arXiv:1601.01272, 2016. [23] A. Krizhevsky, I. Sutskever and G. E. Hinton, "Imagenet classification with deep convolutional neural networks," in Advances in neural information processing systems, 2012. [24] Y. Kim, "Convolutional neural networks for sentence classification," arXiv preprint arXiv:1408.5882, 2014. [25] D. Britz, "Understanding Convolutional Neural Networks for NLP," 2018. [Online]. Available: http://www.wildml.com/2015/11/understanding-convolutional-neural-networks-for-nlp/. [26] A. Krizhevsky and G. E. Hinton, "Using very deep autoencoders for content-based image retrieval.," in ESANN, 2011. [27] P. Vincent, H. Larochelle, Y. Bengio and P.-A. Manzagol, "Extracting and composing robust features with denoising autoencoders," in Proceedings of the 25th international conference on Machine learning, 2008. [28] H. Noh, S. Hong and B. Han, "Learning deconvolution network for semantic segmentation," in Proceedings of the IEEE international conference on computer vision, 2015. [29] P. Baldi, "Autoencoders, unsupervised learning, and deep architectures," in Proceedings of ICML workshop on unsupervised and transfer learning, 2012. [30] M. D. Zeiler and R. Fergus, "Visualizing and understanding convolutional networks," in European conference on computer vision, 2014. [31] M. D. Zeiler, D. Krishnan, G. W. Taylor and R. Fergus, "Deconvolutional networks," 2010. [32] N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever and R. Salakhutdinov, "Dropout: a simple way to prevent neural networks from overfitting," The Journal of Machine Learning Research, vol. 15, pp. 1929-1958, 2014. [33] T. Salimans and D. P. Kingma, "Weight normalization: A simple reparameterization to accelerate training of deep neural networks," in Advances in Neural Information Processing Systems, 2016. [34] J. Patel, S. Shah, P. Thakkar and K. Kotecha, "Predicting stock and stock price index movement using trend deterministic data preparation and machine learning techniques," Expert Systems with Applications, vol. 42, pp. 259-268, 2015. [35] C. Sun, A. Shrivastava, S. Singh and A. Gupta, "Revisiting unreasonable effectiveness of data in deep learning era," in Computer Vision (ICCV), 2017 IEEE International Conference on, 2017. [36] R. Rothe, "Applying deep learning to real-world problems," 2018. [Online]. Available: https://medium.com/merantix/applying-deep-learning-to-real-world-problems-ba2d86ac5837. [37] I. Sutskever, J. Martens and G. E. Hinton, "Generating text with recurrent neural networks," in Proceedings of the 28th International Conference on Machine Learning (ICML-11), 2011. [38] H. Sak, A. Senior and F. Beaufays, "Long short-term memory based recurrent neural network architectures for large vocabulary speech recognition," arXiv preprint arXiv:1402.1128, 2014. [39] R. Pascanu, T. Mikolov and Y. Bengio, "On the difficulty of training recurrent neural networks," in International Conference on Machine Learning, 2013. [40] S. Kombrink, T. Mikolov, M. Karafiát and L. Burget, "Recurrent neural network based language modeling in meeting recognition," in Twelfth annual conference of the international speech communication association, 2011. [41] Y. Goldberg, "Neural network methods for natural language processing," Synthesis Lectures on Human Language Technologies, vol. 10, pp. 1-309, 2017. [42] W. S. Cleveland and R. McGill, "Graphical perception: Theory, experimentation, and application to the development of graphical methods," Journal of the American statistical association, vol. 79, pp. 531-554, 1984. [43] D. Bahdanau, K. Cho and Y. Bengio, "Neural machine translation by jointly learning to align and translate," arXiv preprint arXiv:1409.0473, 2014.
|