|
1.Devlin, J., Chang, M.W., Lee, K. and Toutanova, K. (2019) “ BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding ” arXiv:1810.04805v2. 2.Sanh, V., Debut, L., Chaumond, J. and Wolf, T. (2020) “ DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter ” , arXiv:1910.01108v4. 3.Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L. and Stoyanov, V. (2019) “ RoBERTa: A Robustly Optimized BERT Pretraining Approach ” , arXiv:1907.11692v1. 4.Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, R. and Le, Q.V. (2020) “ XLNet: Generalized Autoregressive Pretraining for Language Understanding ” , arXiv:1906.08237v2. 5.Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L. and Polosukhin, I. (2017) “ Attention Is All You Need ” , arXiv:1706.03762. 6.Hinton, G., Vinyals, O. and Dean, J. (2015) “ Distilling the Knowledge in a Neural Network ” , arXiv:1503.02531v1. 7.Dai, Z., Yang, Z., Yang, Y., Carbonell, J., Le, Q.N. and Salakhutdinov, R. (2019) “ Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context ” , arXiv:1901.02860v3. 8.Radford, A., Wu, J., Child, R., Luan, D., Amodei, D. and Sutskever, H.(2019) “ Language Models are Unsupervised Multitask Learners ” , https://github.com/openai/gpt-2. 9.Katharopoulos, A., Vyas, A., Pappas, N. and Fleuret, F. (2020) “ Transformers are RNNs: Fast Autoregressive Transformers with Linear Attention ” , arXiv:2006.16236v3. 10.Howard, J. and Ruder, S. (2018) “ Universal Language Model Fine-tuning for Text Classification ” , arXiv:1801.06146v5. 11.Xu, L., Zhang, X.W. and Dong. Q.Q. (2020) “ CLUECorpus2020: A Large-scale Chinese Corpus for Pre-training Language Model ” , arXiv:2003.01355v2. 12.吳晨皓、黃承龍,2020,BERT與GPT-2分別應用於刑事案件之罪名分類及判決書生成。 13.陳世榮,2012,“社會科學研究中的文字探勘應用:以文意為基礎的文件分類及其問題” 。 14.吳孟瑾、傅詞源、李佳衛、張耀中,2019, “利用自然語言處理進行自動新聞分類之研究” 。 15.Lee, M. (2019) “進擊的BERT:NLP 界的巨人之力與遷移學習” , https://leemeng.tw/attack_on_bert_transfer_learning_in_nlp.html. 16.Cui, Y.M. (2019) “中文XLNet” , https://github.com/ymcui/Chinese-XLNet. 17.Cui, Y.M. (2019) “中文BERT-wwm” , https://github.com/ymcui/Chinese-BERT-wwm. 18.Xu, L. (2019) “中文預訓練RoBERTa模型” , https://github.com/brightmart/roberta_zh. 19.Du, Z.Y. (2019) “GPT2-Chinese” , https://github.com/Morizeyao/GPT2-Chinese. 20.Yang, J.X. (2019) “GPT2-chitchat” , https://github.com/yangjianxin1/GPT2-chitchat.
|