|
[1]"ISO 8601", [Online] Available at: https://zh.wikipedia.org/zh-tw/ISO_8601, [Accessed: January. 2024] [2]Khaled Shaalan, "Rule-based Approach in Arabic Natural Language Processing", International Journal on Information and Communication Technologies, Vol. 3, No. 3, June 2010, [Online] Available at: https://m.marefa.org/w/images/1/17/Rule_based_Arabic_NLP.pdf, [Accessed: January. 2024] [3]Sudha Morwal, Nusrat Jahan, Deepti Chopra, "Named Entity Recognition using Hidden Markov Model (HMM)", Internation“l Journal on Natural Language Computing (IJNLC) Vol. 1, ”o.4, December 2012, [Online] Available at: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3758852, [Accessed: January. 2024] [4]J. Lafferty, A. McCallum, and F. C. Pereira, "Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data", 2001, [Online] Available at: https://repository.upenn.edu/entities/publication/c9aea099-b5c8-4fdd-901c-15b6f889e4a7, [Accessed: January. 2024] [5]Silvestri, S., Esposito, A., Gargiulo, F., Sicuranza, M., Ciampi, M., & De Pietro, G., "A Big Data Architecture for the Extraction and Analysis of EHR Data", 2019 IEEE World Congress on Services (SERVICES), [Online] Available at: https://ieeexplore.ieee.org/document/8817262, [Accessed: January. 2024] [6]L. Li, L. Jin, Z. Jiang, D. Song, and D. Huang, "Biomedical named entity recognition based on extended Recurrent Neural Networks", 2015 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), [Online] Available at: https://ieeexplore.ieee.org/abstract/document/7359761, [Accessed: January. 2024] [7]A. Graves, "Long short-term memory. Supervised sequence labelling with recurrent neural networks", Studies in Computational Intelligence (SCI, volume 385) 2012, [Online] Available at: https://link.springer.com/book/10.1007/978-3-642-24797-2, [Accessed: January. 2024] [8]Zhiheng Huang, Wei Xu, Kai Yu, "Bidirectional LSTM-CRF Models for Sequence Tagging", arXiv:1508.019“1v1 [cs.CL] 9 Aug 2015, [Online] Available at: https://arxiv.org/pdf/1508.01991.pdf , [Accessed: January. 2024] [9]Łukasz Kaiser, Samy Bengio, "Can Active Memory Replace Attention?", 30th Conference Neural Information Processing Systems (NIPS2016” ,Barcelona, Spain. [Online] Available at: https://proceedings.neurips.cc/paper_files/paper/2016/file/fb8feff253bb6c834deb61ec76baa893-Paper.pdf, [Accessed: January. 2024] [10]Nal Kalchbrenner, Lasse Espeholt, Karen Simonyan, Aaronvanden Oord, Alex Graves, Koray Kavukcuoglu, " Neural Machine Translation in Linear Time ", arXiv preprint arXiv:1610.10“99v2, 2017, [Online] Available at: https://arxiv.org/pdf/1610.10099.pdf , [Accessed: January. 2024] [11]Jonas Gehring, Michael Auli, David Grangier, Denis Yarats, Yann N. Dauphin, " Convolutional sequence to sequence learning", arXiv preprint arXiv:1705.03122v2, 2017, [Online] Available at: https://arxiv.org/pdf/1705.03122.pdf , [Accessed: January. 2024] [12]John F. Kolen, Stefan C. Kremer, "Gradient Flow in Recurrent Nets: The Difficulty of Learning LongTerm Dependencies", Wiley-IEEE Press, 2001, [Online] Available at: https://ieeexplore.ieee.org/document/5264952, [Accessed: January. 2024] [13]Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N.Gomez, Łukasz Kaiser, "Attention Is All You Need", arXiv preprint arXiv:1706.03762, [Online] Available at: https://arxiv.org/pdf/1706.03762.pdf, [Accessed: January. 2024] [14]Jianpeng Cheng, Li Dong, Mirella Lapata, "Long short-term memory-networks for machine reading", arXiv preprint arXiv:601.06733,2016, [Online] Available at: https://arxiv.org/pdf/1601.06733.pdf, [Accessed: January. 2024] [15]Ankur P. Parikh, Oscar Täckström, Dipanjan Das, Jakob Uszkoreit, "A Decomposable Attention Model for Natural Language Inference", Empirical Methods in Natural Language Processing, 2016, [Online] Available at: https://arxiv.org/pdf/1606.01933.pdf, [Accessed: January. 2024] [16]Romain Paulus, Caiming Xiong, Richard Socher, "A Deep Reinforced Model for Abstractive Summarization", arXiv preprint arXiv:1705.04304, 2017, [Online] Available at: https://arxiv.org/pdf/1705.04304.pdf, [Accessed: January. 2024] [17]Zhouhan Lin, Minwei Feng, Cicero Nogueira dos Santos, Mo Yu, Bing Xiang, Bowen Zhou, Yoshua Bengio, "A Structured Self-attentive Sentence Embedding", arXiv preprint arXiv:1703.03130, 2017, [Online] Available at: https://arxiv.org/pdf/1703.03130.pdf, [Accessed: January. 2024] [18]Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova, "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding", arXiv preprint arXiv:1810.04805, 2018, [Online] Available at: https://arxiv.org/pdf/1810.04805.pdf, [Accessed: January. 2024] [18] Andrea Setzer, and Robert Gaizauskas, "Annotating Events and Temporal Information in Newswire Texts", In Proceedings of the Second International Conference on Language Resources and Evaluation (LREC’00) , Athens, Greece. European Language Resources Association (ELRA), May 2000, [Online] Available at: http://www.lrec-conf.org/proceedings/lrec2000/pdf/321.pdf, [Accessed: January. 2024] [20]Naushad UzZaman, Hector Llorens, Leon Derczynski, James Allen, Marc Verhagen, and James Pustejovsky, "SemEval-2013 Task 1: TempEval-3: Evaluating Time Expressions, Events, and Temporal Relations", In Second Joint Conference on Lexical and Computational Semantics (*SEM), Volume 2: Proceedings of the Seventh International Workshop on Semantic Evaluation (SemEval 2013), pages 1–9, Atlanta, Georgia, USA. Association for Computational Linguistics, June 2013, [Online] Available at: https://aclanthology.org/S13-2001, [Accessed: January. 2024] [21]Wentao Ding, Jianhao Chen, Jinmao Li, and Yuzhong Qu, "Automatic rule generation for time expression normalization", In Findings of the Association for Computational Linguistics: EMNLP 2021, pages 3135–3144, Punta Cana, Dominican Republic. Association for Computational Linguistics, November 2021, [Online] Available at: https://aclanthology.org/2021.findings-emnlp.269, [Accessed: January. 2024] [22]Joseph L. Fleis“, "Fleiss' kappa", [Online] Available at: https://en.wikipedia.org/wiki/Fleiss%27_kappa, [Accessed: January. 2024] [23]Anthony J Viera, Joanne M Garret“, " Understanding interobserver agreement: the kappa statistic", I Fam Med, May 2005, [Online] Available at: https://www1.cs.columbia.edu/~julia/courses/CS6998/Interrater_agreement.Kappa_statistic.pdf, [Accessed: January. 2024] [24]Lutz Prechelt, "Early Stopping — But When?", Part of the Lecture Notes in Computer Science book series (LNTCS,volume 7700), 2012, [Online] Available at: https://link.springer.com/chapter/10.1007/978-3-642-35289-8_5, [Accessed: January. 2024] [25]Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, "Language Models are Unsupervised Multitask Learns", Computer Science, Linguistics, 2019, [Online] Available at: https://cdn.openai.com/better-language-models/language_models_are_unsupervised_multitask_learners.pdf, [Accessed: January. 2024] [26]Stella Biderman, Hailey Schoelkopf, Quentin Anthony, Herbie Bradley, Kyl’ O'Brien, Eric Hallahan, Mohammad Aflah Khan, Shivanshu Purohit, USVSN Sai Prashanth, Edward Raff, Aviya Skowron, Lintang Sutawika, Oskar van der Wal, "Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling", arXiv preprint arXiv:2304.01373, 2023, [Online] Available at: https://arxiv.org/pdf/2304.01373.pdf, [Accessed: January. 2024] [27]Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, Dario Amodei, "Language Models are Few-Shot Learners”, arXiv preprint arXiv:2005.14165, 2020, [Online] Available at: https://arxiv.org/pdf/2005.14165.pdf, [Accessed: January. 2024] [28]Andonian, Alex and Anthony, Quentin and Biderman, Stella and Black, Sid and Gali, Preetham and Gao, Leo and Hallahan, Eric and Levy-Kramer, Josh and Leahy, Connor and Nestler, Lucas and Parker, Kip and Pieler, Michael and Phang, Jason and Purohit, Shivanshu and Schoelkopf, Hailey and Stander, Dashiell and Songz, Tri and Tigges, Curt and Thérien, Benjamin and Wang, Phil and Weinbach, Samuel, "{GPT-NeoX-20B}: An Open-Source Autoregressive Language Model", arXiv preprint arXiv:2204.06745, 2022, [Online] Available at: https://arxiv.org/pdf/2204.06745.pdf, [Accessed: January. 2024] [29]Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, "LoRA: Low-Rank Adaptation of Large Language Models", arXiv preprint arXiv:2106.09685, [Online] Available at: https://arxiv.org/pdf/2106.09685.pdf, [Accessed: January. 2024] [30]Ilya Loshchilov, Frank Hutter, "Decoupled Weight Decay Regularization", arXiv preprint arXiv:1711.05101, [Online] Available at: https://arxiv.org/pdf/1711.05101.pdf, [Accessed: January. 2024] [31]Liu, J., Shen, D., Zhang, Y., Dolan, W. B., Carin, L., & Chen, W. (2022, May), "What Makes Good In-Context Examples for GPT-3?", In Proceedings of Deep Learning Inside Out (DeeLIO 2022): The 3rd Workshop on Knowledge Extraction and Integration for Deep Learning Architectures (pp. 100-114), [Online] Available at: https://arxiv.org/pdf/2101.06804.pdf, [Accessed: January. 2024]
|