[1] 王忠义,夏立新,李玉海.基于知识内容的数字图书馆跨学科多粒度知识表示模型构建[J].中国图书馆学报,2019,45(6):50-64. [2] 化柏林.学术论文中方法知识元的类型与描述规则研究[J].中国图书馆学报,2016,42(1):30-40. [3] STORRER A, WELLINGHOFF S. Automated detection and annotation of term definitions in German text corpora[C]//LREC. Proceedings of the fifth international conference on language resources and evaluation. Genoa:European Language Resources Association, 2006:2373-2376. [4] CASTRO-SANCHEZ N, SIDOROV G. Analysis of definitions of verbs in an explanatory dictionary for automatic extraction of actants based on detection of patterns[C]//NLDB. NLDB 2010:Natural language processing and information systems. Cardiff:NLDB, 2010:233-239. [5] ESPONOSA-ANKE L,SAGGION H. Applying dependency relations to definition extraction[C]//METAIS E, MATHIEU R, TEISSEIRE M, et al. Natural language processing and information systems. New York:Springer, 2014:63-74. [6] HEARST M. Automatic acquisition of hyponyms from large text corpora[C]//Proceedings of the 14th conference on computational linguistics-Volume 2. Gothenburg:Association for Computational Linguistics, 1992:539-545. [7] BOELLA G, DI-CARO L. Extracting definitions and hypernym relations relying on syntactic dependencies and support vector machines[C]//Proceedings of the 51st annual meeting of the Association for Computational Linguistics (Volume 2:Short Papers). Sofia:Association for Computational Linguistics, 2013:532-537. [8] BORG C, ROSNER M, PACE G. Automatic grammar rule extraction and ranking for definitions[C]//Proceedings of the seventh international conference on language resources and evaluation. Valletta:European Language Resources Association, 2010:2577-2584. [9] JIN Y, MIN-YEN K, JUN-PING N, et al. Mining scientific terms and their definitions:a study of the ACL anthology[C]//Proceedings of the 2013 conference on empirical methods in natural language processing. Washington:Association for Computational Linguistics, 2013:780-790. [10] LI SL, XU B, CHUNG T. Definition extraction with LSTM recurrent neural networks[C]//Chinese computational linguistics and natural language processing based on naturally annotated big data. Yantai:Computational Linguistics Professional Committee of the Chinese Information Society of China, 2016:177-189. [11] ESPONOSA-ANKE L, SCHOCHAET S. Syntactically aware neural architectures for definition extraction[C]//Proceeding of the 2018 conference of the North American Chapter of the Association for Computational Linguistics:human language technologies, volume 2(short papers). New Orleans:Association for Computational Linguistics, 2018:378-385. [12] SHVETS A, WANNER L. Concept extraction using pointer-generator networks[J]. ArXiv,2020,abs/2008.11295. [13] AMARANDEI S, FLECAN A, IONI G, et al. Key biomedical concepts extraction[C]//Workshop on curative power of medical Data. Romania:EUROLAN, 2017:234-240. [14] YANG X, BIAN J, HOGAN WR, et al. Clinical concept extraction using transformers[J]. Journal of the American Medical Informatics Association, 2020, 27(12):1935-1942. [15] NAVIGLI R, VELARDI P. Learning word-class lattices for definition and hypernym extraction[C]//Proceedings of the 48th annual meeting of the Association for Computational Linguistics. Uppsala:ACL, 2010:1318-1327. [16] DEVLIN J, CHANG MW, LEE K, et al. Bert:Pre-training of deep bidirectional transformers for language understanding[J]. ArXiv,2018,abs/1810.04805. [17] GRAVES A, SCHMIDHUBER J. Framewise phoneme classification with bidirectional LSTM and other neural network architectures[J]. Neural Network, 2005, 18(5/6):602-610. [18] LAFFERTY J, MCCALLUM A, PEREIRA F. Conditional random fields:probabilistic models for segmenting and labeling sequence data[C]//Proceedings of the eighteenth international conference on machine learning. San Francisco:Association for Computer Machinery, 2001:282-289. [19] 毛明毅,吴晨,钟义信,等.加入自注意力机制的BERT命名实体识别模型[J].智能系统学报, 2020, 84(4):146-153. [20] JAWAHAR G, SAGOT B, SEDDAH D. What does BERT learn about the structure of language?[C]//Proceedings of the 57th annual meeting of the Association for Computational Linguistics. Florence:Association for Computational Linguistics, 2019:3651-3657. [21] VASWANI A, SHAZEER N, PARMAR N, et al. Attention is all you need[J]. ArXiv,2017,abs/1706.03762,2017. [22] HOCHREITER S, SCHMIDHUBER J. Long short-term memory[J]. Neural computation, 1997, 9(8):1735-1780. [23] HUANG Z, WEI X, KAI Y. Bidirectional LSTM-CRF models for sequence tagging[J]. ArXiv,2015,abs/1508.01991. [24] YAN H, DENG B, LI X, et al. TENER:Adapting transformer encoder for named entity recognition[J]. ArXiv,2019,abs/1911.04474.作者贡献说明:李雪思:负责论文撰写、模型试验;张智雄:负责思路指导,实验整体把控;刘欢:负责论文修改。 |