图书情报工作 ›› 2022, Vol. 66 ›› Issue (14): 119-127.DOI: 10.13266/j.issn.0252-3116.2022.14.012

• 知识组织 • 上一篇    下一篇

基于混合深度学习的中文医学实体抽取研究

韩普1,2, 顾亮1   

  1. 1. 南京邮电大学管理学院 南京 210003;
    2. 江苏省数据工程与知识服务重点实验室 南京 210023
  • 收稿日期:2021-12-14 修回日期:2022-05-15 出版日期:2022-07-20 发布日期:2022-07-28
  • 作者简介:韩普,副教授,博士,硕士生导师,E-mail:hanpu@njupt.edu.cn;顾亮,硕士研究生。
  • 基金资助:
    本文系国家社会科学基金项目"大数据环境下健康领域实体语义挖掘研究"(项目编号: 17CTQ022)和江苏省研究生科研创新计划基金项目"基于深度学习的医学文献实体识别研究"(项目编号:KYCX21_0844)研究成果之一。

Research on Extraction of Chinese Medical Entities Based on Hybrid Deep Learning

Han Pu1,2, Gu Liang1   

  1. 1. School of Management, Nanjing University of Posts & Telecommunications, Nanjing 210003;
    2. Jiangsu Provincial Key Laboratory of Data Engineering and Knowledge Service, Nanjing 210023
  • Received:2021-12-14 Revised:2022-05-15 Online:2022-07-20 Published:2022-07-28

摘要: [目的/意义]医学实体抽取是医疗健康领域信息组织和知识挖掘的关键环节。针对中文医学实体专业性强、命名规则复杂和抽取难度大的现状,探究如何利用多种深度学习方法混合协作以提升中文医学实体抽取的准确性。[方法/过程]首先在深度学习模型BiLSTM-CRF基础上,引入语言模型BERT和迭代膨胀卷积神经网络IDCNN,增强文本语义表征能力和局部特征捕获能力;接着利用BERT预训练进行外部医学语料资源的知识迁移,实现多语义特征融合;然后引入自注意力机制捕获全局上下文重要信息,并加入Highway优化深层网络训练,解决网络加深导致的精度下降问题,最终提出MF-HDL (Multi Feature-Hybrid Deep Learning)模型。[结果/结论]MF-HDL模型在中文糖尿病数据集上效果显著,其F1值较基准模型IDCNN-CRF和BiLSTM-CRF分别提升18.42%和17.18%,此方法在中文医学实体抽取任务上表现优异。

关键词: 混合深度学习, 医学实体抽取, BERT, 注意力, Highway

Abstract: [Purpose/significance] Medical entity extraction is a key link in information organization and knowledge mining in the medical and health field. Aiming at the current situation of strong professionalism of Chinese medical entities, complex naming rules and difficulty in extraction, this paper explores how to use a variety of deep learning methods to mix and cooperate to enhance the accuracy of Chinese medical entity extraction. [Method/process] Firstly, on the basis of the deep learning model BiLSTM-CRF, this study introduced the language model BERT and iterative expanded convolutional neural network IDCNN to enhance the text semantic representation ability and local feature capture ability. Secondly, it utilized the BERT pre-training to transfer the knowledge of external medical corpus resources and realize the fusion of multiple semantic features. In addition, the self-attention mechanism was introduced to capture important global contextual information, and Highway was added to optimize deep network training to solve the problem of reduced accuracy caused by network deepening. Finally, MF-HDL model (Multi Feature-Hybrid Deep Learning) was proposed. [Result/conclusion] The MF-HDL model has a significant performance on the Chinese diabetes dataset. Compared with the benchmark models IDCNN-CRF and BiLSTM-CRF, the F1 value of MF-HDL has increased by 18.42% and 17.18%, respectively, which verifies the excellent performance of the method in the Chinese medical entity extraction task.

Key words: hybrid deep learning, medical entity extraction, BERT, attention, Highway

中图分类号: