|Table of Contents|

Domain Expert Entity Extraction Method Based on Multi-FeatureBidirectional Gated Neural Network(PDF)

《南京师大学报(自然科学版)》[ISSN:1001-4616/CN:32-1239/N]

Issue:
2021年01期
Page:
128-135
Research Field:
·计算机科学与技术·
Publishing date:

Info

Title:
Domain Expert Entity Extraction Method Based on Multi-FeatureBidirectional Gated Neural Network
Author(s):
Zhang KewenLi XiangYan YunyangZhu QuanyinMa Jialin
Faculty of Computer and Software Engineering,Huaiyin Institute of Technology,Huai’an 223005,China
Keywords:
named entity recognitionnatural language processinginformation extractionmulti-featureboundary feature
PACS:
TP301.6
DOI:
10.3969/j.issn.1001-4616.2021.01.018
Abstract:
Named entity recognition is the basic task of natural language processing(NLP)and information extraction(IE). Traditional expert named entity recognition methods have problems,such as excessive reliance on artificial feature labeling and word segmentation effects,and the inability to recognize a large number of professional new words in the expert profile. This paper proposes a method based on multi-features bidirectional gated neural network structure combined with conditional random field model for the domain expert entity extraction. Firstly,train the entity extraction model by constructing a domain expert corpus. Secondly,use the Bert method to represent the word embedding,and perform feature analysis on the vocabulary structure elements of the professional field of the corpus and extract the boundary features. Thirdly,use the bidirectional gated neural network and attention mechanism to effectively obtain the long-distance dependence of specific words. Finally,combine the conditional random field model to achieve named entity recognition. The experimental comparison and analysis of five methods on the same data set show that the F1 value of the model is improved by more than 9.98% compared with BiLSTM-CRF and IDCNN-CRF.

References:

[1] 邹博伟,钱忠,陈站成,等. 面向自然语言文本的否定性与不确定性信息抽取[J]. 软件学报,2016,27(2):309-328.
[2]LI J,SUN A,HAN J,et al. A survey on deep learning for named entity recognition[J]. IEEE transactions on knowledge and data engineering,2020,32(3):1558-2191.
[3]GE H,CAVERLEE J,LU H. Taper:a contextual tensor-based approach for personalized expert recommendation[C]//Proceedings of the 10th ACM Conference on Recommender Systems. Boston,2016:261-268.
[4]LI X,WANG Z,GAO S,et al. An intelligent context-aware management framework for cold chain logistics distribution[J]. IEEE transactions on intelligent transportation systems,2019,20(12):4553-4566.
[5]BARTOLI A,DE LORENZO A,MEDVET E,et al. Active learning of regular expressions for entity extraction[J]. IEEE transactions on cybernetics,2017,48(3):1067-1080.
[6]ZHANG Y,YANG J. Chinese NER using lattice LSTM[C]//Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics. Melbourne,2018:1554-1564.
[7]汪诚愚,何晓丰,宫学庆,等. 面向上下位关系预测的词嵌入投影模型[J]. 计算机学报,2019,43(5):868-883.
[8]MORWAL S,JAHAN N,CHOPRA D. Named entity recognition using hidden Markov model(HMM)[J]. International journal on natural language computing,2012,1(4):15-23.
[9]MCCALLUM A,FREITAG D,PEREIRA F C N. Maximum entropy Markov models for information extraction and segmentation[C]//Proceedings of International Conference on Machine Learning. Stanford,2000:591-598.
[10]LAFFERTY J,MCCALLUM A,PEREIRA F C N. Conditional random fields:probabilistic models for segmenting and labeling sequence data[C]//Proceedings of the Eighteenth International Conference on Machine Learning. Williams College,MA,2001:282-289.
[11]DEVLIN J,CHANG M W,LEE K,et al. Bert:pre-training of deep bidirectional transformers for language understanding[J]. Computation and language,2018,23(2):3-19.
[12]COLLOBERT R,WESTON J,BOTTOU L,et al. Natural language processing(almost)from scratch[J]. Journal of machine learning research,2011,12(1):2493-2537.
[13]STRUBELL E,VERGA P,BELANGER D,et al. Fast and accurate entity recognition with iterated dilated convolutions[C]//Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing. Copenhagen,2017:2670-2680.
[14]HUANG Z,XU W,YU K. Bidirectional LSTM-CRF models for sequence tagging[J]. Computer science,2015(8):1508-1518.
[15]WANG S,LI Y,LIU N,et al. Noisy-data-disposing algorithm of data clean on the attribute level[J]. Computer engineering,2005(9):86-87.



[16]张华平,吴林芳,张芯铭,等. 领域知识图谱小样本构建与应用[J]. 人工智能,2020(1):113-124.

[17]唐明,朱磊,邹显春. 基于Word2Vec的一种文档向量表示[J]. 计算机科学,2016,43(6):214-217,269.
[18]MIKOLOV T,SUTSKEVER I,CHEN K,et al. Distributed representations of words and phrases and their compositionality[J]. Advances in neural information processing systems,2013,26:3111-3119.
[19]PENNINGTON J,SOCHER R,MANNING C D. Glove:global vectors for word representation[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing. Doha,2014:1532-1543.
[20]严云洋,瞿学新,朱全银,等. 基于离群点检测的分类结果置信度的度量方法[J]. 南京大学学报(自然科学版),2019,55(1):102-109.
[21]GOUTTE C,GAUSSIER E. A probabilistic interpretation of precision,recall and F-score,with implication for uation[C]//Proceedings of European Conference on Information Retri. Springer,Berlin,Heidelberg,2005:345-359.

Memo

Memo:
-
Last Update: 2021-03-15