Full Text:   <7804>

Summary:  <431>

CLC number: TP391.1

On-line Access: 2024-08-27

Received: 2023-10-17

Revision Accepted: 2024-05-08

Crosschecked: 2022-03-03

Cited: 0

Clicked: 2521

Citations:  Bibtex RefMan EndNote GB/T7714

 ORCID:

Hongbin ZHANG

https://orcid.org/0000-0001-6568-5117

Weiwen ZHANG

https://orcid.org/0000-0002-5098-6459

-   Go to

Article info.
Open peer comments

Frontiers of Information Technology & Electronic Engineering  2022 Vol.23 No.11 P.1620-1630

http://doi.org/10.1631/FITEE.2100495


Improving entity linking with two adaptive features


Author(s):  Hongbin ZHANG, Quan CHEN, Weiwen ZHANG

Affiliation(s):  School of Computer Science and Technology, Guangdong University of Technology, Guangzhou 510006, China

Corresponding email(s):   zhangww@gdut.edu.cn

Key Words:  Entity linking, Local model, Global model, Adaptive features, Entity type


Hongbin ZHANG, Quan CHEN, Weiwen ZHANG. Improving entity linking with two adaptive features[J]. Frontiers of Information Technology & Electronic Engineering, 2022, 23(11): 1620-1630.

@article{title="Improving entity linking with two adaptive features",
author="Hongbin ZHANG, Quan CHEN, Weiwen ZHANG",
journal="Frontiers of Information Technology & Electronic Engineering",
volume="23",
number="11",
pages="1620-1630",
year="2022",
publisher="Zhejiang University Press & Springer",
doi="10.1631/FITEE.2100495"
}

%0 Journal Article
%T Improving entity linking with two adaptive features
%A Hongbin ZHANG
%A Quan CHEN
%A Weiwen ZHANG
%J Frontiers of Information Technology & Electronic Engineering
%V 23
%N 11
%P 1620-1630
%@ 2095-9184
%D 2022
%I Zhejiang University Press & Springer
%DOI 10.1631/FITEE.2100495

TY - JOUR
T1 - Improving entity linking with two adaptive features
A1 - Hongbin ZHANG
A1 - Quan CHEN
A1 - Weiwen ZHANG
J0 - Frontiers of Information Technology & Electronic Engineering
VL - 23
IS - 11
SP - 1620
EP - 1630
%@ 2095-9184
Y1 - 2022
PB - Zhejiang University Press & Springer
ER -
DOI - 10.1631/FITEE.2100495


Abstract: 
entity linking (EL) is a fundamental task in natural language processing. Based on neural networks, existing systems pay more attention to the construction of the global model, but ignore latent semantic information in the local model and the acquisition of effective entity type information. In this paper, we propose two adaptive features, in which the first adaptive feature enables the local and global models to capture latent information, and the second adaptive feature describes effective information for entity type embeddings. These adaptive features can work together naturally to handle some uncertain entity type information for EL. Experimental results demonstrate that our EL system achieves the best performance on the AIDA-B and MSNBC datasets, and the best average performance on out-domain datasets. These results indicate that the proposed adaptive features, which are based on their own diverse contexts, can capture information that is conducive for EL.

利用两个自适应特征改进实体链接

张鸿彬,陈权,张伟文
广东工业大学计算机学院,中国广州市,510006
摘要:实体链接是自然语言处理中的一项基本任务。现有的基于神经网络的系统更多地关注全局模型的构建,而忽略了局部模型中潜在的语义信息和有效实体类型信息的获取。本文提出两个自适应特征,其中第一个自适应特征使得局部和全局模型能够捕获潜在信息,第二个自适应特征能够描述实体类型嵌入的有效信息。这些自适应特征可以很自然地协同工作来处理一些不确定的实体类型信息。实验结果表明,我们的实体链接系统在AIDA-B和MSNBC数据集上取得了最佳的性能,并在域外数据集上达到了最佳的平均性能。这些结果表明,所提出的自适应特征能够基于其自身不同的上下文来捕获有利于实体链接的信息。

关键词:实体链接;局部模型;全局模型;自适应特征;实体类型

Darkslateblue:Affiliate; Royal Blue:Author; Turquoise:Article

Reference

[1]Bunescu R, Paşca M, 2006. Using encyclopedic knowledge for named entity disambiguation. Proc 11th Conf of the European Chapter of the Association for Computational Linguistics, p.9-16.

[2]Cao YX, Hou L, Li JZ, et al., 2018. Neural collective entity linking. Proc 27th Int Conf on Computational Linguistics, p.675-686.

[3]Chen S, Wang JP, Jiang F, et al., 2020. Improving entity linking by modeling latent entity type information. Proc 34th AAAI Conf on Artificial Intelligence, p.7529-7537.

[4]Deng ZH, Li ZX, Yang Q, et al., 2020. Improving entity linking with graph networks. Proc 21st Int Conf on Web Information Systems Engineering, p.343-354.

[5]Devlin J, Chang MW, Lee K, et al., 2019. BERT: pre-training of deep bidirectional transformers for language understanding. Proc Conf of the North American Chapter of the Association for Computational Linguistics, p.4171-4186.

[6]Durrett G, Klein D, 2014. A joint model for entity analysis: coreference, typing, and linking. Trans Assoc Comput Linguist, 2:477-490.

[7]Fang W, Zhang JW, Wang DL, et al., 2016. Entity disambiguation by knowledge and text jointly embedding. Proc 20th SIGNLL Conf on Computational Natural Language Learning, p.260-269.

[8]Fang Z, Cao YN, Li Q, et al., 2019. Joint entity linking with deep reinforcement learning. Proc World Wide Web Conf, p.438-447.

[9]Francis-Landau M, Durrett G, Klein D, 2016. Capturing semantic similarity for entity linking with convolutional neural networks. Proc Conf of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, p.1256-1261.

[10]Ganea OE, Hofmann T, 2017. Deep joint entity disambiguation with local neural attention. Proc Conf on Empirical Methods in Natural Language Processing, p.2619-2629.

[11]Globerson A, Lazic N, Chakrabarti S, et al., 2016. Collective entity resolution with multi-focal attention. Proc 54th Annual Meeting of the Association for Computational Linguistics, p.621-631.

[12]Guo ZC, Barbosa D, 2018. Robust named entity disambiguation with random walks. Semant Web, 9(4):459-479.

[13]Han XP, Sun L, 2011. A generative entity-mention model for linking entities with knowledge base. Proc 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, p.945-954.

[14]Hoffart J, Yosef MA, Bordino I, et al., 2011. Robust disambiguation of named entities in text. Proc Conf on Empirical Methods in Natural Language Processing, p.782-792.

[15]Hoffmann R, Zhang CL, Ling X, et al., 2011. Knowledge-based weak supervision for information extraction of overlapping relations. Proc 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, p.541-550.

[16]Honnibal M, Dale R, 2009. DAMSEL: the DSTO/Macquarie system for entity-linking. Proc 2nd Text Analysis Conf.

[17]Hou F, Wang RL, He J, et al., 2020. Improving entity linking through semantic reinforced entity embeddings. Proc 58th Annual Meeting of the Association for Computational Linguistics, p.6843-6848.

[18]Le P, Titov I, 2018. Improving entity linking by modeling latent relations between mentions. Proc 56th Annual Meeting of the Association for Computational Linguistics, p.1595-1604.

[19]Li ZZ, Feng DW, Li DS, et al., 2020. Learning to select pseudo labels: a semi-supervised method for named entity recognition. Front Inform Technol Electron Eng, 21(6):903-916.

[20]Luo G, Huang XJ, Lin CY, et al., 2015. Joint entity recognition and disambiguation. Proc Conf on Empirical Methods in Natural Language Processing, p.879-888.

[21]Martins PH, Marinho Z, Martins AFT, 2019. Joint learning of named entity recognition and entity linking. Proc 57th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop, p.190-196.

[22]Medelyan O, Milne D, Legg C, et al., 2009. Mining meaning from Wikipedia. Int J Hum-Comput Stud, 67(9):716-754.

[23]Milne D, Witten IH, 2008. Learning to link with Wikipedia. Proc 17th ACM Conf on Information and Knowledge Management, p.509-518.

[24]Mulang IO, Singh K, Prabhu C, et al., 2020. Evaluating the impact of knowledge graph context on entity disambiguation models. Proc 29th ACM Int Conf on Information & Knowledge Management, p.2157-2160.

[25]Nguyen DB, Theobald M, Weikum G, 2016. J-NERD: joint named entity recognition and disambiguation with rich linguistic features. Trans Assoc Comput Linguist, 4:215-229.

[26]Pennington J, Socher R, Manning C, 2014. GloVe: global vectors for word representation. Proc Conf on Empirical Methods in Natural Language Processing, p.1532-1543.

[27]Pershina M, He YF, Grishman R, 2015. Personalized page rank for named entity disambiguation. Proc Conf of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, p.238-243.

[28]Runge A, Hovy E, 2020. Exploring neural entity representations for semantic information. Proc 3rd BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP, p.204-216.

[29]Sevgili O, Shelmanov A, Arkhipov M, et al., 2020. Neural entity linking: a survey of models based on deep learning. https://arxiv.org/abs/2006.00575

[30]Shen W, Wang JY, Han JW, 2015. Entity linking with a knowledge base: issues, techniques, and solutions. IEEE Trans Knowl Data Eng, 27(2):443-460.

[31]Vaswani A, Shazeer N, Parmar N, et al., 2017. Attention is all you need. Proc 31st Int Conf on Neural Information Processing Systems, p.6000-6010.

[32]Wang HF, Liu ZQ, 2019. An error recognition method for power equipment defect records based on knowledge graph technology. Front Inform Technol Electron Eng, 20(11):1564-1577.

[33]Wang Z, Zhang JW, Feng JL, et al., 2014. Knowledge graph and text jointly embedding. Proc Conf on Empirical Methods in Natural Language Processing, p.1591-1601.

[34]Wu JS, Zhang RC, Mao YY, et al., 2020. Dynamic graph convolutional networks for entity linking. Proc Web Conf, p.1149-1159.

[35]Xin KX, Hua W, Liu Y, et al., 2019. Entity disambiguation based on parse tree neighbours on graph attention network. Proc 20th Int Conf on Web Information Systems Engineering, p.523-537.

[36]Xu P, Barbosa D, 2018. Neural fine-grained entity type classification with hierarchy-aware loss. Proc Conf of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, p.16-25.

[37]Xue MG, Cai WM, Su JS, et al., 2019. Neural collective entity linking based on recurrent random walk network learning. Proc 28th Int Joint Conf on Artificial Intelligence, p.5327-5333.

[38]Yamada I, Shindo H, Takeda H, et al., 2016. Joint learning of the embedding of words and entities for named entity disambiguation. Proc 20th SIGNLL Conf on Computational Natural Language Learning, p.250-259.

[39]Yang XY, Gu XT, Lin S, et al., 2019. Learning dynamic context augmentation for global entity linking. Proc Conf on Empirical Methods in Natural Language Processing and the 9th Int Joint Conf on Natural Language Processing, p.271-281.

[40]Zhang HB, Chen Q, Zhang WW, et al., 2022. HSIE: improving named entity disambiguation with hidden semantic information extractor. Proc 14th Int Conf on Machine Learning and Computing, p.251-257.

[41]Zhong HP, Zhang JW, Wang Z, et al., 2015. Aligning knowledge and text embeddings by entity descriptions. Proc Conf on Empirical Methods in Natural Language Processing, p.267-272.

Open peer comments: Debate/Discuss/Question/Opinion

<1>

Please provide your name, email address and a comment





Journal of Zhejiang University-SCIENCE, 38 Zheda Road, Hangzhou 310027, China
Tel: +86-571-87952783; E-mail: cjzhang@zju.edu.cn
Copyright © 2000 - 2024 Journal of Zhejiang University-SCIENCE