Academic Journal

Leveraging Concept-Enhanced Pre-Training Model and Masked-Entity Language Model for Named Entity Disambiguation

التفاصيل البيبلوغرافية
العنوان: Leveraging Concept-Enhanced Pre-Training Model and Masked-Entity Language Model for Named Entity Disambiguation
المؤلفون: Zizheng Ji, Lin Dai, Jin Pang, Tingting Shen
المصدر: IEEE Access, Vol 8, Pp 100469-100484 (2020)
بيانات النشر: IEEE, 2020.
سنة النشر: 2020
المجموعة: LCC:Electrical engineering. Electronics. Nuclear engineering
مصطلحات موضوعية: Named entity disambiguation, pre-training, lexical knowledge, Electrical engineering. Electronics. Nuclear engineering, TK1-9971
الوصف: Named Entity Disambiguation (NED) refers to the task of resolving multiple named entity mentions in an input-text sequence to their correct references in a knowledge graph. We tackle NED problem by leveraging two novel objectives for pre-training framework, and propose a novel pre-training NED model. Especially, the proposed pre-training NED model consists of: (i) concept-enhanced pre-training, aiming at identifying valid lexical semantic relations with the concept semantic constraints derived from external resource Probase; and (ii) masked entity language model, aiming to train the contextualized embedding by predicting randomly masked entities based on words and non-masked entities in the given input-text. Therefore, the proposed pre-training NED model could merge the advantage of pre-training mechanism for generating contextualized embedding with the superiority of the lexical knowledge (e.g., concept knowledge emphasized here) for understanding language semantic. We conduct experiments on the CoNLL dataset and TAC dataset, and various datasets provided by GERBIL platform. The experimental results demonstrate that the proposed model achieves significantly higher performance than previous models.
نوع الوثيقة: article
وصف الملف: electronic resource
اللغة: English
تدمد: 2169-3536
Relation: https://ieeexplore.ieee.org/document/9091850/; https://doaj.org/toc/2169-3536
DOI: 10.1109/ACCESS.2020.2994247
URL الوصول: https://doaj.org/article/176fbfa4ba0845e99129abc4f7c98e97
رقم الانضمام: edsdoj.176fbfa4ba0845e99129abc4f7c98e97
قاعدة البيانات: Directory of Open Access Journals
ResultId 1
Header edsdoj
Directory of Open Access Journals
edsdoj.176fbfa4ba0845e99129abc4f7c98e97
930
3
Academic Journal
academicJournal
930.017944335938
PLink https://search.ebscohost.com/login.aspx?direct=true&site=eds-live&scope=site&db=edsdoj&AN=edsdoj.176fbfa4ba0845e99129abc4f7c98e97&custid=s6537998&authtype=sso
FullText Array ( [Availability] => 0 )
Array ( [0] => Array ( [Url] => https://doaj.org/article/176fbfa4ba0845e99129abc4f7c98e97 [Name] => EDS - DOAJ [Category] => fullText [Text] => View record in DOAJ [MouseOverText] => View record in DOAJ ) [1] => Array ( [Url] => https://resolver.ebscohost.com/openurl?custid=s6537998&groupid=main&authtype=ip,guest&sid=EBSCO:edsdoj&genre=article&issn=21693536&ISBN=&volume=8&issue=&date=20200101&spage=100469&pages=100469-100484&title=IEEE Access&atitle=Leveraging%20Concept-Enhanced%20Pre-Training%20Model%20and%20Masked-Entity%20Language%20Model%20for%20Named%20Entity%20Disambiguation&id=DOI:10.1109/ACCESS.2020.2994247 [Name] => Full Text Finder (s6537998api) [Category] => fullText [Text] => Full Text Finder [Icon] => https://imageserver.ebscohost.com/branding/images/FTF.gif [MouseOverText] => Full Text Finder ) )
Items Array ( [Name] => Title [Label] => Title [Group] => Ti [Data] => Leveraging Concept-Enhanced Pre-Training Model and Masked-Entity Language Model for Named Entity Disambiguation )
Array ( [Name] => Author [Label] => Authors [Group] => Au [Data] => <searchLink fieldCode="AR" term="%22Zizheng+Ji%22">Zizheng Ji</searchLink><br /><searchLink fieldCode="AR" term="%22Lin+Dai%22">Lin Dai</searchLink><br /><searchLink fieldCode="AR" term="%22Jin+Pang%22">Jin Pang</searchLink><br /><searchLink fieldCode="AR" term="%22Tingting+Shen%22">Tingting Shen</searchLink> )
Array ( [Name] => TitleSource [Label] => Source [Group] => Src [Data] => IEEE Access, Vol 8, Pp 100469-100484 (2020) )
Array ( [Name] => Publisher [Label] => Publisher Information [Group] => PubInfo [Data] => IEEE, 2020. )
Array ( [Name] => DatePubCY [Label] => Publication Year [Group] => Date [Data] => 2020 )
Array ( [Name] => Subset [Label] => Collection [Group] => HoldingsInfo [Data] => LCC:Electrical engineering. Electronics. Nuclear engineering )
Array ( [Name] => Subject [Label] => Subject Terms [Group] => Su [Data] => <searchLink fieldCode="DE" term="%22Named+entity+disambiguation%22">Named entity disambiguation</searchLink><br /><searchLink fieldCode="DE" term="%22pre-training%22">pre-training</searchLink><br /><searchLink fieldCode="DE" term="%22lexical+knowledge%22">lexical knowledge</searchLink><br /><searchLink fieldCode="DE" term="%22Electrical+engineering%2E+Electronics%2E+Nuclear+engineering%22">Electrical engineering. Electronics. Nuclear engineering</searchLink><br /><searchLink fieldCode="DE" term="%22TK1-9971%22">TK1-9971</searchLink> )
Array ( [Name] => Abstract [Label] => Description [Group] => Ab [Data] => Named Entity Disambiguation (NED) refers to the task of resolving multiple named entity mentions in an input-text sequence to their correct references in a knowledge graph. We tackle NED problem by leveraging two novel objectives for pre-training framework, and propose a novel pre-training NED model. Especially, the proposed pre-training NED model consists of: (i) concept-enhanced pre-training, aiming at identifying valid lexical semantic relations with the concept semantic constraints derived from external resource Probase; and (ii) masked entity language model, aiming to train the contextualized embedding by predicting randomly masked entities based on words and non-masked entities in the given input-text. Therefore, the proposed pre-training NED model could merge the advantage of pre-training mechanism for generating contextualized embedding with the superiority of the lexical knowledge (e.g., concept knowledge emphasized here) for understanding language semantic. We conduct experiments on the CoNLL dataset and TAC dataset, and various datasets provided by GERBIL platform. The experimental results demonstrate that the proposed model achieves significantly higher performance than previous models. )
Array ( [Name] => TypeDocument [Label] => Document Type [Group] => TypDoc [Data] => article )
Array ( [Name] => Format [Label] => File Description [Group] => SrcInfo [Data] => electronic resource )
Array ( [Name] => Language [Label] => Language [Group] => Lang [Data] => English )
Array ( [Name] => ISSN [Label] => ISSN [Group] => ISSN [Data] => 2169-3536 )
Array ( [Name] => NoteTitleSource [Label] => Relation [Group] => SrcInfo [Data] => https://ieeexplore.ieee.org/document/9091850/; https://doaj.org/toc/2169-3536 )
Array ( [Name] => DOI [Label] => DOI [Group] => ID [Data] => 10.1109/ACCESS.2020.2994247 )
Array ( [Name] => URL [Label] => Access URL [Group] => URL [Data] => <link linkTarget="URL" linkTerm="https://doaj.org/article/176fbfa4ba0845e99129abc4f7c98e97" linkWindow="_blank">https://doaj.org/article/176fbfa4ba0845e99129abc4f7c98e97</link> )
Array ( [Name] => AN [Label] => Accession Number [Group] => ID [Data] => edsdoj.176fbfa4ba0845e99129abc4f7c98e97 )
RecordInfo Array ( [BibEntity] => Array ( [Identifiers] => Array ( [0] => Array ( [Type] => doi [Value] => 10.1109/ACCESS.2020.2994247 ) ) [Languages] => Array ( [0] => Array ( [Text] => English ) ) [PhysicalDescription] => Array ( [Pagination] => Array ( [PageCount] => 16 [StartPage] => 100469 ) ) [Subjects] => Array ( [0] => Array ( [SubjectFull] => Named entity disambiguation [Type] => general ) [1] => Array ( [SubjectFull] => pre-training [Type] => general ) [2] => Array ( [SubjectFull] => lexical knowledge [Type] => general ) [3] => Array ( [SubjectFull] => Electrical engineering. Electronics. Nuclear engineering [Type] => general ) [4] => Array ( [SubjectFull] => TK1-9971 [Type] => general ) ) [Titles] => Array ( [0] => Array ( [TitleFull] => Leveraging Concept-Enhanced Pre-Training Model and Masked-Entity Language Model for Named Entity Disambiguation [Type] => main ) ) ) [BibRelationships] => Array ( [HasContributorRelationships] => Array ( [0] => Array ( [PersonEntity] => Array ( [Name] => Array ( [NameFull] => Zizheng Ji ) ) ) [1] => Array ( [PersonEntity] => Array ( [Name] => Array ( [NameFull] => Lin Dai ) ) ) [2] => Array ( [PersonEntity] => Array ( [Name] => Array ( [NameFull] => Jin Pang ) ) ) [3] => Array ( [PersonEntity] => Array ( [Name] => Array ( [NameFull] => Tingting Shen ) ) ) ) [IsPartOfRelationships] => Array ( [0] => Array ( [BibEntity] => Array ( [Dates] => Array ( [0] => Array ( [D] => 01 [M] => 01 [Type] => published [Y] => 2020 ) ) [Identifiers] => Array ( [0] => Array ( [Type] => issn-print [Value] => 21693536 ) ) [Numbering] => Array ( [0] => Array ( [Type] => volume [Value] => 8 ) ) [Titles] => Array ( [0] => Array ( [TitleFull] => IEEE Access [Type] => main ) ) ) ) ) ) )
IllustrationInfo