InferNER: an attentive model leveraging the sentence-level information for Named Entity Recognition in Microblogs

We investigate the problem of named entity recognition in the user-generated text such as social media posts. This task is rendered particularly difficult by the restricted length and limited grammatical coherence of this data type. Current state-of-the-art approaches rely on external sources such a...

Full description

Saved in:
Bibliographic Details
Main Authors: Moemmur Shahzad, Ayesha Amin, Diego Esteves, Axel-Cyrille Ngonga Ngomo
Format: Article
Language:English
Published: LibraryPress@UF 2021-04-01
Series:Proceedings of the International Florida Artificial Intelligence Research Society Conference
Subjects:
Online Access:https://journals.flvc.org/FLAIRS/article/view/128538
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1849736503489462272
author Moemmur Shahzad
Ayesha Amin
Diego Esteves
Axel-Cyrille Ngonga Ngomo
author_facet Moemmur Shahzad
Ayesha Amin
Diego Esteves
Axel-Cyrille Ngonga Ngomo
author_sort Moemmur Shahzad
collection DOAJ
description We investigate the problem of named entity recognition in the user-generated text such as social media posts. This task is rendered particularly difficult by the restricted length and limited grammatical coherence of this data type. Current state-of-the-art approaches rely on external sources such as gazetteers to alleviate some of these restrictions. We present a neural model able to outperform state of the art on this task without recurring to gazetteers or similar external sources of information. Our approach relies on word-, character-, and sentence-level information for NER in short-text. Social media posts like tweets often have associated images that may provide auxiliary context relevant to understand these texts. Hence, we also incorporate visual information and introduce an attention component which computes attention weight probabilities over textual and text-relevant visual contexts separately. Our model outperforms the current state of the art on various NER datasets. On WNUT 2016 and 2017, our model achieved 53.48% and 50.52% F1 score, respectively. With Multimodal model, our system also outperforms the current SOTA with an F1 score of 74% on the multimodal dataset. Our evaluation further suggests that our model also goes beyond the current state-of-the-art on newswire data, hence corroborating its suitability for various NER tasks.
format Article
id doaj-art-e860535fb1be4384a362868ea15b90d5
institution DOAJ
issn 2334-0754
2334-0762
language English
publishDate 2021-04-01
publisher LibraryPress@UF
record_format Article
series Proceedings of the International Florida Artificial Intelligence Research Society Conference
spelling doaj-art-e860535fb1be4384a362868ea15b90d52025-08-20T03:07:16ZengLibraryPress@UFProceedings of the International Florida Artificial Intelligence Research Society Conference2334-07542334-07622021-04-013410.32473/flairs.v34i1.12853862928InferNER: an attentive model leveraging the sentence-level information for Named Entity Recognition in MicroblogsMoemmur Shahzad0Ayesha Amin1Diego Esteves2Axel-Cyrille Ngonga Ngomo3Universität PaderbornUniversität des SaarlandesFarfetchUniversität PaderbornWe investigate the problem of named entity recognition in the user-generated text such as social media posts. This task is rendered particularly difficult by the restricted length and limited grammatical coherence of this data type. Current state-of-the-art approaches rely on external sources such as gazetteers to alleviate some of these restrictions. We present a neural model able to outperform state of the art on this task without recurring to gazetteers or similar external sources of information. Our approach relies on word-, character-, and sentence-level information for NER in short-text. Social media posts like tweets often have associated images that may provide auxiliary context relevant to understand these texts. Hence, we also incorporate visual information and introduce an attention component which computes attention weight probabilities over textual and text-relevant visual contexts separately. Our model outperforms the current state of the art on various NER datasets. On WNUT 2016 and 2017, our model achieved 53.48% and 50.52% F1 score, respectively. With Multimodal model, our system also outperforms the current SOTA with an F1 score of 74% on the multimodal dataset. Our evaluation further suggests that our model also goes beyond the current state-of-the-art on newswire data, hence corroborating its suitability for various NER tasks.https://journals.flvc.org/FLAIRS/article/view/128538nlpnamed entity recognitionattention modelsmicroblogsshort-textneural networks
spellingShingle Moemmur Shahzad
Ayesha Amin
Diego Esteves
Axel-Cyrille Ngonga Ngomo
InferNER: an attentive model leveraging the sentence-level information for Named Entity Recognition in Microblogs
Proceedings of the International Florida Artificial Intelligence Research Society Conference
nlp
named entity recognition
attention models
microblogs
short-text
neural networks
title InferNER: an attentive model leveraging the sentence-level information for Named Entity Recognition in Microblogs
title_full InferNER: an attentive model leveraging the sentence-level information for Named Entity Recognition in Microblogs
title_fullStr InferNER: an attentive model leveraging the sentence-level information for Named Entity Recognition in Microblogs
title_full_unstemmed InferNER: an attentive model leveraging the sentence-level information for Named Entity Recognition in Microblogs
title_short InferNER: an attentive model leveraging the sentence-level information for Named Entity Recognition in Microblogs
title_sort inferner an attentive model leveraging the sentence level information for named entity recognition in microblogs
topic nlp
named entity recognition
attention models
microblogs
short-text
neural networks
url https://journals.flvc.org/FLAIRS/article/view/128538
work_keys_str_mv AT moemmurshahzad inferneranattentivemodelleveragingthesentencelevelinformationfornamedentityrecognitioninmicroblogs
AT ayeshaamin inferneranattentivemodelleveragingthesentencelevelinformationfornamedentityrecognitioninmicroblogs
AT diegoesteves inferneranattentivemodelleveragingthesentencelevelinformationfornamedentityrecognitioninmicroblogs
AT axelcyrillengongangomo inferneranattentivemodelleveragingthesentencelevelinformationfornamedentityrecognitioninmicroblogs