Position-aware Joint Entity and Relation Extraction with Attention Mechanism

Position-aware Joint Entity and Relation Extraction with Attention Mechanism

Chenglong Zhang, Shuyong Gao, Haofen Wang, Wenqiang Zhang

Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence
Main Track. Pages 4496-4502. https://doi.org/10.24963/ijcai.2022/624

Named entity recognition and relation extraction are two important core subtasks of information extraction, which aim to identify named entities and extract relations between them. In recent years, span representation methods have received a lot of attention and are widely used to extract entities and corresponding relations from plain texts. Most recent works focus on how to obtain better span representations from pre-trained encoders, but ignore the negative impact of a large number of span candidates on slowing down the model performance. In our work, we propose a joint entity and relation extraction model with an attention mechanism and position-attentive markers. The attention score of each candidate span is calculated, and most of the candidate spans with low attention scores are pruned before being fed into the span classifier, thus achieving the goal of removing the most irrelevant spans. At the same time, in order to explore whether the position information can improve the performance of the model, we add position-attentive markers to the model. The experimental results show that our model is effective. With the same pre-trained encoder, our model achieves the new state-of-the-art on standard benchmarks (ACE05, CoNLL04 and SciERC), obtaining a 4.7%-17.8% absolute improvement in relation F1.
Keywords:
Natural Language Processing: Information Extraction
Natural Language Processing: Knowledge Extraction
Data Mining: Knowledge Graphs and Knowledge Base Completion
Natural Language Processing: Language Models
Natural Language Processing: Named Entities