Login (DCU Staff Only)
Login (DCU Staff Only)

DORAS | DCU Research Repository

Explore open access research and scholarly works from DCU

Advanced Search

Multi-level structured self-attentions for distantly supervised relation extraction

Du, Jinhua orcid logoORCID: 0000-0002-3267-4881, Han, Jingguang, Way, Andy orcid logoORCID: 0000-0001-5736-5930 and Wan, Dadong (2018) Multi-level structured self-attentions for distantly supervised relation extraction. In: 2018 Conference on Empirical Methods in Natural Language Processing, 31 Oct - 4 Nov 2018, Brussels, Belgium.

Abstract
Attention mechanisms are often used in deep neural networks for distantly supervised relation extraction (DS-RE) to distinguish valid from noisy instances. However, traditional 1- D vector attention models are insufficient for the learning of different contexts in the selection of valid instances to predict the relationship for an entity pair. To alleviate this issue, we propose a novel multi-level structured (2-D matrix) self-attention mechanism for DS-RE in a multi-instance learning (MIL) framework using bidirectional recurrent neural networks. In the proposed method, a structured word-level self-attention mechanism learns a 2-D matrix where each row vector represents a weight distribution for different aspects of an instance regarding two entities. Targeting the MIL issue, the structured sentence-level attention learns a 2-D matrix where each row vector represents a weight distribution on selection of different valid instances. Experiments conducted on two publicly available DS-RE datasets show that the proposed framework with a multi-level structured self-attention mechanism significantly outperform state-of-the-art baselines in terms of PR curves, P@N and F1 measures.
Metadata
Item Type:Conference or Workshop Item (Paper)
Event Type:Conference
Refereed:Yes
Subjects:Computer Science > Machine translating
DCU Faculties and Centres:DCU Faculties and Schools > Faculty of Engineering and Computing > School of Computing
Research Institutes and Centres > ADAPT
Published in: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. . Association for Computational Linguistics (ACL).
Publisher:Association for Computational Linguistics (ACL)
Official URL:https://www.aclweb.org/anthology/D18-1245
Copyright Information:© 2018 Association for Computational Linguistics (ACL)
Funders:ADAPT Centre for Digital Content Technology, funded under the SFI Research Centres Programme (Grant 13/RC/2106), and by SFI Industry Fellowship Programme 2016 (Grant 16/IFB/4490), and is supported by Accenture Labs Dublin.
ID Code:23354
Deposited On:24 May 2019 15:11 by Thomas Murtagh . Last Modified 24 May 2019 15:11
Documents

Full text available as:

[thumbnail of Multi-Level_Structured_Self-Attentions_for_Distantly_Supervised_Relation_Extraction[1].pdf]
Preview
PDF - Requires a PDF viewer such as GSview, Xpdf or Adobe Acrobat Reader
657kB
Downloads

Downloads

Downloads per month over past year

Archive Staff Only: edit this record