Structured Minimally Supervised Learning for Neural Relation Extraction

Fan Bai, Alan Ritter


Abstract
We present an approach to minimally supervised relation extraction that combines the benefits of learned representations and structured learning, and accurately predicts sentence-level relation mentions given only proposition-level supervision from a KB. By explicitly reasoning about missing data during learning, our approach enables large-scale training of 1D convolutional neural networks while mitigating the issue of label noise inherent in distant supervision. Our approach achieves state-of-the-art results on minimally supervised sentential relation extraction, outperforming a number of baselines, including a competitive approach that uses the attention layer of a purely neural model.
Anthology ID:
N19-1310
Volume:
Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)
Month:
June
Year:
2019
Address:
Minneapolis, Minnesota
Editors:
Jill Burstein, Christy Doran, Thamar Solorio
Venue:
NAACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
3057–3069
Language:
URL:
https://aclanthology.org/N19-1310
DOI:
10.18653/v1/N19-1310
Bibkey:
Cite (ACL):
Fan Bai and Alan Ritter. 2019. Structured Minimally Supervised Learning for Neural Relation Extraction. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 3057–3069, Minneapolis, Minnesota. Association for Computational Linguistics.
Cite (Informal):
Structured Minimally Supervised Learning for Neural Relation Extraction (Bai & Ritter, NAACL 2019)
Copy Citation:
PDF:
https://aclanthology.org/N19-1310.pdf
Video:
 https://vimeo.com/359696937
Code
 bflashcp3f/PCNN-NMAR