Neural Discontinuous Constituency Parsing

Miloš Stanojević, Raquel G. Alhama


Abstract
One of the most pressing issues in discontinuous constituency transition-based parsing is that the relevant information for parsing decisions could be located in any part of the stack or the buffer. In this paper, we propose a solution to this problem by replacing the structured perceptron model with a recursive neural model that computes a global representation of the configuration, therefore allowing even the most remote parts of the configuration to influence the parsing decisions. We also provide a detailed analysis of how this representation should be built out of sub-representations of its core elements (words, trees and stack). Additionally, we investigate how different types of swap oracles influence the results. Our model is the first neural discontinuous constituency parser, and it outperforms all the previously published models on three out of four datasets while on the fourth it obtains second place by a tiny difference.
Anthology ID:
D17-1174
Volume:
Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing
Month:
September
Year:
2017
Address:
Copenhagen, Denmark
Editors:
Martha Palmer, Rebecca Hwa, Sebastian Riedel
Venue:
EMNLP
SIG:
SIGDAT
Publisher:
Association for Computational Linguistics
Note:
Pages:
1666–1676
Language:
URL:
https://aclanthology.org/D17-1174
DOI:
10.18653/v1/D17-1174
Bibkey:
Cite (ACL):
Miloš Stanojević and Raquel G. Alhama. 2017. Neural Discontinuous Constituency Parsing. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pages 1666–1676, Copenhagen, Denmark. Association for Computational Linguistics.
Cite (Informal):
Neural Discontinuous Constituency Parsing (Stanojević & Alhama, EMNLP 2017)
Copy Citation:
PDF:
https://aclanthology.org/D17-1174.pdf