YNU-HPCC at SemEval-2018 Task 12: The Argument Reasoning Comprehension Task Using a Bi-directional LSTM with Attention Model

Quanlei Liao, Xutao Yang, Jin Wang, Xuejie Zhang


Abstract
An argument is divided into two parts, the claim and the reason. To obtain a clearer conclusion, some additional explanation is required. In this task, the explanations are called warrants. This paper introduces a bi-directional long short term memory (Bi-LSTM) with an attention model to select a correct warrant from two to explain an argument. We address this question as a question-answering system. For each warrant, the model produces a probability that it is correct. Finally, the system chooses the highest correct probability as the answer. Ensemble learning is used to enhance the performance of the model. Among all of the participants, we ranked 15th on the test results.
Anthology ID:
S18-1187
Volume:
Proceedings of the 12th International Workshop on Semantic Evaluation
Month:
June
Year:
2018
Address:
New Orleans, Louisiana
Editors:
Marianna Apidianaki, Saif M. Mohammad, Jonathan May, Ekaterina Shutova, Steven Bethard, Marine Carpuat
Venue:
SemEval
SIG:
SIGLEX
Publisher:
Association for Computational Linguistics
Note:
Pages:
1109–1113
Language:
URL:
https://aclanthology.org/S18-1187
DOI:
10.18653/v1/S18-1187
Bibkey:
Cite (ACL):
Quanlei Liao, Xutao Yang, Jin Wang, and Xuejie Zhang. 2018. YNU-HPCC at SemEval-2018 Task 12: The Argument Reasoning Comprehension Task Using a Bi-directional LSTM with Attention Model. In Proceedings of the 12th International Workshop on Semantic Evaluation, pages 1109–1113, New Orleans, Louisiana. Association for Computational Linguistics.
Cite (Informal):
YNU-HPCC at SemEval-2018 Task 12: The Argument Reasoning Comprehension Task Using a Bi-directional LSTM with Attention Model (Liao et al., SemEval 2018)
Copy Citation:
PDF:
https://aclanthology.org/S18-1187.pdf