Judicious Selection of Training Data in Assisting Language for Multilingual Neural NER

Rudra Murthy, Anoop Kunchukuttan, Pushpak Bhattacharyya


Abstract
Multilingual learning for Neural Named Entity Recognition (NNER) involves jointly training a neural network for multiple languages. Typically, the goal is improving the NER performance of one of the languages (the primary language) using the other assisting languages. We show that the divergence in the tag distributions of the common named entities between the primary and assisting languages can reduce the effectiveness of multilingual learning. To alleviate this problem, we propose a metric based on symmetric KL divergence to filter out the highly divergent training instances in the assisting language. We empirically show that our data selection strategy improves NER performance in many languages, including those with very limited training data.
Anthology ID:
P18-2064
Volume:
Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)
Month:
July
Year:
2018
Address:
Melbourne, Australia
Editors:
Iryna Gurevych, Yusuke Miyao
Venue:
ACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
401–406
Language:
URL:
https://aclanthology.org/P18-2064
DOI:
10.18653/v1/P18-2064
Bibkey:
Cite (ACL):
Rudra Murthy, Anoop Kunchukuttan, and Pushpak Bhattacharyya. 2018. Judicious Selection of Training Data in Assisting Language for Multilingual Neural NER. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 401–406, Melbourne, Australia. Association for Computational Linguistics.
Cite (Informal):
Judicious Selection of Training Data in Assisting Language for Multilingual Neural NER (Murthy et al., ACL 2018)
Copy Citation:
PDF:
https://aclanthology.org/P18-2064.pdf
Presentation:
 P18-2064.Presentation.pdf
Video:
 https://aclanthology.org/P18-2064.mp4
Code
 murthyrudra/NeuralNER