Assessing Phrasal Representation and Composition in Transformers

Lang Yu, Allyson Ettinger


Abstract
Deep transformer models have pushed performance on NLP tasks to new limits, suggesting sophisticated treatment of complex linguistic inputs, such as phrases. However, we have limited understanding of how these models handle representation of phrases, and whether this reflects sophisticated composition of phrase meaning like that done by humans. In this paper, we present systematic analysis of phrasal representations in state-of-the-art pre-trained transformers. We use tests leveraging human judgments of phrase similarity and meaning shift, and compare results before and after control of word overlap, to tease apart lexical effects versus composition effects. We find that phrase representation in these models relies heavily on word content, with little evidence of nuanced composition. We also identify variations in phrase representation quality across models, layers, and representation types, and make corresponding recommendations for usage of representations from these models.
Anthology ID:
2020.emnlp-main.397
Volume:
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)
Month:
November
Year:
2020
Address:
Online
Editors:
Bonnie Webber, Trevor Cohn, Yulan He, Yang Liu
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
4896–4907
Language:
URL:
https://aclanthology.org/2020.emnlp-main.397
DOI:
10.18653/v1/2020.emnlp-main.397
Bibkey:
Cite (ACL):
Lang Yu and Allyson Ettinger. 2020. Assessing Phrasal Representation and Composition in Transformers. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 4896–4907, Online. Association for Computational Linguistics.
Cite (Informal):
Assessing Phrasal Representation and Composition in Transformers (Yu & Ettinger, EMNLP 2020)
Copy Citation:
PDF:
https://aclanthology.org/2020.emnlp-main.397.pdf
Video:
 https://slideslive.com/38939299
Code
 yulang/phrasal-composition-in-transformers
Data
BiRD