Overcoming Language Variation in Sentiment Analysis with Social Attention

Yi Yang, Jacob Eisenstein


Abstract
Variation in language is ubiquitous, particularly in newer forms of writing such as social media. Fortunately, variation is not random; it is often linked to social properties of the author. In this paper, we show how to exploit social networks to make sentiment analysis more robust to social language variation. The key idea is linguistic homophily: the tendency of socially linked individuals to use language in similar ways. We formalize this idea in a novel attention-based neural network architecture, in which attention is divided among several basis models, depending on the author’s position in the social network. This has the effect of smoothing the classification function across the social network, and makes it possible to induce personalized classifiers even for authors for whom there is no labeled data or demographic metadata. This model significantly improves the accuracies of sentiment analysis on Twitter and on review data.
Anthology ID:
Q17-1021
Volume:
Transactions of the Association for Computational Linguistics, Volume 5
Month:
Year:
2017
Address:
Cambridge, MA
Editors:
Lillian Lee, Mark Johnson, Kristina Toutanova
Venue:
TACL
SIG:
Publisher:
MIT Press
Note:
Pages:
295–307
Language:
URL:
https://aclanthology.org/Q17-1021
DOI:
10.1162/tacl_a_00062
Bibkey:
Cite (ACL):
Yi Yang and Jacob Eisenstein. 2017. Overcoming Language Variation in Sentiment Analysis with Social Attention. Transactions of the Association for Computational Linguistics, 5:295–307.
Cite (Informal):
Overcoming Language Variation in Sentiment Analysis with Social Attention (Yang & Eisenstein, TACL 2017)
Copy Citation:
PDF:
https://aclanthology.org/Q17-1021.pdf
Video:
 https://aclanthology.org/Q17-1021.mp4
Code
 yiyang-gt/social-attention
Data
Ciao