Options
Exploring fine-tuned embeddings that model intensifiers for emotion analysis
Bostan, Laura; Klinger, Roman (2019): Exploring fine-tuned embeddings that model intensifiers for emotion analysis, in: Alexandra Balahur, Roman Klinger, Veronique Hoste, u. a. (Hrsg.), Proceedings of the Tenth Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, Minneapolis: Association for Computational Linguistics, S. 25–34, doi: 10.18653/v1/W19-1304.
Faculty/Chair:
Author:
Title of the compilation:
Proceedings of the Tenth Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis
Editors:
Conference:
Tenth Workshop on Computational Approaches to Subjectivity and Sentiment Analysis (WASSA), Juni 2019 ; Minneapolis
Publisher Information:
Year of publication:
2019
Pages:
Language:
English
DOI:
Abstract:
Adjective phrases like “a little bit surprised”, “completely shocked”, or “not stunned at all” are not handled properly by current state-of-the-art emotion classification and intensity prediction systems. Based on this finding, we analyze differences between embeddings used by these systems in regard to their capability of handling such cases and argue that intensifiers in context of emotion words need special treatment, as is established for sentiment polarity classification, but not for more fine-grained emotion prediction. To resolve this issue, we analyze different aspects of a post-processing pipeline which enriches the word representations of such phrases. This includes expansion of semantic spaces at the phrase level and sub-word level followed by retrofitting to emotion lexicons. We evaluate the impact of these steps with ‘A La Carte and Bag-of-Substrings extensions based on pretrained GloVe,Word2vec, and fastText embeddings against a crowd-sourced corpus of intensity annotations for tweets containing our focus phrases. We show that the fastText-based models do not gain from handling these specific phrases under inspection. For Word2vec embeddings, we show that our post-processing pipeline improves the results by up to 8% on a novel dataset densly populated with intensifiers while it does not decrease the performance on the established EmoInt dataset.
GND Keywords: ;
Computerlinguistik
Gefühl
Keywords:
fine-tuned embeddings
DDC Classification:
RVK Classification:
Peer Reviewed:
Yes:
International Distribution:
Yes:
Open Access Journal:
Yes:
Type:
Conferenceobject
Activation date:
March 7, 2024
Versioning
Question on publication
Permalink
https://fis.uni-bamberg.de/handle/uniba/93954