Learning Sentiment-Specific Word Embedding for Twitter Sentiment Classification

Learning Sentiment-Specific Word Embedding for Twitter Sentiment Classification

June 23-25, 2014 | Duyu Tang, Furu Wei, Nan Yang, Ming Zhou, Ting Liu, Bing Qin
This paper introduces a method for learning sentiment-specific word embeddings (SSWE) for Twitter sentiment classification. Existing word embedding methods typically model syntactic context but ignore sentiment, leading to poor performance in sentiment analysis as words with opposite sentiment polarities (e.g., good and bad) are mapped to similar vectors. To address this, the authors propose SSWE, which encodes sentiment information into word embeddings. They develop three neural networks to incorporate sentiment polarity into loss functions, enabling the model to learn embeddings that distinguish words with opposite sentiment polarities. The SSWE is learned from massive distant-supervised tweets using positive and negative emoticons. Experiments on the SemEval 2013 Twitter sentiment classification benchmark show that SSWE performs comparably to hand-crafted features and improves performance when combined with existing features. The SSWE feature is also evaluated by measuring word similarity in the embedding space for sentiment lexicons, where it outperforms existing word embedding methods. The authors propose three models: SSWE_h, SSWE_r, and SSWE_u. SSWE_h uses a softmax layer to predict sentiment polarity, while SSWE_r uses a ranking objective function to relax constraints. SSWE_u combines both syntactic and sentiment losses to capture both aspects. The models are trained on 10 million tweets, with SSWE_u achieving the best performance. The results show that SSWE outperforms traditional methods in sentiment classification, particularly in capturing sentiment information and distinguishing words with opposite sentiment polarities. The SSWE feature is also effective in improving the performance of existing sentiment analysis systems when combined with other features. The study demonstrates the effectiveness of learning sentiment-specific word embeddings for Twitter sentiment classification.This paper introduces a method for learning sentiment-specific word embeddings (SSWE) for Twitter sentiment classification. Existing word embedding methods typically model syntactic context but ignore sentiment, leading to poor performance in sentiment analysis as words with opposite sentiment polarities (e.g., good and bad) are mapped to similar vectors. To address this, the authors propose SSWE, which encodes sentiment information into word embeddings. They develop three neural networks to incorporate sentiment polarity into loss functions, enabling the model to learn embeddings that distinguish words with opposite sentiment polarities. The SSWE is learned from massive distant-supervised tweets using positive and negative emoticons. Experiments on the SemEval 2013 Twitter sentiment classification benchmark show that SSWE performs comparably to hand-crafted features and improves performance when combined with existing features. The SSWE feature is also evaluated by measuring word similarity in the embedding space for sentiment lexicons, where it outperforms existing word embedding methods. The authors propose three models: SSWE_h, SSWE_r, and SSWE_u. SSWE_h uses a softmax layer to predict sentiment polarity, while SSWE_r uses a ranking objective function to relax constraints. SSWE_u combines both syntactic and sentiment losses to capture both aspects. The models are trained on 10 million tweets, with SSWE_u achieving the best performance. The results show that SSWE outperforms traditional methods in sentiment classification, particularly in capturing sentiment information and distinguishing words with opposite sentiment polarities. The SSWE feature is also effective in improving the performance of existing sentiment analysis systems when combined with other features. The study demonstrates the effectiveness of learning sentiment-specific word embeddings for Twitter sentiment classification.
Reach us at info@study.space
[slides] Learning Sentiment-Specific Word Embedding for Twitter Sentiment Classification | StudySpace