Learning Sentiment-Specific Word Embedding for Twitter Sentiment Classification

NLPIR SEMINAR Y2019#40

INTRO

In the new semester, our Lab, Web Search Mining and Security Lab, plans to hold an academic seminar every Monday, and each time a keynote speaker will share understanding of papers on his/her related research with you.

Arrangement

Tomorrow’s seminar is organized as follows:

  1. The seminar time is 1:20.pm, Mon (December 9, 2019), at Zhongguancun Technology Park ,Building 5, 1306.
  2. Baohua Zhang is going to give a presentation on the paper, Learning Sentiment-Specific Word Embedding for Twitter Sentiment Classification.(Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, pages 1555-1565, Baltimore, Maryland, USA, June 23-25 2014)
  3. The seminar will be hosted by Ziyu Liu.

Everyone interested in this topic is welcomed to join us.

Learning Sentiment-Specific Word Embedding for Twitter Sentiment Classification

Duyu Tang, Furu Wei, Nan Yang, Ming Zhou, Ting Liu, Bing Qin

Abstract

We present a method that learns word emedding for Twitter sentiment classification in this paper. Most existing algorithms for learning continuous word representations typically only model the syntactic context of words but ignore the sentiment of text. This is problematic for sentiment analysis as they usually map words with similar syntactic context but opposite sentiment polarity, such as good and bad, to neighboring word vectors. We address this issue by learning sentiment-specific word embedding (SSWE), which encodes sentiment information in the continuous representation of words. Specifically, we develop three neural networks to effectively incorporate the supervision from sentiment polarity of text (e.g. sentences or tweets) in their loss functions. To obtain large scale training corpora, we learn the sentiment-specific word embedding from massive distant-supervised tweets collected by positive and negative emoticons. Experiments on applying SSWE to a benchmark Twitter sentiment classification dataset in SemEval 2013 show that (1) the SSWE feature performs comparably with hand-crafted features in the top-performed system; (2) the performance is further improved by concatenating SSWE with existing feature set.

You May Also Like

About the Author: nlpvv

发表回复