TY - JOUR
T1 - Learning Word Representations for Sentiment Analysis
AU - Li, Yang
AU - Pan, Quan
AU - Yang, Tao
AU - Wang, Suhang
AU - Tang, Jiliang
AU - Cambria, Erik
N1 - Publisher Copyright:
© 2017, Springer Science+Business Media, LLC.
PY - 2017/12/1
Y1 - 2017/12/1
N2 - Word embedding has been proven to be a useful model for various natural language processing tasks. Traditional word embedding methods merely take into account word distributions independently from any specific tasks. Hence, the resulting representations could be sub-optimal for a given task. In the context of sentiment analysis, there are various types of prior knowledge available, e.g., sentiment labels of documents from available datasets or polarity values of words from sentiment lexicons. We incorporate such prior sentiment information at both word level and document level in order to investigate the influence each word has on the sentiment label of both target word and context words. By evaluating the performance of sentiment analysis in each category, we find the best way of incorporating prior sentiment information. Experimental results on real-world datasets demonstrate that the word representations learnt by DLJT2 can significantly improve the sentiment analysis performance. We prove that incorporating prior sentiment knowledge into the embedding process has the potential to learn better representations for sentiment analysis.
AB - Word embedding has been proven to be a useful model for various natural language processing tasks. Traditional word embedding methods merely take into account word distributions independently from any specific tasks. Hence, the resulting representations could be sub-optimal for a given task. In the context of sentiment analysis, there are various types of prior knowledge available, e.g., sentiment labels of documents from available datasets or polarity values of words from sentiment lexicons. We incorporate such prior sentiment information at both word level and document level in order to investigate the influence each word has on the sentiment label of both target word and context words. By evaluating the performance of sentiment analysis in each category, we find the best way of incorporating prior sentiment information. Experimental results on real-world datasets demonstrate that the word representations learnt by DLJT2 can significantly improve the sentiment analysis performance. We prove that incorporating prior sentiment knowledge into the embedding process has the potential to learn better representations for sentiment analysis.
UR - http://www.scopus.com/inward/record.url?scp=85027529941&partnerID=8YFLogxK
U2 - 10.1007/s12559-017-9492-2
DO - 10.1007/s12559-017-9492-2
M3 - 文章
AN - SCOPUS:85027529941
SN - 1866-9956
VL - 9
SP - 843
EP - 851
JO - Cognitive Computation
JF - Cognitive Computation
IS - 6
ER -