TY - JOUR
T1 - FMSA-SC
T2 - A Fine-Grained Multimodal Sentiment Analysis Dataset Based on Stock Comment Videos
AU - Song, Lingyun
AU - Chen, Siyu
AU - Meng, Ziyang
AU - Sun, Mingxuan
AU - Shang, Xuequn
N1 - Publisher Copyright:
© 1999-2012 IEEE.
PY - 2024
Y1 - 2024
N2 - Previous Sentiment Analysis (SA) studies have demonstrated that exploring sentiment cues from multiple synchronized modalities can effectively improve the SA results. Unfortunately, until now there is no publicly available dataset for multimodal SA of the stock market. Existing datasets for stock market SA only provide textual stock comments, which usually contain words with ambiguous sentiments or even sarcasm words expressing opposite sentiments of literal meaning. To address this issue, we introduce a Fine-grained Multimodal Sentiment Analysis dataset built upon 1,247 Stock Comment videos, called FMSA-SC. It provides both multimodal sentiment annotations for the videos and unimodal sentiment annotations for the textual, visual, and acoustic modalities of the videos. In addition, FMSA-SC also provides fine-grained annotations that align text at the phrase level with visual and acoustic modalities. Furthermore, we present a new fine-grained multimodal multi-task framework as the baseline for multimodal SA on the FMSA-SC.
AB - Previous Sentiment Analysis (SA) studies have demonstrated that exploring sentiment cues from multiple synchronized modalities can effectively improve the SA results. Unfortunately, until now there is no publicly available dataset for multimodal SA of the stock market. Existing datasets for stock market SA only provide textual stock comments, which usually contain words with ambiguous sentiments or even sarcasm words expressing opposite sentiments of literal meaning. To address this issue, we introduce a Fine-grained Multimodal Sentiment Analysis dataset built upon 1,247 Stock Comment videos, called FMSA-SC. It provides both multimodal sentiment annotations for the videos and unimodal sentiment annotations for the textual, visual, and acoustic modalities of the videos. In addition, FMSA-SC also provides fine-grained annotations that align text at the phrase level with visual and acoustic modalities. Furthermore, we present a new fine-grained multimodal multi-task framework as the baseline for multimodal SA on the FMSA-SC.
KW - Multimedia databases
KW - neural networks
KW - sentiment analysis
KW - video signal processing
UR - http://www.scopus.com/inward/record.url?scp=85187278942&partnerID=8YFLogxK
U2 - 10.1109/TMM.2024.3363641
DO - 10.1109/TMM.2024.3363641
M3 - 文章
AN - SCOPUS:85187278942
SN - 1520-9210
VL - 26
SP - 7294
EP - 7306
JO - IEEE Transactions on Multimedia
JF - IEEE Transactions on Multimedia
ER -