TY - GEN
T1 - An attention-based user profiling model by leveraging multi-modal social media contents
AU - Li, Zhimin
AU - Guo, Bin
AU - Sun, Yueqi
AU - Wang, Zhu
AU - Wang, Liang
AU - Yu, Zhiwen
N1 - Publisher Copyright:
© Springer Nature Singapore Pte Ltd. 2019.
PY - 2019
Y1 - 2019
N2 - With the popularization of social media, inferring user profiles from the user-generated content has aroused wide attention for its applications in marketing, advertising, recruiting, etc. Most existing works focus on using data from single modality (such as texts and profile photos) and fail to notice that the combination of multi-modal data can supplement with each other and can therefore improve the prediction accuracy. In this paper, we propose AMUP model, namely the Attention-based Multi-modal User Profiling model, which uses different tailored neural networks to extract and fuse semantic information from three modalities, i.e., texts, avatar, and relation network. We propose a dual attention mechanism. The word-level attention network selects informative words from the noisy and prolix texts and the modality-level attention network addresses the problem of imbalanced contribution among different modalities. Experimental results on more than 1.5K users’ real-world data extracted from a popular Q&A social platform show that our proposed model outperforms the single-modality methods and achieves better accuracy when compared with existing approaches that utilize multi-modal data.
AB - With the popularization of social media, inferring user profiles from the user-generated content has aroused wide attention for its applications in marketing, advertising, recruiting, etc. Most existing works focus on using data from single modality (such as texts and profile photos) and fail to notice that the combination of multi-modal data can supplement with each other and can therefore improve the prediction accuracy. In this paper, we propose AMUP model, namely the Attention-based Multi-modal User Profiling model, which uses different tailored neural networks to extract and fuse semantic information from three modalities, i.e., texts, avatar, and relation network. We propose a dual attention mechanism. The word-level attention network selects informative words from the noisy and prolix texts and the modality-level attention network addresses the problem of imbalanced contribution among different modalities. Experimental results on more than 1.5K users’ real-world data extracted from a popular Q&A social platform show that our proposed model outperforms the single-modality methods and achieves better accuracy when compared with existing approaches that utilize multi-modal data.
KW - Deep neural networks
KW - Dual attention mechanism
KW - Multi-modal learning
KW - User profile inferring
UR - http://www.scopus.com/inward/record.url?scp=85076897312&partnerID=8YFLogxK
U2 - 10.1007/978-981-15-1925-3_20
DO - 10.1007/978-981-15-1925-3_20
M3 - 会议稿件
AN - SCOPUS:85076897312
SN - 9789811519246
T3 - Communications in Computer and Information Science
SP - 272
EP - 284
BT - Cyberspace Data and Intelligence, and Cyber-Living, Syndrome, and Health - International 2019 Cyberspace Congress, CyberDI and CyberLife, Proceedings
A2 - Ning, Huansheng
PB - Springer
T2 - 3rd International Conference on Cyberspace Data and Intelligence, Cyber DI 2019, and the International Conference on Cyber-Living, Cyber-Syndrome, and Cyber-Health, CyberLife 2019
Y2 - 16 December 2019 through 18 December 2019
ER -