T-LBERT with Domain Adaptation for Cross-Domain Sentiment Classification

Hongye Cao, Qianru Wei, Jiangbin Zheng

Research output: Contribution to journalArticlepeer-review

Abstract

Cross-domain sentiment classification transfers the knowledge from the source domain to the target domain lacking supervised information for sentiment classification. Existing cross-domain sentiment classification methods establish connections by extracting domain-invariant features manually. However, these methods have poor adaptability to bridge connections across different domains and ignore important sentiment information. Hence, we propose a Topic Lite Bidirectional Encoder Representations from Transformers (T-LBERT) model with domain adaption to improve the adaptability of cross-domain sentiment classification. It combines the learning content of the source domain and the topic information of the target domain to improve the domain adaptability of the model. Due to the unbalanced distribution of information in the combined data, we apply a two-layer attention adaptive mechanism for classification. A shallow attention layer is applied to weigh the important features of the combined data. Inspired by active learning, we propose a deep domain adaption layer, which actively adjusts model parameters to balance the difference and representativeness between domains. Experimental results on Amazon review datasets demonstrate that the T-LBERT model considerably outperforms other state-of-the-art methods. T-LBERT shows stable classification performance on multiple metrics.

Original languageEnglish
Pages (from-to)141-150
Number of pages10
JournalInternational Arab Journal of Information Technology
Volume20
Issue number1
DOIs
StatePublished - Jan 2023

Keywords

  • Cross-domain
  • attention
  • domain adaption
  • sentiment classification
  • topic model

Fingerprint

Dive into the research topics of 'T-LBERT with Domain Adaptation for Cross-Domain Sentiment Classification'. Together they form a unique fingerprint.

Cite this