TY - JOUR
T1 - BioKnowPrompt
T2 - Incorporating imprecise knowledge into prompt-tuning verbalizer with biomedical text for relation extraction
AU - Li, Qing
AU - Wang, Yichen
AU - You, Tao
AU - Lu, Yantao
N1 - Publisher Copyright:
© 2022 Elsevier Inc.
PY - 2022/12
Y1 - 2022/12
N2 - Domain tuning pre-trained language models (PLMs) with task-specific prompts have achieved great success in different domains. By using cloze-style language prompts to stimulate the versatile knowledge of PLMs, which directly bridges the gap between pre-training tasks and various downstream tasks. Large unlabelled corpora in the biomedical domain have been created in the last decade(i.e., PubMed, PMC, MIMIC, and ScienceDirect). In this paper, we introduce BioKnowPrompt, a prompt-tuning PLMs model that has been incorporating imprecise knowledge into verbalizer for biomedical text relation extraction. In particular, we use learnable words and learnable relation words to infuse entity and relation information into quick creation, and we use biomedical domain knowledge constraints to synergistically improve their representation. By using additional prompts to fine-tune PLMs, we can further stimulate the rich knowledge distributed in PLMs to better serve downstream tasks such as relation extraction. BioKnowPrompt has a lot of significant potential in few-shot learning, which outperforms the previous models and achieves state-of-the-art on the 5 datasets.
AB - Domain tuning pre-trained language models (PLMs) with task-specific prompts have achieved great success in different domains. By using cloze-style language prompts to stimulate the versatile knowledge of PLMs, which directly bridges the gap between pre-training tasks and various downstream tasks. Large unlabelled corpora in the biomedical domain have been created in the last decade(i.e., PubMed, PMC, MIMIC, and ScienceDirect). In this paper, we introduce BioKnowPrompt, a prompt-tuning PLMs model that has been incorporating imprecise knowledge into verbalizer for biomedical text relation extraction. In particular, we use learnable words and learnable relation words to infuse entity and relation information into quick creation, and we use biomedical domain knowledge constraints to synergistically improve their representation. By using additional prompts to fine-tune PLMs, we can further stimulate the rich knowledge distributed in PLMs to better serve downstream tasks such as relation extraction. BioKnowPrompt has a lot of significant potential in few-shot learning, which outperforms the previous models and achieves state-of-the-art on the 5 datasets.
KW - Biomedical knowledge-aware
KW - Few-shot learning
KW - Pre-trained language models
KW - Prompt tuning
KW - Relation extraction
UR - http://www.scopus.com/inward/record.url?scp=85141252014&partnerID=8YFLogxK
U2 - 10.1016/j.ins.2022.10.063
DO - 10.1016/j.ins.2022.10.063
M3 - 文章
AN - SCOPUS:85141252014
SN - 0020-0255
VL - 617
SP - 346
EP - 358
JO - Information Sciences
JF - Information Sciences
ER -