TY - JOUR
T1 - PneumoLLM
T2 - Harnessing the power of large language model for pneumoconiosis diagnosis
AU - Song, Meiyue
AU - Wang, Jiarui
AU - Yu, Zhihua
AU - Wang, Jiaxin
AU - Yang, Le
AU - Lu, Yuting
AU - Li, Baicun
AU - Wang, Xue
AU - Wang, Xiaoxu
AU - Huang, Qinghua
AU - Li, Zhijun
AU - Kanellakis, Nikolaos I.
AU - Liu, Jiangfeng
AU - Wang, Jing
AU - Wang, Binglu
AU - Yang, Juntao
N1 - Publisher Copyright:
© 2024 Elsevier B.V.
PY - 2024/10
Y1 - 2024/10
N2 - The conventional pretraining-and-finetuning paradigm, while effective for common diseases with ample data, faces challenges in diagnosing data-scarce occupational diseases like pneumoconiosis. Recently, large language models (LLMs) have exhibits unprecedented ability when conducting multiple tasks in dialogue, bringing opportunities to diagnosis. A common strategy might involve using adapter layers for vision–language alignment and diagnosis in a dialogic manner. Yet, this approach often requires optimization of extensive learnable parameters in the text branch and the dialogue head, potentially diminishing the LLMs’ efficacy, especially with limited training data. In our work, we innovate by eliminating the text branch and substituting the dialogue head with a classification head. This approach presents a more effective method for harnessing LLMs in diagnosis with fewer learnable parameters. Furthermore, to balance the retention of detailed image information with progression towards accurate diagnosis, we introduce the contextual multi-token engine. This engine is specialized in adaptively generating diagnostic tokens. Additionally, we propose the information emitter module, which unidirectionally emits information from image tokens to diagnosis tokens. Comprehensive experiments validate the superiority of our methods.
AB - The conventional pretraining-and-finetuning paradigm, while effective for common diseases with ample data, faces challenges in diagnosing data-scarce occupational diseases like pneumoconiosis. Recently, large language models (LLMs) have exhibits unprecedented ability when conducting multiple tasks in dialogue, bringing opportunities to diagnosis. A common strategy might involve using adapter layers for vision–language alignment and diagnosis in a dialogic manner. Yet, this approach often requires optimization of extensive learnable parameters in the text branch and the dialogue head, potentially diminishing the LLMs’ efficacy, especially with limited training data. In our work, we innovate by eliminating the text branch and substituting the dialogue head with a classification head. This approach presents a more effective method for harnessing LLMs in diagnosis with fewer learnable parameters. Furthermore, to balance the retention of detailed image information with progression towards accurate diagnosis, we introduce the contextual multi-token engine. This engine is specialized in adaptively generating diagnostic tokens. Additionally, we propose the information emitter module, which unidirectionally emits information from image tokens to diagnosis tokens. Comprehensive experiments validate the superiority of our methods.
KW - Foundational model
KW - Large language model
KW - Medical image diagnosis
UR - http://www.scopus.com/inward/record.url?scp=85196958593&partnerID=8YFLogxK
U2 - 10.1016/j.media.2024.103248
DO - 10.1016/j.media.2024.103248
M3 - 文章
C2 - 38941859
AN - SCOPUS:85196958593
SN - 1361-8415
VL - 97
JO - Medical Image Analysis
JF - Medical Image Analysis
M1 - 103248
ER -