TY - GEN
T1 - ThermRNet
T2 - 4th International Conference on Frontiers of Artificial Intelligence and Machine Learning, FAIML 2025
AU - Wang, Haipeng
AU - Liu, Qihang
N1 - Publisher Copyright:
© 2025 Copyright held by the owner/author(s).
PY - 2025/8/15
Y1 - 2025/8/15
N2 - Respiratory rate is a vital physiological indicator, yet traditional monitoring methods often require physical contact, limiting their usability in unobtrusive or remote settings. Infrared thermography enables non-contact respiratory monitoring but faces challenges in robustness and accuracy under complex conditions. To address these issues, we propose ThermRNet, a novel deep spatiotemporal learning model designed for reliable and precise respiratory monitoring using thermal video data. ThermRNet fuses 3D convolutional neural networks (3D-CNN) for localized spatial-temporal feature extraction with a Transformer-based spatiotemporal attention mechanism that captures long-range dependencies across frames. This architecture enables the model to both recognize instantaneous breathing states on a per-frame basis and accurately estimate the continuous respiratory rate over time. On a thermal video dataset collected from 15 participants, our model achieves 95.1% classification accuracy, a mean absolute error of 0.2166 bpm, and a Pearson correlation coefficient of 0.9958. These results demonstrate ThermRNet's strong potential for reliable, contactless respiratory monitoring in real-world applications such as telemedicine and intelligent health surveillance applications.
AB - Respiratory rate is a vital physiological indicator, yet traditional monitoring methods often require physical contact, limiting their usability in unobtrusive or remote settings. Infrared thermography enables non-contact respiratory monitoring but faces challenges in robustness and accuracy under complex conditions. To address these issues, we propose ThermRNet, a novel deep spatiotemporal learning model designed for reliable and precise respiratory monitoring using thermal video data. ThermRNet fuses 3D convolutional neural networks (3D-CNN) for localized spatial-temporal feature extraction with a Transformer-based spatiotemporal attention mechanism that captures long-range dependencies across frames. This architecture enables the model to both recognize instantaneous breathing states on a per-frame basis and accurately estimate the continuous respiratory rate over time. On a thermal video dataset collected from 15 participants, our model achieves 95.1% classification accuracy, a mean absolute error of 0.2166 bpm, and a Pearson correlation coefficient of 0.9958. These results demonstrate ThermRNet's strong potential for reliable, contactless respiratory monitoring in real-world applications such as telemedicine and intelligent health surveillance applications.
KW - Infrared thermography
KW - Respiratory rate estimation
KW - Spatiotemporal deep learning
UR - https://www.scopus.com/pages/publications/105016678779
U2 - 10.1145/3748382.3748387
DO - 10.1145/3748382.3748387
M3 - 会议稿件
AN - SCOPUS:105016678779
T3 - FAIML 2025 - Proceedings of the 2025 4th International Conference on Frontiers of Artificial Intelligence and Machine Learning
SP - 21
EP - 25
BT - FAIML 2025 - Proceedings of the 2025 4th International Conference on Frontiers of Artificial Intelligence and Machine Learning
PB - Association for Computing Machinery, Inc
Y2 - 25 April 2025 through 27 April 2025
ER -