Presentation + Paper
7 June 2024 Heart rate and oxygen level estimation from facial videos using a hybrid deep learning model
Author Affiliations +
Abstract
Vital signs can be inferred from facial videos for health monitoring remotely, while facial videos can be easily obtained through phone cameras, webcams, or surveillance systems. In this study, we propose a hybrid deep learning model to estimate heart rate (HR) and blood oxygen saturation level (SpO2) from facial videos. The hybrid model has a mixed network architecture consisting of convolutional neural network (CNN), convolutional long short-term memory (convLSTM), and video vision transformer (ViViT). Temporal resolution is emphasized in feature extraction since both HR and SpO2 are varying over time. A clip of video consists of a set of frame images within a time segment. CNN is performed with regard to each frame (e.g., time distributed), convLSTM and ViViT can be configured to process a sequence of frames. These high-resolution temporal features are combined to predict HR and SpO2, which are expected to capture these signal variations. Our vital video dataset is fairly large by including 891 subjects from difference races and ages. Facial detection and data normalization are performed in preprocessing. Our experiments show that the proposed hybrid model can predict HR and SpO2 accurately. In addition, those models can be extended to infer HR fluctuations, respiratory rates, and blood pressure variations from facial videos.
Conference Presentation
(2024) Published by SPIE. Downloading of the abstract is permitted for personal use only.
Yufeng Zheng "Heart rate and oxygen level estimation from facial videos using a hybrid deep learning model", Proc. SPIE 13033, Multimodal Image Exploitation and Learning 2024, 1303307 (7 June 2024); https://doi.org/10.1117/12.3013956
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Vital signs

Deep learning

Transformers

Back to Top