The source code and instructions for obtaining a copy of the test data are made available with this paper. We compare its performance with that of recent state-of-the-art pulse estimation methods, and show that both RemotePulseNet and the low frame rate 3DCNNs produce high-quality pulse signals from faces captured under a challenging interview scenario. Using this insight, we propose RemotePulseNet, a novel 3DCNN architecture that exploits temporally dilated convolutions with increasing dilation rate to drastically increase the receptive field.
Realistic fps prefab 1.23 cgpeers full#
In particular, we found that models trained on videos with frame rates as low as 12.9 frames per second (fps) perform better than those trained on videos recorded at a full 90 fps, perhaps due to the temporal receptive fields becoming larger in time dimension when the fps decreases. We show that previous studies reporting null effects of frame rate changes on pulse estimators may no longer be valid when using CNNs, and decreasing the frame rate may actually improve performance. We train and evaluate 32 3DCNNs with different temporal parameters on a new large-scale database for physiological monitoring in an interview scenario. In this paper, we explore the effect of frame rate, temporal kernel width, and – more generally – temporal receptive field on the reliability of heart rate and waveform estimation carried out by 3DCNNs. As the frame rate decreases substantially, from 8 skipped frames (11.25 fps) to 20 skipped frames (4.5 fps), the performance on MAE, RMSE, and correlation degrades. 3-Dimensional Convolutional Neural Networks (3DCNNs) have shown promising performance on the rPPG task, although it is critical that we understand the impact of both video and model parameters. Skipping 7 frames is equivalent to a 12.86 fps video, which is more likely feasible for mobile devices than the native recording rate found in our dataset. Remote photoplethysmography (rPPG) is the monitoring of blood volume pulse from a camera at a distance.