Skip to content

FactorizePhys: Matrix Factorization for Multidimensional Attention in Remote Physiological Sensing [NeurIPS 2024]

License

Notifications You must be signed in to change notification settings

PhysiologicAILab/FactorizePhys

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

FactorizePhys: Matrix Factorization for Multidimensional Attention in Remote Physiological Sensing

Jitesh Joshi1, Sos S. Agaian2 and Youngjun Cho1

1 Department of Computer Science, University College London, UK

2 Department of Computer Science, College of Staten Island, City University of New York, USA


Introduction

This work introduces the Factorized Self-Attention Module (FSAM), which uses nonnegative matrix factorization to compute multidimensional attention across spatial, temporal, and channel dimensions collectively, rather than separately. It proposes FactorizePhys, an end-to-end 3D-CNN model leverages FSAM to estimate blood volume pulse signals from video frames. The method enhances signal extraction tasks by effectively factorizing voxel embeddings for multidimensional attention and demonstrates adaptability by integrating into EfficientPhys, an established 2D-CNN-based architecture. FSAM's effectiveness and cross-dataset generalization are validated through experiments on four datasets, surpassing state-of-the-art rPPG techniques. The study highlights FSAM's potential as a general multidimensional attention mechanism and provides insights into the role of nonnegative matrix factorization in rPPG, offering a new approach for estimating accurate physiological signals from spatial-temporal data.

This work is accepted at NeurIPS, 2024. Please find our paper at this link.


๐Ÿ”ฅ Please remember to โญ this repo if you find it useful and cite our work if you end up using it in your work! ๐Ÿ”ฅ

๐Ÿ”ฅ If you have any questions or concerns, please create an issue ๐Ÿ“! ๐Ÿ”ฅ


Proposed Factorized Self-Attention Module (FSAM)

Factorized Self-Attention Module (FSAM)

Please refer to our paper for notations and symbols used in this figure.

FactorizePhys and EfficientPhys with FSAM

(A) Proposed FactorizePhys with FSAM; (B) FSAM Adapted for EfficientPhys

Cross-Dataset Generalization and Learned Spatial-Temporal Features

Cumulative Cross-Dataset Performance

(A) Cumulative cross-dataset performance (MAE) v/s latency* plot; (B) Visualization of learned spatial-temporal features for base 3D-CNN model trained with and without FSAM. * System specs: Ubuntu 22.04 OS, NVIDIA GeForce RTX 3070 Laptop GPU, Intelยฎ Coreโ„ข i7-10870H CPU @ 2.20GHz, 16 GB RAM

About the Repository

๐Ÿ““ Algorithms

The repo currently supports the following algorithms:

๐Ÿ“ Datasets

The repo supports four datasets, namely iBVP, PURE, SCAMPS, and UBFC-rPPG. To use these datasets in a deep learning model, you should organize the files as follows.

  • iBVP
    • Joshi, J.; Cho, Y. iBVP Dataset: RGB-Thermal rPPG Dataset with High Resolution Signal Quality Labels. Electronics 2024, 13, 1334.

      iBVP_Dataset/
      |   |-- p01_a/
      |      |-- p01_a_rgb/
      |      |-- p01_a_t/
      |      |-- p01_a_bvp.csv
      |   |-- p01_b/
      |      |-- p01_b_rgb/
      |      |-- p01_b_t/
      |      |-- p01_b_bvp.csv
      |...
      |   |-- pii_x/
      |      |-- pii_x_rgb/
      |      |-- pii_x_t/
      |      |-- pii_x_bvp.csv
    

  • PURE
    • Stricker, R., Mรผller, S., Gross, H.-M.Non-contact "Video-based Pulse Rate Measurement on a Mobile Service Robot" in: Proc. 23st IEEE Int. Symposium on Robot and Human Interactive Communication (Ro-Man 2014), Edinburgh, Scotland, UK, pp. 1056 - 1062, IEEE 2014

     data/PURE/
     |   |-- 01-01/
     |      |-- 01-01/
     |      |-- 01-01.json
     |   |-- 01-02/
     |      |-- 01-02/
     |      |-- 01-02.json
     |...
     |   |-- ii-jj/
     |      |-- ii-jj/
     |      |-- ii-jj.json
    

  • SCAMPS
    • D. McDuff, M. Wander, X. Liu, B. Hill, J. Hernandez, J. Lester, T. Baltrusaitis, "SCAMPS: Synthetics for Camera Measurement of Physiological Signals", NeurIPS, 2022

     data/SCAMPS/Train/
        |-- P00001.mat
        |-- P00002.mat
     |...
     data/SCAMPS/Val/
        |-- P00001.mat
        |-- P00002.mat
     |...
     data/SCAMPS/Test/
        |-- P00001.mat
        |-- P00002.mat
     |...
    

  • UBFC-rPPG
    • S. Bobbia, R. Macwan, Y. Benezeth, A. Mansouri, J. Dubois, "Unsupervised skin tissue segmentation for remote photoplethysmography", Pattern Recognition Letters, 2017.

     data/UBFC-rPPG/
     |   |-- subject1/
     |       |-- vid.avi
     |       |-- ground_truth.txt
     |   |-- subject2/
     |       |-- vid.avi
     |       |-- ground_truth.txt
     |...
     |   |-- subjectn/
     |       |-- vid.avi
     |       |-- ground_truth.txt
    

๐Ÿ”ง Setup

STEP 1: bash setup.sh

STEP 2: conda activate fsam

STEP 3: pip install -r requirements.txt

๐Ÿ’ป Example of Using Pre-trained Models

Please use config files under ./configs/infer_configs

For example, if you want to run The model trained on PURE and tested on UBFC-rPPG, use python main.py --config_file configs/infer_configs/PURE_UBFC-rPPG_FactorizePhys_FSAM_Res.yaml

๐Ÿ’ป Examples of Neural Network Training

Please use config files under ./configs/train_configs

Training on PURE and Testing on iBVP With FactorizePhys

STEP 1: Download the PURE raw data by asking the paper authors.

STEP 2: Download the iBVP raw data by asking the paper authors.

STEP 3: Modify configs/train_configs/PURE_iBVP_FactorizePhys_FSAM_Res.yaml

STEP 4: Run python main.py --config_file configs/train_configs/PURE_iBVP_FactorizePhys_FSAM_Res.yaml

Note 1: Preprocessing requires only once; thus turn it off on the yaml file when you train the network after the first time.

Note 2: The example yaml setting will allow 100% of PURE to train and and test on iBVP after training 10 for epochs. Alternatively, this can be changed to train using 80% of PURE, validate with 20% of PURE and use the best model(with the least validation loss) to test on iBVP.

Detailed Cross-Dataset Generalization for Reproducibility

Performance Evaluation on iBVP Dataset, for Models Trained with PURE Dataset:

Model Attention Module MAE (HR) โ†“ RMSE (HR) โ†“ MAPE (HR)โ†“ Corr (HR)โ†‘ SNR (BVP)โ†‘ MACC(BVP)โ†‘
PhysNet - 1.63 3.77 2.17 0.92 6.08 0.55
PhysFormer TD-MHSA* 2.50 7.09 3.39 0.79 5.21 0.52
EfficientPhys SASN 3.80 14.82 5.15 0.56 2.93 0.45
EfficientPhys FSAM (Ours) 2.10 4.00 2.94 0.91 4.19 0.49
FactorizePhys (Ours) FSAM (Ours) 1.66 3.55 2.31 0.93 6.78 0.58

Performance Evaluation on iBVP Dataset, for Models Trained with SCAMPS Dataset:

Model Attention Module MAE (HR) โ†“ RMSE (HR) โ†“ MAPE (HR)โ†“ Corr (HR)โ†‘ SNR (BVP)โ†‘ MACC(BVP)โ†‘
PhysNet - 31.85 37.40 45.62 -0.10 -6.11 0.16
PhysFormer TD-MHSA* 41.73 43.89 58.56 0.15 -9.13 0.14
EfficientPhys SASN 26.19 44.55 38.11 -0.12 -2.36 0.30
EfficientPhys FSAM (Ours) 13.40 22.10 19.93 0.05 -3.46 0.24
FactorizePhys (Ours) FSAM (Ours) 2.71 6.22 3.87 0.81 2.36 0.43

Performance Evaluation on iBVP Dataset, for Models Trained with UBFC-rPPG Dataset:

Model Attention Module MAE (HR) โ†“ RMSE (HR) โ†“ MAPE (HR)โ†“ Corr (HR)โ†‘ SNR (BVP)โ†‘ MACC(BVP)โ†‘
PhysNet - 3.18 7.65 4.84 0.70 5.54 0.56
PhysFormer TD-MHSA* 7.86 17.13 11.44 0.38 1.71 0.43
EfficientPhys SASN 2.74 7.07 4.02 0.74 4.03 0.49
EfficientPhys FSAM (Ours) 2.56 6.13 3.71 0.79 4.65 0.50
FactorizePhys (Ours) FSAM (Ours) 1.74 4.39 2.42 0.90 6.59 0.56

Performance Evaluation on PURE Dataset, for Models Trained with iBVP Dataset:

Model Attention Module MAE (HR) โ†“ RMSE (HR) โ†“ MAPE (HR)โ†“ Corr (HR)โ†‘ SNR (BVP)โ†‘ MACC(BVP)โ†‘
PhysNet - 7.78 19.12 8.94 0.59 9.90 0.70
PhysFormer TD-MHSA* 6.58 16.55 6.93 0.76 9.75 0.71
EfficientPhys SASN 0.56 1.40 0.87 1.00 11.96 0.73
EfficientPhys FSAM (Ours) 0.44 1.19 0.64 1.00 12.64 0.75
FactorizePhys (Ours) FSAM (Ours) 0.60 1.70 0.87 1.00 15.19 0.77

Performance Evaluation on PURE Dataset, for Models Trained with SCAMPS Dataset:

Model Attention Module MAE (HR) โ†“ RMSE (HR) โ†“ MAPE (HR)โ†“ Corr (HR)โ†‘ SNR (BVP)โ†‘ MACC(BVP)โ†‘
PhysNet - 26.74 36.19 46.73 0.45 -2.21 0.31
PhysFormer TD-MHSA* 16.64 28.13 30.58 0.51 0.84 0.42
EfficientPhys SASN 6.21 18.45 12.16 0.74 4.39 0.51
EfficientPhys FSAM (Ours) 8.03 19.09 15.12 0.73 3.81 0.48
FactorizePhys (Ours) FSAM (Ours) 5.43 15.80 11.10 0.80 11.40 0.67

Performance Evaluation on PURE Dataset, for Models Trained with UBFC-rPPG Dataset:

Model Attention Module MAE (HR) โ†“ RMSE (HR) โ†“ MAPE (HR)โ†“ Corr (HR)โ†‘ SNR (BVP)โ†‘ MACC(BVP)โ†‘
PhysNet - 10.38 21.14 20.91 0.66 11.01 0.72
PhysFormer TD-MHSA* 8.90 18.77 17.68 0.71 8.73 0.66
EfficientPhys SASN 4.71 14.52 7.63 0.80 8.77 0.66
EfficientPhys FSAM (Ours) 3.69 13.27 5.85 0.83 9.65 0.68
FactorizePhys (Ours) FSAM (Ours) 0.48 1.39 0.72 1.00 14.16 0.78

Performance Evaluation on UBFC-rPPG Dataset, for Models Trained with iBVP Dataset:

Model Attention Module MAE (HR) โ†“ RMSE (HR) โ†“ MAPE (HR)โ†“ Corr (HR)โ†‘ SNR (BVP)โ†‘ MACC(BVP)โ†‘
PhysNet - 3.09 10.72 2.83 0.81 7.13 0.81
PhysFormer TD-MHSA* 9.88 19.59 8.72 0.44 2.80 0.70
EfficientPhys SASN 1.14 2.85 1.42 0.99 8.71 0.84
EfficientPhys FSAM (Ours) 1.17 2.87 1.31 0.99 8.54 0.85
FactorizePhys (Ours) FSAM (Ours) 1.04 2.40 1.23 0.99 8.84 0.86

Performance Evaluation on UBFC-rPPG Dataset, for Models Trained with PURE Dataset:

Model Attention Module MAE (HR) โ†“ RMSE (HR) โ†“ MAPE (HR)โ†“ Corr (HR)โ†‘ SNR (BVP)โ†‘ MACC(BVP)โ†‘
PhysNet - 1.23 2.65 1.42 0.99 8.34 0.85
PhysFormer TD-MHSA* 1.01 2.40 1.21 0.99 8.42 0.85
EfficientPhys SASN 1.41 3.16 1.68 0.98 6.87 0.79
EfficientPhys FSAM (Ours) 1.20 2.92 1.50 0.99 7.37 0.79
FactorizePhys (Ours) FSAM (Ours) 1.04 2.44 1.23 0.99 8.88 0.87

Performance Evaluation on UBFC-rPPG Dataset, for Models Trained with SCAMPS Dataset:

Model Attention Module MAE (HR) โ†“ RMSE (HR) โ†“ MAPE (HR)โ†“ Corr (HR)โ†‘ SNR (BVP)โ†‘ MACC(BVP)โ†‘
PhysNet - 11.24 18.81 13.55 0.38 -0.09 0.48
PhysFormer TD-MHSA* 8.42 17.73 11.27 0.49 2.29 0.61
EfficientPhys SASN 2.18 4.82 2.35 0.96 4.40 0.67
EfficientPhys FSAM (Ours) 2.69 5.20 3.16 0.95 3.74 0.63
FactorizePhys (Ours) FSAM (Ours) 1.17 2.56 1.35 0.99 8.41 0.82

Acknowledgement

We would like to express sincere thanks to the authors of rPPG-Toolbox, Liu et al., 2023, building upon which, we developed this repo. For detailed usage related instructions, please refer the GitHub repo of the rPPG-Toolbox.

Citation

@inproceedings{joshi2024factorizephys,
 title={FactorizePhys: Matrix Factorization for Multidimensional Attention in Remote Physiological Sensing},
 author={Jitesh Joshi and Sos Agaian and Youngjun Cho},
 booktitle={The Thirty-eighth Annual Conference on Neural Information Processing Systems},
 year={2024},
 url={https://openreview.net/forum?id=qrfp4eeZ47}
}