Skip to content

This is the official implementation for ECCV20 paper: Key Frame Proposal Network for Efficient Pose Estimation in Videos

License

Notifications You must be signed in to change notification settings

Yuexiaoxi10/Key-Frame-Proposal-Network-for-Efficient-Pose-Estimation-in-Videos

Repository files navigation

Key-Frame-Proposal-Network-for-Efficient-Pose-Estimation-in-Videos

Paper: https://arxiv.org/abs/2007.15217

This is the official implementation for ECCV20 paper: Key Frame Proposal Network for Efficient Pose Estimation in Videos

All implementations are based on Pytorch and up to date

Prerequires

python = 3.5/3.6

pytorch >= 0.4.0

cuda >= cuda 9.0

Usage

  1. Downloading data:
  1. Downloading pre-train models:
  • We provided our best model for Penn Action, sub-JHMDB, and online updating experiments
  • Please download models: GoogleDrive OneDrive
  • Please place models under folder 'models'

NOTE:

  1. Our demo code is implemented with our best model on validation set for each dataset

  2. Our implementations are able to integrate other Resnet backbones

  3. We updated our Human Pose Interpolation Module, instead of using Least-Square solution in paper, we implemented with FISTA that is able to integrate into deep network for later use. Results were improved with comparable running time.
    Details of FISTA can be found: FISTA

  4. For more details of our dynamic-based dictionary, please refer: DYAN

Details

1. Training:

  • We provded training files for each of experiments
  • For training Penn Action and Sub-JHMDB: please not, parameter 'alpha' that we disscussed in our paper is linearly increasing with the training epoch, you may need to tune by yourself while training.
  • For online updating experiment on sub-JHMDB: we used our best model as the backbone

2. Testing:

  • We provded testing files for each of experiments, all results reported in Table 3,4,5 were based on the baseline, we directly applied their model on both Penn Action and sub-JHMDB dataset, no fine-tuning involved
  • Please referOur baseline, we used architecture of '384x384_pose_resnet_101_d256d256d256' which trained on MPII
  • You can find 'sampling comparison' experiment in 'test_penn.py', and 'robustness comparison' experiment in 'test_jhmdb.py'
  • Note: set 'if_occ' is 'True' for doing 'robustness comparison' experiment, and you can change occlusion ratio by setting 'occRatio'

Citation

If you find this repo and our paper is useful, please consider to cite:

@inproceedings{k-fpn, author = {Yuexi Zhang and Yin Wang and Octavia Camps and Mario Sznaier}, booktitle = {Proceedings of the European Conference on Computer Vision(ECCV)}, title = {Key {F}rame {P}roposal {N}etwork for {E}fficient {P}ose {E}stimation in {V}ideos}, year = {2020}, month = {August} }

Zhang, Y., Wang, Y., Camps, O., Sznaier, M.: Key Frame Proposal Network forEfficient Pose Estimation in Videos. In: Proceedings of the European Conference on Computer Vision(ECCV) (August 2020)

About

This is the official implementation for ECCV20 paper: Key Frame Proposal Network for Efficient Pose Estimation in Videos

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages