Paper: https://arxiv.org/abs/2007.15217
This is the official implementation for ECCV20 paper: Key Frame Proposal Network for Efficient Pose Estimation in Videos
All implementations are based on Pytorch and up to date
python = 3.5/3.6
pytorch >= 0.4.0
cuda >= cuda 9.0
- Downloading data:
- Penn Action: http://dreamdragon.github.io/PennAction/
- JHMDB: http://jhmdb.is.tue.mpg.de/dataset
- please refer get bbox to generate bbox for later use
- Downloading pre-train models:
- We provided our best model for Penn Action, sub-JHMDB, and online updating experiments
- Please download models: GoogleDrive OneDrive
- Please place models under folder 'models'
-
Our demo code is implemented with our best model on validation set for each dataset
-
Our implementations are able to integrate other Resnet backbones
-
We updated our Human Pose Interpolation Module, instead of using Least-Square solution in paper, we implemented with FISTA that is able to integrate into deep network for later use. Results were improved with comparable running time.
Details of FISTA can be found: FISTA -
For more details of our dynamic-based dictionary, please refer: DYAN
- We provded training files for each of experiments
- For training Penn Action and Sub-JHMDB: please not, parameter 'alpha' that we disscussed in our paper is linearly increasing with the training epoch, you may need to tune by yourself while training.
- For online updating experiment on sub-JHMDB: we used our best model as the backbone
- We provded testing files for each of experiments, all results reported in Table 3,4,5 were based on the baseline, we directly applied their model on both Penn Action and sub-JHMDB dataset, no fine-tuning involved
- Please referOur baseline, we used architecture of '384x384_pose_resnet_101_d256d256d256' which trained on MPII
- You can find 'sampling comparison' experiment in 'test_penn.py', and 'robustness comparison' experiment in 'test_jhmdb.py'
- Note: set 'if_occ' is 'True' for doing 'robustness comparison' experiment, and you can change occlusion ratio by setting 'occRatio'
If you find this repo and our paper is useful, please consider to cite:
@inproceedings{k-fpn, author = {Yuexi Zhang and Yin Wang and Octavia Camps and Mario Sznaier}, booktitle = {Proceedings of the European Conference on Computer Vision(ECCV)}, title = {Key {F}rame {P}roposal {N}etwork for {E}fficient {P}ose {E}stimation in {V}ideos}, year = {2020}, month = {August} }
Zhang, Y., Wang, Y., Camps, O., Sznaier, M.: Key Frame Proposal Network forEfficient Pose Estimation in Videos. In: Proceedings of the European Conference on Computer Vision(ECCV) (August 2020)