Official codebase for "Continual Spatio-Temporal Graph Convolutional Networks" (Pattern Recognition, 2023), including:
-
Models: Co ST-GCN, Co AGCN, Co S-TR, and more ... (see Models section for full overview).
-
Datasets: NTU RGB+D 60, NTU RGB+D 120, and Kinetics Skeleton 400.
Graph-based reasoning over skeleton data has emerged as a promising approach for human action recognition. However, the application of prior graph-based methods, which predominantly employ whole temporal sequences as their input, to the setting of online inference entails considerable computational redundancy. In this paper, we tackle this issue by reformulating the Spatio-Temporal Graph Convolutional Neural Network as a Continual Inference Network, which can perform step-by-step predictions in time without repeat frame processing. To evaluate our method, we create a continual version of ST-GCN, CoST-GCN, alongside two derived methods with different self-attention mechanisms, CoAGCN and CoS-TR. We investigate weight transfer strategies and architectural modifications for inference acceleration, and perform experiments on the NTU RGB+D 60, NTU RGB+D 120, and Kinetics Skeleton 400 datasets. Retaining similar predictive accuracy, we observe up to 109x reduction in time complexity, on-hardware accelerations of 26x, and reductions in maximum allocated memory of 52% during online inference.
Fig. 1: Continual Spatio-temporal Graph Convolution Blocks consist of an in-time Graph Convolution followed by an across-time Continual Convolution (here a kernel size of three is depicted). The residual connection is delayed to ensure temporal alignment with the continual temporal convolution that is weight-compatible with non-continual networks.
Fig. 2: Accuracy/complexity trade-off on NTU RGB+D 60 X-Sub for ⬥ Continual and ■ prior methods during online inference. Numbers denote streams for each method. *Architecture modification with stride one and no padding.
- Clone this repository and enter it:
git clone https://github.com/LukasHedegaard/continual-skeletons.git cd continual-skeletons
- Optionally create and activate conda environment:
conda create --name continual-skeletons python=3.8 conda activate continual-skeletons
- Install as editable module
pip install -e .[dev]
The repository is s
root
|- datasets/ # Dataset loaders
|- models/ # Individual models and shared base-code
|- ...
|- st_gcn/ # Baseline model
|- cost_gcn/ # Continual version of model
|- st_gcn_mod/ # Modified baseline with stride one and no padding
|- cost_gcn_mod/ # Continual version of modified baseline model
|- cost_gcn_mod.py # Python entry-point
|- scripts/ # Scripts used to achieve results from paper. Please run from root.
|- evaluate_ntu60.py
|- evaluate_ntu120.py
|- evaluate_kinetics.py
|- ...
|- tests/ # Unit tests for custom modules
|- weights/ # Place pretrained weights here
|- preds/ # Place extracted predictions here to perform multi-stream eval
|- Makefile # Commands for testing, linting, cleaning.
|- .env # Modify path to your dataset here, i.e. DATASETS_PATH=/my/path
Download the skeleton data of NTU-RGBD-60 and NTU-RGBD-120 from here and put them in nturgbd_raw directory.
Name the folder of the downloaded skeletons folder for NTU-RGBD-60 and NTU-RGBD-120 as nturgb+d_skeletons60
, nturgb+d_skeletons120
, respectively.
The skeleton data for Kinetics dataset is extracted using Openpose toolbox by ST-GCN authors.
The extracted skeleton data called Kinetics-skeleton (7.5GB) can be directly downloaded from GoogleDrive, and it should be placed in kinetics_raw directory.
Before training and testing the models, the datasets should be preprocessed. The downloaded data should be placed in the following directories:
root
|- datasets/
|- data_preparation/
|- nturgbd_raw\ # Raw NTU-RGBD skeleton data
|- nturgb+d_skeletons60\ # Skeleton data for NTU-RGBD-60
|- nturgb+d_skeletons120\ # Skeleton data for NTU-RGBD-120
|- ...
|- ntu60_samples_with_missing_skeletons.txt # Sample IDs with missing skeletons in NTU-RGBD-60
|- ntu120_samples_with_missing_skeletons.txt # Sample IDs with missing skeletons in NTU-RGBD-120
|- kinetics_raw\ # Raw Kinetics data
|- kinetics_train\
|- ...
|- kinetics_val\
|- ...
|- kinetics_train_label.json
|- kinetics_val_label.json
For generating the preprocedded data you need to run the following commands:
# NTU-RGBD-60
python datasets/data_preparation/ntu60_prep.py
# NTU-RGBD-120
python datasets/data_preparation/ntu120_prep.py
# Kinetics
python datasets/data_preparation/kinetics400_prep.py
For generating bone and motion data for each of the datasets run the following commands:
# Bone generation
python datasets/data_preparation/bone_data_prep.py
# Motion generation
python datasets/data_preparation/motion_data_prep.py
The joint and bone skeleton data can be concatenated by running the following command:
# joint_bone data concatenation
python datasets/data_preparation/merge_joint_bone_data.py
Individual folders with relevant scripts are avilable under /models
for the following models:
- ST-GCN (baseline)
- ST-GCN*
- Co ST-GCN
- Co ST-GCN*
- AGCN (baseline)
- AGCN*
- Co AGCN
- Co AGCN*
- S-TR (baseline)
- S-TR*
- Co S-TR
- Co S-TR*
To see an overview of available commands for a model, check the help, e.g.:
python models/cost_gcn/cost_gcn.py --help
The commands used to produce the paper results are found in the associated scripts folder, e.g.:
python models/cost_gcn/scripts/evaluate_ntu60.py
Trained model weights are available here.
To reproduce results:
- Prepare datasets
- Download and preprocessing guidelines coming up
- Add
DATASET_PATH=/your/dataset/path
to.env
.
- Download pretrained weights and place them in
~/weights
. - Run evaluation script. For instance, to evaluate the CoST-GCN* model on NTU RGB+D 120 and save its predictions, the command would be:
python models/cost_gcn/scripts/evaluate_ntu120.py
@article{hedegaard2023continual,
title = {Continual spatio-temporal graph convolutional networks},
author = {Lukas Hedegaard and Negar Heidari and Alexandros Iosifidis},
journal = {Pattern Recognition},
volume = {140},
pages = {109528},
year = {2023},
issn = {0031-3203},
doi = {https://doi.org/10.1016/j.patcog.2023.109528},
}