Skip to content

[MICCAI 2024] Surgformer: Surgical Transformer with Hierarchical Temporal Attention for Surgical Phase Recognition

License

Notifications You must be signed in to change notification settings

isyangshu/Surgformer

Repository files navigation

[MICCAI 2024] Surgformer: Surgical Transformer with Hierarchical Temporal Attention for Surgical Phase Recognition

License: MIT

Abstract

Existing state-of-the-art methods for surgical phase recognition either rely on the extraction of spatial-temporal features at short-range temporal resolution or adopt the sequential extraction of the spatial and temporal features across the entire temporal resolution. However, these methods have limitations in modeling spatial-temporal dependency and addressing spatial-temporal redundancy: 1) These methods fail to effectively model spatial-temporal dependency, due to the lack of long-range information or joint spatial-temporal modeling. 2) These methods utilize dense spatial features across the entire temporal resolution, resulting in significant spatial-temporal redundancy. In this paper, we propose the Surgical Transformer (Surgformer) to address the issues of spatial-temporal modeling and redundancy in an end-to-end manner, which employs divided spatial-temporal attention and takes a limited set of sparse frames as input. Moreover, we propose a novel Hierarchical Temporal Attention (HTA) to capture both global and local information within varied temporal resolutions from a target frame-centric perspective. Distinct from conventional temporal attention that primarily emphasizes dense long-range similarity, HTA not only captures long-term information but also considers local latent consistency among informative frames. HTA then employs pyramid feature aggregation to effectively utilize temporal information across diverse temporal resolutions, thereby enhancing the overall temporal representation. Extensive experiments on two challenging benchmark datasets verify that our proposed Surgformer performs favorably against the state-of-the-art methods.

NOTES

2024-10-01: Without any adjustment and surgical parameters, we win the 2nd in OSS challenge and 3rd in PhaKIR.

2024-06-26: We release the full version of Surgformer.

2024-05-14: Our paper is early accepted for MICCAI 2024.

Installation

  • Environment: CUDA 11.4 / Python 3.8
  • Device: NVIDIA GeForce RTX 3090
  • Create a virtual environment
> conda env create -f Surgformer.yml

How to Train

Prepare your data

  1. Download raw video data from Cholec80 and AutoLaparo;

You need to fill the request form to access both datasets.

  1. Use the pre-processing tools provided by us to extract frames and generate pickle files for training.
# Extract frames form raw videos
python datasets/data_preprosses/extract_frames_ch80.py
python datasets/data_preprosses/extract_frames_autolaparo.py

# Generate .pkl for training
python datasets/data_preprosses/generate_labels_ch80.py
python datasets/data_preprosses/generate_labels_autolaparo.py

# Note that you can change the size of each frame to reduce I/O load.
  1. You can also use the cutting tool provided by TMRNet to cut black margin for surgical videos in Cholec80, which may help to get better performance.
# Cut black margin
python datasets/data_preprosses/frame_cutmargin.py

The final structure of datasets should be as following:

Note that you can change fps in the Step.2 to generate more frames.

data/
    └──Cholec80/
        └──frames/
            └──video01
                ├──00000.png
                ├──00001.png
                └──...
            ├──...    
            └──video80
        └──frames_cutmargin/
        └──labels/
            └──train
                ├── 1pstrain.pickle
                └── ...
            └──test
                ├── 1psval_test.pickle
                └── ...
    └──AutoLaparo/
        └──frames/
            └──01
                ├──00000.png
                ├──00001.png
                └──...
            ├──...    
            └──21
        └──labels_pkl/
            └──train
            ├──val
            └──test

Pretrained Parameters

We use the parameters of TimeSformer trained on K400 dataset with frames 8 and spatial size 224 as initialization parameters.

You can try to use ImageNet(MAE), which may help to get better performance.

Training

We provide the script for training train_phase.sh.

run the following code for training

sh scripts/train.sh

You need to modify data_path, eval_data_path, output_dir and log_dir according to your own setting.

Optional settings
Model: surgformer_base surgformer_HTA surgformer_HTA_KCA
Dataset: Cholec80 AutoLaparo

Test

Currently, the test and evaluation codes we provide are only applicable to two-GPU inference.

  1. run the following code for testing, and get 0.txt and 1.txt;
sh scripts/test.sh
  1. Merge the files and generate separate txt file for each video;
python datasets/convert_results/convert_cholec80.py
python datasets/convert_results/convert_autolaparo.py
  1. Use Matlab Evaluation Code to compute metrics;

Acknowledgements

Huge thanks to the authors of following open-source projects:

License & Citation

If you find our work useful in your research, please consider citing our paper at:

coming soon

This code is available for non-commercial academic purposes. If you have any question, feel free to email Shu YANG.

About

[MICCAI 2024] Surgformer: Surgical Transformer with Hierarchical Temporal Attention for Surgical Phase Recognition

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published