Skip to content

SeungjunNah/DeepDeblur_release

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

50 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

DeepDeblur_release

Single image deblurring with deep learning.

This is a project page for our research. Please refer to our CVPR 2017 paper for details:

Deep Multi-scale Convolutional Neural Network for Dynamic Scene Deblurring [paper] [supplementary] [slide]

If you find our work useful in your research or publication, please cite our work:

@InProceedings{Nah_2017_CVPR,
  author = {Nah, Seungjun and Kim, Tae Hyun and Lee, Kyoung Mu},
  title = {Deep Multi-Scale Convolutional Neural Network for Dynamic Scene Deblurring},
  booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
  month = {July},
  year = {2017}
}

PyTorch version

PyTorch version is now available: https://github.com/SeungjunNah/DeepDeblur-PyTorch

New dataset released!

Check out our new REDS dataset! In CVPR 2019, I co-organized the 4th NTIRE workshop and the corresponding video restoration challenges. We released the REDS dataset for challenge participants to train and evaluate video deblurring / super-resolution methods. Special thanks go to my colleagues, Sungyong Baik, Seokil Hong, Gyeongsik Moon, Sanghyun Son, Radu Timofte and Kyoung Mu Lee for collecting, processing, and releasing the dataset together.

Updates

Downloads are now available for training, validation, and test input data. A public leaderboard site is under construction. Download page: https://seungjunnah.github.io/Datasets/reds

Dependencies

  luarocks install torchx
  • cudnn
    cd ~/torch/extra/cudnn
    git checkout R7 # R7 is for cudnn v7
    luarocks make

Code

To run demo, download and extract the trained models into "experiment" folder.

Type following command in "code" folder.

qlua -i demo.lua -load -save release_scale3_adv_gamma -blur_type gamma2.2 -type cudaHalf
qlua -i demo.lua -load -save release_scale3_adv_lin -blur_type linear -type cudaHalf

To train a model, clone this repository and download below dataset in "dataset" directory.

The data structure should look like "dataset/GOPRO_Large/train/GOPRxxxx_xx_xx/blur/xxxxxx.png"

Then run main.lua in "code" directory with optional parameters.

th main.lua -nEpochs 450 -save scale3 # Train for 450 epochs, save in 'experiment/scale3'
th main.lua -load -save scale3  # Load saved model
> blur_dir, output_dir = ...
> deblur_dir(blur_dir, output_dir)

Optional parameters are listed in opts.lua

ex) -type: Operation type option. Supports cuda and cudaHalf. Half precision CNN has similar accuracy as single precision in evaluation mode. However, fp16 training is not meant to be supported in this code. ADAM optimizer is hard to use with fp16.

Dataset

In this work, we proposed a new dataset of realistic blurry and sharp image pairs using a high-speed camera. However, we do not provide blur kernels as they are unknown.

  • Downloads available here
Statistics Training Test Total
sequences 22 11 33
image pairs 2103 1111 3214

Here are some example images.

Blurry image example 1 Blurry image

Sharp image example 1 Sharp image

Blurry image example 2 Blurry image

Sharp image example 2 Sharp image

Acknowledgment

This project is partially funded by Microsoft Research Asia

About

Deep Multi-scale CNN for Dynamic Scene Deblurring

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages