Deep Video Deblurring
This is the demo code for Deep Video Deblurring. Given a stack of pre-aligned input frames, our network predicts a sharper central image.
Prepare data
-
Download and unzip test videos to
dataset
, from this link, or place your own test video frames underdataset/qualitative_datasets/[video_file_name]/input
. -
Align frames in Matlab, by running
preprocess/launcher.m
Outputs should be stored at data/testing_real_all_nostab_[alignment]
under structure
/image_-2
/image_-1
/image_0
/image_1
/image_2
Alternatively, you can download the pre-aligned qualitative videos from here, here, and here.
Download pretrained weights
- Download and unzip pretrained weights into
logs
, from here.
Run prediction script
- Run script:
sh run_pred.sh
- Results will be saved to
outImg
.
Citation
If you find this code useful for your research, please cite:
@article{su2016deep,
title={Deep Video Deblurring},
author={Su, Shuochen and Delbracio, Mauricio and Wang, Jue and Sapiro, Guillermo and Heidrich, Wolfgang and Wang, Oliver},
journal={arXiv preprint arXiv:1611.08387},
year={2016}
}