forked from coupriec/VideoPredictionICLR2016
-
Notifications
You must be signed in to change notification settings - Fork 0
License
namanjain007/VideoPredictionICLR2016
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
July 2016 Authors: Michael Mathieu, Camille Couprie Update: due to large files that could not be stored on github, the trained models and dataset may be found at: http://perso.esiee.fr/~coupriec/MathieuICLR16TestCode.zip This repository contains: - Test code for the ICLR 2016 paper: [1] Michael Mathieu, Camille Couprie, Yann LeCun: "Deep multi-scale video prediction beyond mean square error". http://arxiv.org/abs/1511.05440 http://cs.nyu.edu/~mathieu/iclr2016.html - Two trained models (using adversarial+l2norm training or adversarial+l1norm+gdl training). - A subset of the UCF101 test dataset [2] with optical flow results to perform an evaluation in moving area as described in [1]. - A training script for the model. Because the Sports1m dataset is hard to get, we cannot provide an easy script to train on it. Instead, we propose a script to train on UCF101, which is significantly smaller. Main files: - For testing: test-frame-prediction-on-ucf-rec_gdl.lua Script to test 2 trained models to predict future frames in video from 4 previous ones on a subset of the UCF101 test dataset. - For training: - For training: train_iclr_model.lua Script to train a model from scratch on the UCF101 dataset. If you want to train on the Sports1m dataset, you will need to download it and write a datareader, similar to datasources/ucf101.lua . Usage: 1- Install torch and the packages (standard packages + nngraph, cudnn.torch, gfx.js) For testing: 2- Uncompress the provided archives. 3- Run the main script : th test-frame-prediction-on-ucf-rec_gdl.lua It generates results (2 predicted images + animated gifs) in a directory named 'AdvGDL'. It also display the average PSNR and SSIM of the 2 first predicted frames following the evaluation presented in [1]. For training: 2- Get the UCF101 dataset (requires unrar, modify the script if you have another .rar extractor): cd datasources python get_datasource.py 3- Get thffpmeg from https://github.com/MichaelMathieu/THFFmpeg 4- Run the training script: th train_iclr_model.lua 5- For visualizing the intermediate results, start the gfx.js server th -lgfx.start And go to http://localhost:8000 in your internet browser. [2]:Khurram Soomro, Amir Roshan Zamir and Mubarak Shah, UCF101: A Dataset of 101 Human Action Classes From Videos in The Wild., CRCV-TR-12-01, November, 2012.
About
No description, website, or topics provided.
Resources
License
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published
Languages
- Lua 98.6%
- Python 1.4%