Skip to content

Official code release accompanying the paper "Revealing disocclusions through temporal view synthesis"

License

Notifications You must be signed in to change notification settings

NagabhushanSN95/IVP

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

IVP - Infilling Vector Prediction

Official Code Release accompanying the WACV 2022 paper "Revealing Disocclusions in Temporal View Synthesis through Infilling Vector Prediction"

Databases

  • Download the IISc VEED Database. Extract the zip files and place them in Data/Veed.
  • For SceneNet database, download all the ground truth for the training set from here. Extract the zip files and place them in Data/SceneNet. The following steps describe training and testing on IISc VEED dataset. The steps for SceneNet dataset are similar and the code for each step is also provided.

Python Environment

Environment details are available in IVP.yml for conda and requirements.txt for pip. To create the environment using conda

conda env create -f IVP.yml

Training and Inference

  1. To train the IVP model on IISc VEED dataset,
cd src
python Trainer.py --configs configs/Configs_VEED.json --generate_data --correct_depth
cd ..

Our model generates some data before starting training. The --generate_data flag instructs the Trainer.py to generate this data. If the data has already been generated, this flag can be omitted. Many datasets have errors in the depth maps, usually at the foreground-background boundaries. The correct_depth flag corrects this depth before further processing. If you have clean depth and do not want to employ depth correction, omit this flag. More details about depth correction can be found here. To train on SceneNet dataset, use the configs here.

  1. To run inference on IISc VEED dataset,
cd src
python Test.py --configs configs/Configs_VEED.json --generate_data --correct_depth --test_num 3
cd ..

The configs parameter and generate_data and correct_depth flags are similar to the training code. The predicted frames will be saved in a folder named Test0003. You can specify different test_num if testing on different datasets with the same trained model.

Citation

@inproceedings{kanchana2022ivp,
    title = {Revealing Disocclusions in Temporal View Synthesis through Infilling Vector Prediction},
    author = {Kanchana, Vijayalakshmi and Somraj, Nagabhushan and Yadwad, Suraj and Soundararajan, Rajiv},
    booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)},
    pages = {3541--3550},
    month = {January},
    year = {2022},
    doi = {10.1109/WACV51458.2022.00315}
}

Acknowledgments

The code for depth based warping is borrowed from here.

For any queries or bugs related to either the IVP code or the IISc VEED database, please raise an issue.

About

Official code release accompanying the paper "Revealing disocclusions through temporal view synthesis"

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages