We present a novel neural light field representation for efficient view synthesis.

Project page | Paper | Data
This is the official repo for the implementation of Progressively-connected light field network for efficient view synthes.
Clone this repository
git clone https://github.com/Totoro97/ProLiF.git
cd ProLiF
pip install -r requirements.txt- Training for novel view synthesis
python exp_runner.py --config-name=prolif case_name=<case_name>- Training for scene fitting under varing light conditions
python exp_runner.py --config-name=prolif-lpips case_name=<case_name>- Training for text-guided scene style editing
python exp_runner.py --config-name=prolif_clip case_name=<case_name>- Rendering all test views
python exp_runner.py --config-name=<config_name> case_name=<case_name> mode=validate is_continue=true # use latest checkpointThe synthesized images can be found in exp/<case_name>/<exp_name>/validation.
- Rendering video
python exp_runner.py --config-name=<config_name> case_name=<case_name> mode=video is_continue=true # use latestcheck pointThe synthesized video can be found in exp/<case_name>/<exp_name>/video.
We follow the same data convention as LLFF. You may follow the original LLFF instruction for data preparation.
Cite as below if you find this repository is helpful to your project:
@article{wang2021prolif,
author = {Wang, Peng and Liu, Yuan and Lin, Guying and Gu, Jiatao and Liu, Lingjie and Komura, Taku and Wang, Wenping},
title = {Progressive-connected Light Field Network for Efficient View Synthesis},
journal = {Arxiv},
year = {2022},
}