Diffuse3D: Wide-Angle 3D Photography via Bilateral Diffusion
Yutao Jiang, Yang Zhou, Yuan Liang, Wenxi Liu, Jianbo Jiao, Yuhui Quan, and Shengfeng He
IEEE International Conference on Computer Vision (ICCV), 2023.
We recommend running our code with conda environment:
conda create -n d3d python=3.9
conda activate d3d
pip install -r requirements.txt
Next, please download the model weight using the following command:
bash download_model.sh
The input to our model is a RGBD image, we recommend using the DPT to generate depth image.
The dataset directory should have the following structure:
dataset/
├── scene-1
│ ├── src.png
│ ├── src_depth.png
│ ├── ...
├── scene-2
│ ├── ...
├── ...
You can generate novel views with this command:
python inference.py --data_dir <DATASET_DIR> --output_dir <OUTPUT_DIR>
This work is licensed under MIT License.
If you find our work useful, please consider citing our paper:
@inproceedings{jiang2023diffuse3d,
title={Diffuse3D: Wide-Angle 3D Photography via Bilateral Diffusion},
author={Jiang, Yutao and Zhou, Yang and Liang, Yuan and Liu, Wenxi and Jiao, Jianbo and Quan, Yuhui and He, Shengfeng},
booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
pages={8998--9008},
year={2023}
}
Our code is heavily borrowed from 3D Photo and LDM, we thank the author for their great works.