2.7 KiB
Shape Generation and Completion Through Point-Voxel Diffusion
Implementation of Shape Generation and Completion Through Point-Voxel Diffusion
Linqi Zhou, Yilun Du, Jiajun Wu
Requirements:
Install Python environment
module load conda
module load artifactory
mamba env create --file env.yml
Install PyTorchEMD by
cd metrics/PyTorchEMD
module load gcc/11.2.0
module load mpfr/4.0.2
conda activate PVD
python setup.py install
cp build/**/emd_cuda.cpython-*-x86_64-linux-gnu.so .
The code was tested on Unbuntu with Titan RTX.
Data
For generation, we use ShapeNet point cloud, which can be downloaded here.
For completion, we use ShapeNet rendering provided by GenRe.
We provide script convert_cam_params.py
to process the provided data.
For training the model on shape completion, we need camera parameters for each view which are not directly available. To obtain these, simply run
$ python convert_cam_params.py --dataroot DATA_DIR --mitsuba_xml_root XML_DIR
which will create ..._cam_params.npz
in each provided data folder for each view.
Pretrained models
Pretrained models can be downloaded here.
Training:
$ python train_generation.py --category car|chair|airplane
Please refer to the python file for optimal training parameters.
Testing:
$ python train_generation.py --category car|chair|airplane --model MODEL_PATH
Results
Some generation and completion results are as follows.
Multimodal completion on a ShapeNet chair.
Multimodal completion on PartNet.
Multimodal completion on two Redwood 3DScan chairs.
Reference
@inproceedings{Zhou_2021_ICCV,
author = {Zhou, Linqi and Du, Yilun and Wu, Jiajun},
title = {3D Shape Generation and Completion Through Point-Voxel Diffusion},
booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)},
month = {October},
year = {2021},
pages = {5826-5835}
}
Acknowledgement
For any questions related to codes and experiment setting, please contact Linqi Zhou and Yilun Du.