We propose a new method for high quality reconstruction of indoor scenes using Hybrid Representation and Normal Prior Enhancement. The paper has been accepted by TVCG.
The data is organized as follows:
<scene_name>
|-- cameras_sphere.npz # camera parameters
|-- image
|-- 0000.png # target image for each view
|-- 0001.png
...
|-- weight
|-- 0000.npy # uncertainty weight for each view
|-- 0001.npy
...
|-- pose
|-- 0000.txt # camera pose for each view
|-- 0001.txt
...
|-- pred_normal_refine
|-- 0000.npz # predicted normal for each view
|-- 0001.npz
...
|-- xxx.ply # GT mesh or point cloud from MVS
|-- trans_n2w.txt # transformation matrix from normalized coordinates to world coordinates
Please refer to scripts/preprocess.sh
for step-by-step data processing. We also provide demo pre-processed data (download), which can be directly reconstructed using our method.
For the weight of our uncertainty estimation module (U-Net), here is the download link: https://drive.google.com/file/d/18mvYnUISLIG9yKdbWeKgHiB5yZ_6Eym6/view?usp=sharing
conda create -n recon python=3.8
conda activate recon
conda install pytorch=1.9.0 torchvision torchaudio cudatoolkit=10.2 -c pytorch
pip install -r requirements.txt
python ./exp_runner.py --mode train --conf ./confs/scannet_0050.conf --gpu 0
python exp_runner.py --mode validate_mesh --conf ./confs/scannet_0050.conf --is_continue --gpu 0
python ./exp_evaluation.py --scene scene0050_00 --iter 50000 --reso 512