Geometry- and Light-aware ControlNet
The geometry- and light-aware ControlNet uses an object's normal and depth maps as geometry conditions and six predefined materials with a given environment light as lighting conditions. Our model generates images that align with the given geometry and environment light.
Material Generation
This ControlNet can be used in distillation process to generate PBR materials. Please refer to the paper "DreamMat: High-quality PBR Material Generation with Geometry- and Light-aware Diffusion Models" for technical details.
Project Page β’ Arxiv β’ GitHub
Training Dataset
We train the geometry- and light-aware ControlNet from the images which are rendered on the objects in the LVIS subset of the Objaverse. Since the names and tags of objects in this dataset are rather noisy, we employ BLIP for captioning all rendered images. We render 16 random views for every object under randomly chosen environment light maps. The light condition maps are obtained by using ray tracing in Blender, which represents the radiance for different materials under the environment light. For normal maps, we transform the model's normal vectors into view space and flip the x-axis following ScanNetβs protocol. Depth maps are processed by inverting the real depth values and normalizing them.
π Citation
@inproceedings{zhang2024dreammat,
title={DreamMat: High-quality PBR Material Generation with Geometry- and Light-aware Diffusion Models},
author={Zhang, Yuqing and Liu, Yuan and Xie, Zhiyu and Yang, Lei and Liu, Zhongyuan and Yang, Mengzhou and Zhang, Runze and Kou, Qilong and and Lin, Cheng and Wang, Wenping and Jin, Xiaogang},
booktitle={SIGGRAPH},
year={2024}
}
- Downloads last month
- 29
Model tree for zzzyuqing/light-geo-controlnet
Base model
stabilityai/stable-diffusion-2-1-base