An evaluation of SVBRDF Prediction from Generative Image Models for Appearance Modeling of 3D Scenes

Abstract
Digital content creation is experiencing a profound change with the advent of deep generative models. For texturing, conditional image generators now allow the synthesis of realistic RGB images of a 3D scene that align with the geometry of that scene. For appearance modeling, SVBRDF prediction networks recover material parameters from RGB images. Combining these technologies allows us to quickly generate SVBRDF maps for multiple views of a 3D scene, which can be merged to form a SVBRDF texture atlas of that scene. In this paper, we analyze the challenges and opportunities for SVBRDF prediction in the context of such a fast appearance modeling pipeline. On the one hand, single-view SVBRDF predictions might suffer from multiview incoherence and yield inconsistent texture atlases. On the other hand, generated RGB images, and the different modalities on which they are conditioned, can provide additional information for SVBRDF estimation compared to photographs. We compare neural architectures and conditions to identify designs that achieve high accuracy and coherence. We find that, surprisingly, a standard UNet is competitive with more complex designs.
Description

CCS Concepts: Computing methodologies -> Texturing; Reflectance modeling

        
@inproceedings{
10.2312:sr.20251186
, booktitle = {
Eurographics Symposium on Rendering
}, editor = {
Wang, Beibei
and
Wilkie, Alexander
}, title = {{
An evaluation of SVBRDF Prediction from Generative Image Models for Appearance Modeling of 3D Scenes
}}, author = {
Gauthier, Alban
and
Deschaintre, Valentin
and
Lanvin, Alexandre
and
Durand, Fredo
and
Bousseau, Adrien
and
Drettakis, George
}, year = {
2025
}, publisher = {
The Eurographics Association
}, ISSN = {
1727-3463
}, ISBN = {
978-3-03868-292-9
}, DOI = {
10.2312/sr.20251186
} }
Citation