Differentiable Procedural Models for Single-view 3D Mesh Reconstruction

No Thumbnail Available
Date
2023
Authors
Garifullin, Albert
Maiorov, Nikolay
Frolov, Vladimir
Journal Title
Journal ISSN
Volume Title
Publisher
The Eurographics Association
Abstract
Most existing solutions for single-view 3D object reconstruction are based on deep learning with implicit or voxel representations of the scene and are unable to produce detailed and high-quality meshes and textures that can be directly used in practice. Differentiable rendering, on the other hand, is able to produce high-quality meshes but requires several images of an object. We propose a novel approach to single-view 3D reconstruction that uses procedural generator input parameters as a scene representation. Instead of estimating the vertex positions of the mesh directly, we estimate the input parameters of a procedural generator by minimizing the silhouette loss function between reference and rendered images. We use differentiable rendering and create partly differentiable procedural generators to use gradient-based optimization of the loss function. It allows us to create a highly detailed model from a single image taken in an uncontrolled environment. Moreover, the reconstructed model can be further modified in a convenient way by changing the estimated input parameters.
Description

CCS Concepts: Computing methodologies -> Rendering; Shape modeling

        
@inproceedings{
10.2312:cgvc.20231189
, booktitle = {
Computer Graphics and Visual Computing (CGVC)
}, editor = {
Vangorp, Peter
and
Hunter, David
}, title = {{
Differentiable Procedural Models for Single-view 3D Mesh Reconstruction
}}, author = {
Garifullin, Albert
and
Maiorov, Nikolay
and
Frolov, Vladimir
}, year = {
2023
}, publisher = {
The Eurographics Association
}, ISBN = {
978-3-03868-231-8
}, DOI = {
10.2312/cgvc.20231189
} }
Citation