Layout👉Scene

Layout Guided Scene Generation via Geometry and Appearance Diffusion Priors

1Sun Yat-sen University, 2Xiamen University, 3National University of Defense Technology

Introduction

TL;DR: Layout2Scene generates 3D scene from human-provided layout and text prompts.

Methodology

Methodology Overview

Our method generates scene meshes and textures from user-specified scene layouts and style prompts through a two-stage framework: (1) Layout-guided geometry generation: We adopt a hybrid scene representation where objects and background are modeled using 3D Gaussians and polygons, respectively. The representation is optimized with a layout-guided geometry diffusion model and an object-level geometry diffusion model. The final scene mesh is extracted from the optimized hybrid scene representation. (2) Layout-guided appearance generation: Given the generated scene mesh, we employ an object-aware neural texture to represent textures for all objects and the background. This representation is optimized via a layout-guided appearance diffusion model and an object-level appearance diffusion model. The optimized neural texture is then converted into high-quality texture maps for each object and background through a texture sampling process.

Interactive Results on Geometry Generation


Input 3D Layout
Output (Ours)
Output (Frankenstein)
Input 3D Layout
Output (Ours)
Output (Frankenstein)

Comparisons on 3D Scene Generation

Style Customization

More

We provide the Blender Addon for easy 3D layout design, while the generated scenes can be directly imported into Blender for further editing.

BibTeX

If you find our approach helpful, you may consider citing our work.

      @article{chen2025layout2scene,
      title={Layout2Scene: 3D semantic layout guided scene generation via geometry and appearance diffusion priors},
      author={Chen, Minglin and Wang, Longguang and Ao, Sheng and Zhang, Ye and Xu, Kai and Guo, Yulan},
      journal={arXiv preprint arXiv:2501.02519},
      year={2025}}