SPGen: Spherical Projection as Consistent and Flexible Representation for Single Image 3D Shape Generation

1Texas A&M University    2LightSpeed Studios    3HKUST    4Waymo
Affiliation Logos
ACM SIGGRAPH Asia 2025 (Conference Papers)
SPGen Teaser

Teaser: SPGen is a powerful generative model creating consistent 3D shapes with flexible topology from single view images.

Abstract

Existing single-view 3D generative models typically adopt multiview diffusion priors to reconstruct object surfaces, yet they remain prone to inter-view inconsistencies and are unable to faithfully represent complex internal structure or nontrivial topologies.

To address these limitations, we propose SPGen, a novel scalable framework that generates high-quality meshes based on multi-layer Spherical Projection (SP) maps. We encode geometry information by projecting it onto a bounding sphere and unwrapping it into a compact and structural multi-layer 2D representation. This SP-based formulation underpins our image-centric generation pipeline and confers three key advantages:

  • Consistency: The injective SP mapping encodes surface geometry from a spherical viewpoint, naturally eliminating view inconsistency and ambiguity.
  • Flexibility: Multi-layer SP maps represent nested internal structures and support direct lifting to both watertight and open 3D surfaces.
  • Efficiency: Operating in the image domain allows inheriting powerful 2D diffusion priors (like SDXL) and enables efficient finetuning with limited resources.

Methodology

1. Spherical Projection (SP) Representation

Unlike multi-view images which often suffer from inconsistency, or voxel grids which are computationally heavy, we propose Spherical Projection (SP) maps. We enclose the 3D object within a unit sphere and cast rays from the origin. For each ray, we record the intersection depth. To handle complex topologies (e.g., self-occlusions, internal structures), we trace multiple intersections along each ray and store them in multi-layer SP maps.

SP Map Concept
2. Generation Pipeline
SPGen Pipeline

Our pipeline consists of two stages:
(a) AutoEncoder Training: We train a VAE to compress high-resolution multi-layer SP maps into compact latents. We introduce Geometry Regularization (Edge-aware loss & Spectral loss) to preserve high-frequency details.
(b) Diffusion Model Training: We finetune a latent diffusion model (based on SDXL) to generate SP map latents conditioned on a single image. We introduce Layer-wise Self-Attention to ensure alignment between different layers (e.g., ensuring inner structures align with outer shells).

Results

High-Quality Generation

SPGen can reconstruct diverse objects with complex topologies, thin structures, and internal details.

Generation Results

Figure: Generation results on Google Scanned Objects (GSO) and Objaverse.

Comparison with SOTA

Compared to geometry-based (e.g., OpenLRM) and multi-view based (e.g., Wonder3D) methods, SPGen yields more consistent geometry and sharper details.

Comparisons
Comparisons

BibTeX

@article{zhang2025spgen,
  title={SPGen: Spherical Projection as Consistent and Flexible Representation for Single Image 3D Shape Generation},
  author={Zhang, Jingdong and Chen, Weikai and Liu, Yuan and Wang, Jionghao and Yu, Zhengming and Shen, Zhuowen and Yang, Bo and Wang, Wenping and Li, Xin},
  journal={arXiv preprint arXiv:2509.12721},
  year={2025}
}