LRM-Zero: Training Large Reconstruction Models with Synthesized Data

Desai Xie1,2, Sai Bi1, Zhixin Shu1, Kai Zhang1, Zexiang Xu1, Yi Zhou1, Sören Pirk3, Arie Kaufman2, Xin Sun1, Hao Tan1
1Adobe Research    2Stony Brook University    3Kiel University   


Training a Large Reconstruction Model (LRM) entirely on synthesized 3D data, achieving high-quality sparse-view 3D reconstruction!


teaser

Sample data from our Zeroverse and Objaverse dataset. LRM-Zero is trained on Zeroverse.

(This webpage contains a lot of videos and interactive viewers. We suggest using Chrome or Edge for the best experience)

Abstract

We present LRM-Zero, a Large Reconstruction Model (LRM) trained entirely on synthesized 3D data, achieving high-quality sparse-view 3D reconstruction. The core of LRM-Zero is our procedural 3D dataset, Zeroverse, which is automatically synthesized from simple primitive shapes with random texturing and augmentations (e.g., height fields, boolean differences, and wireframes). Unlike previous 3D datasets (e.g., Objaverse) which are often captured or crafted by humans to approximate real 3D data, Zeroverse completely ignores realistic global semantics but is rich in complex geometric and texture details that are locally similar to or even more intricate than real objects. We demonstrate that our LRM-Zero, trained with our fully synthesized Zeroverse, can achieve high visual quality in the reconstruction of real-world objects, competitive with models trained on Objaverse. We also analyze several critical design choices of Zeroverse that contribute to LRM-Zero's capability and training stability. Our work demonstrates that 3D reconstruction, one of the core tasks in 3D vision, can potentially be addressed without the semantics of real-world objects.

Zeroverse Dataset

More examples: page 1, page 2

LRM-Zero vs. GS-LRM Comparison

LRM-Zero

GS-LRM

LRM-Zero Results on Text-to-3D and Image-to-3D

Text-to-3D with Instant3D

Image-to-3D with Zero123++

LRM-Zero Results on Google Scanned Objects

(Click to see more results)

LRM-Zero Results on Amazon Berkeley Objects

(Click to see more results)

BibTeX

@misc{xie2024lrmzero,
      title={LRM-Zero: Training Large Reconstruction Models with Synthesized Data},
      author={Desai Xie and Sai Bi and Zhixin Shu and Kai Zhang and Zexiang Xu and Yi Zhou and Sören Pirk and Arie Kaufman and Xin Sun and Hao Tan},
      year={2024},
      eprint={2406.09371},
      archivePrefix={arXiv},
      primaryClass={id='cs.CV' full_name='Computer Vision and Pattern Recognition' is_active=True alt_name=None in_archive='cs' is_general=False description='Covers image processing, computer vision, pattern recognition, and scene understanding. Roughly includes material in ACM Subject Classes I.2.10, I.4, and I.5.'}
}